diff --git a/benchmark_threading.py b/benchmark_threading.py new file mode 100644 index 0000000000000000000000000000000000000000..2a8378984a9a727156e9e87d45fa84e6cb0c1046 --- /dev/null +++ b/benchmark_threading.py @@ -0,0 +1,302 @@ +import os +import time +import argparse +import subprocess +import platform +from typing import Optional, Tuple, Dict +import threading +import numpy as np +from onnx import load, ModelProto +import onnxruntime as ort + +os.environ["XLNX_ENABLE_CACHE"] = "0" +os.environ["PATH"] += ( + os.pathsep + f"{os.environ['CONDA_PREFIX']}\\Lib\\site-packages\\flexmlrt\\lib" +) + +XRT_SMI_PATH = "C:\\Windows\\System32\\AMD\\xrt-smi.exe" + +ONNX_DTYPE_TO_NP = { + "tensor(float)": np.float32, + "tensor(float16)": np.float16, + "tensor(int64)": np.int64, + "tensor(int32)": np.int32, + "tensor(uint16)": np.uint16, + "tensor(int16)": np.int16, + "tensor(uint8)": np.uint8, + "tensor(int8)": np.int8, +} + + +def generate_rand_data_from_onnx( + model: ModelProto, + lowest_int_val: Optional[int], + highest_int_val: Optional[int], +) -> Dict[str, np.ndarray]: + # Load the models + + sess = ort.InferenceSession( + model.SerializePartialToString(), providers=["CPUExecutionProvider"] + ) + + inps = {} + # Iterate over the first models inputs and generate random data + for inp in sess.get_inputs(): + inp_shapes = list(inp.shape) # mutable + for inp_shape in inp_shapes: + assert isinstance( + inp_shape, int + ), f"Found dynamic axes: {inp_shape}. Please freeze." + np_type = ONNX_DTYPE_TO_NP[inp.type] + if np.issubdtype(np_type, np.integer): + iinfo = np.iinfo(np_type) + if lowest_int_val is None: + lowest_int_val = iinfo.min + if highest_int_val is None: + lowest_int_val = iinfo.max + inps[inp.name] = np.random.randint( + lowest_int_val, highest_int_val, size=tuple(inp_shapes), dtype=np_type + ) + else: + inps[inp.name] = np.random.rand(*inp_shapes).astype(np_type) + + return inps + + +def configure_npu_power(p_mode: Optional[str] = None) -> Tuple[int, str, str]: + """ + Configures the NPU power state using xrt-smi.exe. + + Args: + p_mode (string, optional): The desired power mode (p-mode). + If None, displays current status. + Refer to xrt-smi documentation for valid p-modes. + Returns: + tuple: (return_code, stdout, stderr) from the subprocess call. + return_code is an integer, stdout and stderr are strings. + Raises: + OSError: If xrt-smi.exe is not found. + """ + + if platform.system() != "Windows": + return (-1, "xrt-smi.exe is only available on Windows.", "") + + try: + if p_mode is not None: + command = [XRT_SMI_PATH, "configure", "--pmode", str(p_mode)] + else: + command = [ + XRT_SMI_PATH, + "examine", + "--report", + "platform", + ] # Just display status + + process = subprocess.Popen( + command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, text=True + ) + stdout, stderr = process.communicate() + return_code = process.returncode + + if return_code != 0: + print(f"Error executing xrt-smi.exe: {stderr}") + + return return_code, stdout, stderr + + except FileNotFoundError: + raise OSError("xrt-smi.exe not found.") + except Exception as e: # pylint: disable=broad-except + print(f"An unexpected error occurred: {e}") + return -1, "", str(e) + + +def main( + model_file: str, + vaip_config: str, + cache_path: str, + device: str, + pmode: str, + warmup_runs: int, + inferences: int, + lowest_int_value: Optional[int], + highest_int_value: Optional[int], + threads: int, +): + + assert os.path.exists(model_file) + + assert threads >= 1 + + if device == "cpu": + ort_session = ort.InferenceSession( + model_file, + providers=["CPUExecutionProvider"], + ) + + elif device == "npu": + assert os.path.exists(vaip_config) + assert os.path.exists(cache_path) + cache_dir = os.path.dirname(os.path.abspath(cache_path)) + cache_key = os.path.basename(cache_path) + print(f"Using cache directory {cache_dir} with key {cache_key}") + + ret_code, stdout, stderr = configure_npu_power(pmode) + print(stdout) + if ret_code != 0: + print("Error configuring npu power mode.") + print(stderr) + + sess_options = ort.SessionOptions() + ort_session = ort.InferenceSession( + model_file, + providers=["VitisAIExecutionProvider"], + sess_options=sess_options, + provider_options=[ + { + "config_file": vaip_config, + "cacheDir": cache_dir, + "cacheKey": cache_key, + } + ], + ) + + elif device == "igpu": + ort_session = ort.InferenceSession( + model_file, + providers=["DmlExecutionProvider"], + provider_options=[{"device_id": 2}], + ) + + onnx_inputs = generate_rand_data_from_onnx( + load(model_file), + lowest_int_val=lowest_int_value, + highest_int_val=highest_int_value, + ) + + # Warmup + for _ in range(warmup_runs): + ort_session.run(None, onnx_inputs) + + def run_inference(runs, session, inputs): + for _ in range(runs): + session.run(None, inputs) + + latencies = [] + num_threads = threads + threads_list = [] + inferences_per_thread = inferences // num_threads + remainder = inferences % num_threads + print(f"inferences per thread: {inferences_per_thread}, remainder: {remainder}") + start = time.perf_counter() + for i in range(num_threads): + num_runs = inferences_per_thread + (1 if i < remainder else 0) + thread = threading.Thread( + target=run_inference, args=(num_runs, ort_session, onnx_inputs) + ) + threads_list.append(thread) + thread.start() + + for thread in threads_list: + thread.join() + + end = time.perf_counter() + latencies.append((end - start) / inferences) + print(f"Latencies: {latencies}") + print(f"Benchmark results averaged over {inferences} inferences targeting {device}") + print("Average latency (ms): ", round(np.mean(latencies) * 1e3, 3)) + print("Average throughput (inf/s): ", round(1 / np.mean(latencies), 3)) + + +if __name__ == "__main__": + parser = argparse.ArgumentParser( + description="", + ) + + parser.add_argument( + "--pmode", + type=str, + choices=["default", "powersaver", "balanced", "performance", "turbo"], + default="default", + help="Desired power mode.", + ) + + parser.add_argument( + "onnx_model", + type=str, + help="Provide the onnx model file.", + ) + + parser.add_argument( + "--vai-config", + type=str, + help="Path to the vaip configuration json file.", + ) + + parser.add_argument( + "--cache-path", + required=False, + type=str, + help="Path to the saved compilation directory.", + ) + + parser.add_argument( + "--device", + required=False, + type=str, + default="npu", + choices=["npu", "cpu", "igpu"], + help="Select the device to run the measurements on.", + ) + + parser.add_argument( + "--warmup-runs", + required=False, + default=10, + type=int, + help="The number of inferences to run before capturing performance.", + ) + + parser.add_argument( + "--inferences", + required=False, + default=100, + type=int, + help="The number of inferences to average performance over.", + ) + + parser.add_argument( + "--lowest-int-value", + required=False, + type=int, + help="Lowest value the rng will produce if the model has an integer input type.", + ) + + parser.add_argument( + "--highest-int-value", + required=False, + type=int, + help="Highest value the rng will produce if the model has an integer input type.", + ) + + parser.add_argument( + "--threads", + default=1, + required=False, + type=int, + help="The number of threads that are used to run the inferences.", + ) + + args = parser.parse_args() + + main( + args.onnx_model, + args.vai_config, + args.cache_path, + args.device, + args.pmode, + args.warmup_runs, + args.inferences, + args.lowest_int_value, + args.highest_int_value, + args.threads, + ) \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/aie_unsupported_original_ops.json b/segmentation_1_4_0_fp32_combined/aie_unsupported_original_ops.json new file mode 100644 index 0000000000000000000000000000000000000000..66e0d29a6848fc5b9c7ec338ac2464b80c65870a --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/aie_unsupported_original_ops.json @@ -0,0 +1,13 @@ +[ + "Cast_0", + "Transpose_10", + "Transpose_11", + "Transpose_12", + "Transpose_448", + "Transpose_449", + "Transpose_450", + "Transpose_451", + "Transpose_452", + "Transpose_453", + "Transpose_9" +] diff --git a/segmentation_1_4_0_fp32_combined/cache/1d4c9c71f3a11a5a3ebd0c4f9b9709fb907ad630/te_inter_partition_cache.json b/segmentation_1_4_0_fp32_combined/cache/1d4c9c71f3a11a5a3ebd0c4f9b9709fb907ad630/te_inter_partition_cache.json new file mode 100644 index 0000000000000000000000000000000000000000..31fe4854f947b568c940e9600b1de9ec0558b6d4 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/cache/1d4c9c71f3a11a5a3ebd0c4f9b9709fb907ad630/te_inter_partition_cache.json @@ -0,0 +1,7140 @@ +{ + "10026966830493413190": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [16], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [128], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "10062490070065321210": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [6], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [40], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "10307229498188138001": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "10391845491938675702": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "10492525216073892436": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [240], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "10528946888650888391": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [1], + "OW": [64] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [72], + "OH": [1], + "OW": [920] + } + }, + "id": 0 + }, + "10556554590571750004": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [6], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [40], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "10700264929201052950": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [16], + "OH": [12], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [112], + "ICW": [112], + "OC": [48], + "OCW": [48], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "10727544128276376120": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [6], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [40], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "10846026224603292753": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [2], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [960], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "10943316791272103832": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [184], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "10963227159379047924": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [80], + "OH": [1], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "11092279950177326571": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [56], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [112], + "ICW": [112], + "OC": [672], + "OCW": [672], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "11164559383591593506": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "11302748972995246755": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [8], + "OH": [4], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [32], + "ICW": [32], + "OC": [32], + "OCW": [32], + "OH": [30], + "OW": [160] + } + }, + "id": 0 + }, + "11329025555421356232": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [16], + "OH": [1], + "OW": [40] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [120], + "ICW": [120], + "OC": [48], + "OCW": [48], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "11435033089903747205": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [184], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "11580641143542722475": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [240], + "OH": [12], + "OW": [40] + } + }, + "id": 0 + }, + "11861371018295178770": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [192], + "OC": [16], + "OH": [1], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [960], + "ICW": [960], + "OC": [240], + "OCW": [240], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "11961480353149891231": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [96], + "OC": [48], + "OH": [1], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [672], + "ICW": [672], + "OC": [176], + "OCW": [176], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "12302575668607679705": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [12], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [24], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "12387913687178834142": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [80], + "OC": [48], + "OH": [3], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [80], + "ICW": [80], + "OC": [192], + "OCW": [192], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "12562262880184578398": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [2], + "OW": [12] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [64], + "OH": [8], + "OW": [80] + } + }, + "id": 0 + }, + "12565994600343454157": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OC": [8], + "OH": [12], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [40], + "ICW": [40], + "OC": [32], + "OCW": [32], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "12679339250833983242": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [72], + "OC": [16], + "OH": [4], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [72], + "ICW": [72], + "OC": [32], + "OCW": [32], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "12681149571007282408": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [8], + "OH": [2], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [16], + "ICW": [16], + "OC": [16], + "OCW": [16], + "OH": [45], + "OW": [160] + } + }, + "id": 0 + }, + "13017850651716746849": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "13189125416803032844": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "13209035670543806347": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [128], + "OH": [2], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "13263326172031476657": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [8], + "OH": [1], + "OW": [64] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [16], + "ICW": [16], + "OC": [16], + "OCW": [16], + "OH": [20], + "OW": [320] + } + }, + "id": 0 + }, + "13329352869709623705": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [96], + "OC": [32], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [960], + "ICW": [960], + "OC": [128], + "OCW": [128], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "13411565384986468510": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [2], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [72], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "13412111573476215076": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "13416208119636547976": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [24], + "OH": [16], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [72], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "13471986149316703494": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [23], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [16], + "OH": [90], + "OW": [160] + } + }, + "id": 0 + }, + "13483318002960159390": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [8], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [32], + "OH": [30], + "OW": [160] + } + }, + "id": 0 + }, + "13490407344970413206": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [200], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "13652266070640989979": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [80], + "OC": [40], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [160], + "ICW": [160], + "OC": [960], + "OCW": [960], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "13747696295478231722": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [23], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [16], + "OH": [90], + "OW": [160] + } + }, + "id": 0 + }, + "1391098372311032925": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [128], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "13918169330471704249": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [6], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [240], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "13936367568965699966": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [200], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "14028426727709315730": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [184], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "14060132029398729627": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [120], + "OC": [16], + "OH": [1], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [120], + "ICW": [120], + "OC": [32], + "OCW": [32], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "14095873884357302375": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [80], + "OC": [40], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [80], + "ICW": [80], + "OC": [480], + "OCW": [480], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "14182534222412593824": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [64], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "1418352089729515507": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "OC": [8], + "OH": [1], + "OW": [40] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [8], + "OH": [30], + "OW": [160] + } + }, + "id": 0 + }, + "14754737803834076977": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [8], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [120], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "14839517760405474988": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [1], + "OW": [64] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [120], + "OH": [1], + "OW": [920] + } + }, + "id": 0 + }, + "14961823622376311822": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [240], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "14964850546419308832": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "15084648172903469886": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [200], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "15252199640930893805": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [80], + "OC": [40], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [672], + "ICW": [672], + "OC": [160], + "OCW": [160], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "15366874041051119534": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [6], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [40], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "15625897650431120792": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [6], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [72], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "1589904112053860198": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [256], + "OH": [2], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "15998025502086655817": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [5], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [8], + "OH": [20], + "OW": [320] + } + }, + "id": 0 + }, + "1616091849014086957": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [16], + "OH": [3], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [128], + "ICW": [128], + "OC": [64], + "OCW": [64], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "16402378040432170905": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [12], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [24], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "16424411160853745392": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [120], + "OC": [32], + "OH": [1], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [120], + "ICW": [120], + "OC": [480], + "OCW": [480], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "16543734521308364643": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [96], + "OC": [32], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [480], + "ICW": [480], + "OC": [112], + "OCW": [112], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "16660180594794636685": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [16], + "OH": [9], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [16], + "ICW": [16], + "OC": [64], + "OCW": [64], + "OH": [18], + "OW": [160] + } + }, + "id": 0 + }, + "16837037073375920586": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [24], + "OH": [2], + "OW": [40] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [80], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "16909720421144648317": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [16], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [120], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "16955326269104416555": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [5], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [8], + "OH": [20], + "OW": [320] + } + }, + "id": 0 + }, + "1701679871653718655": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [64], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "17047541884526191672": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OC": [8], + "OH": [4], + "OW": [64] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [40], + "ICW": [40], + "OC": [16], + "OCW": [16], + "OH": [15], + "OW": [320] + } + }, + "id": 0 + }, + "17118891849331379395": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [8], + "OH": [4], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [32], + "ICW": [32], + "OC": [16], + "OCW": [16], + "OH": [30], + "OW": [160] + } + }, + "id": 0 + }, + "1719072286549233503": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [96], + "OC": [32], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [672], + "ICW": [672], + "OC": [112], + "OCW": [112], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "17257814780956834928": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [64], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "17320045243823864997": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [200], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "17380671368068608883": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [192], + "OH": [2], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "17484157022346861046": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [40], + "OH": [1], + "OW": [48] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [1], + "OW": [240] + } + }, + "id": 0 + }, + "17503289085723527663": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [256], + "OH": [1], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "17907637872848229909": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [80], + "OC": [40], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [160], + "ICW": [160], + "OC": [960], + "OCW": [960], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "17979705192304779350": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [8], + "OH": [5], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [16], + "ICW": [16], + "OC": [16], + "OCW": [16], + "OH": [20], + "OW": [320] + } + }, + "id": 0 + }, + "18122791096039507878": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "18140749361584105424": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [4], + "OW": [24] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [200], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "18179375394603587330": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "18232259666445886115": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [32], + "OH": [2], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [40], + "ICW": [40], + "OC": [128], + "OCW": [128], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "18247157116488252783": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [32], + "OH": [2], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [40], + "ICW": [40], + "OC": [240], + "OCW": [240], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "1835817167490747973": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [184], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "18402862448763518559": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [24], + "OH": [20], + "OW": [4] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [184], + "OH": [320], + "OW": [4] + } + }, + "id": 0 + }, + "18439626577884786362": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [4], + "OW": [4] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [240], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "192911495949292927": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "OC": [8], + "OH": [6], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [8], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "1983107868992015926": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [12], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [120], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "209030779376412955": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [128], + "OH": [3], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "2146746443699465338": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [16], + "OH": [3], + "OW": [48] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [80], + "ICW": [80], + "OC": [80], + "OCW": [80], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "2344071606782349046": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [128], + "OH": [4], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "2371879921412442941": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [80], + "OH": [1], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "2374470417915057613": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [184], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "2579917799359443473": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [80], + "OH": [1], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "2704563428560469458": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "OC": [8], + "OH": [6], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [8], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "271032871684098807": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [8], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [16], + "OH": [30], + "OW": [160] + } + }, + "id": 0 + }, + "2716124995627608748": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [112], + "OC": [24], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [960], + "ICW": [960], + "OC": [160], + "OCW": [160], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "2920138582029460520": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OC": [8], + "OH": [4], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [8], + "ICW": [8], + "OC": [16], + "OCW": [16], + "OH": [30], + "OW": [160] + } + }, + "id": 0 + }, + "3066705598883327806": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [120], + "OC": [32], + "OH": [1], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [480], + "ICW": [480], + "OC": [128], + "OCW": [128], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "3168718690300333904": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [64], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "318270813535495553": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [6], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [120], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "3221831423333948378": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [5], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [8], + "OH": [20], + "OW": [320] + } + }, + "id": 0 + }, + "3339293583222280981": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [5], + "OW": [40] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [8], + "OH": [20], + "OW": [320] + } + }, + "id": 0 + }, + "3367109530377759015": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [4], + "OW": [24] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [184], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "3435078970592335176": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [16], + "OH": [12], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [40], + "ICW": [40], + "OC": [48], + "OCW": [48], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "3692032364510103941": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [8], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [16], + "OH": [30], + "OW": [160] + } + }, + "id": 0 + }, + "3726406705491134724": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [24], + "OH": [2], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [24], + "ICW": [24], + "OC": [80], + "OCW": [80], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "3843426769902129985": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [6], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [72], + "OH": [23], + "OW": [80] + } + }, + "id": 0 + }, + "3911351982411912144": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [80], + "OC": [48], + "OH": [1], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [240], + "ICW": [240], + "OC": [960], + "OCW": [960], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "3941150613822858515": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [72], + "OC": [16], + "OH": [1], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [72], + "ICW": [72], + "OC": [32], + "OCW": [32], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "3987208456301126832": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [8], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [16], + "OH": [30], + "OW": [160] + } + }, + "id": 0 + }, + "4008165570430271579": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [56], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [1], + "OW": [240] + } + }, + "id": 0 + }, + "4023880041000524543": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [184], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "4169644168289759965": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [240], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "4177912110409701844": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [6], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [240], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "4255029763023846553": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [8], + "OH": [5], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [64], + "ICW": [64], + "OC": [32], + "OCW": [32], + "OH": [18], + "OW": [160] + } + }, + "id": 0 + }, + "4374783739139617614": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [16], + "OH": [3], + "OW": [48] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [80], + "ICW": [80], + "OC": [48], + "OCW": [48], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "4592911515747169522": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [12], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [24], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "462922538141428877": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [6], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [240], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "4645702201520416946": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [23], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [16], + "OH": [90], + "OW": [160] + } + }, + "id": 0 + }, + "4801621660652861968": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [8], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [16], + "OH": [30], + "OW": [160] + } + }, + "id": 0 + }, + "4825666642853207080": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [12], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [16], + "OH": [45], + "OW": [160] + } + }, + "id": 0 + }, + "4921520940930150677": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [64], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "4997824680240234763": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [80], + "OH": [1], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "5063133298956954947": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [6], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [120], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "5249313651399056560": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [24], + "OH": [2], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [6], + "OW": [20] + } + }, + "id": 0 + }, + "5333885574942786145": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [24], + "OH": [2], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [6], + "OW": [20] + } + }, + "id": 0 + }, + "574394342223455753": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [16], + "OH": [3], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [128], + "ICW": [128], + "OC": [128], + "OCW": [128], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "6173144424694747470": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [32], + "OH": [1], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [32], + "ICW": [32], + "OC": [128], + "OCW": [128], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "6225751594267881183": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [192], + "OC": [16], + "OH": [1], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [960], + "ICW": [960], + "OC": [128], + "OCW": [128], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "6322166321066499673": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [80], + "OC": [32], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [80], + "ICW": [80], + "OC": [208], + "OCW": [208], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "6347008778693772383": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [72], + "OC": [16], + "OH": [6], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [72], + "ICW": [72], + "OC": [48], + "OCW": [48], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "6657658657472651100": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [168], + "OC": [16], + "OH": [1], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [168], + "ICW": [168], + "OC": [672], + "OCW": [672], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "6782098364165148542": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [12] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [1], + "K": [1], + "R": [1] + }, + "OW": { + "C": [4], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [32], + "OH": [1], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [24], + "ICW": [24], + "OC": [80], + "OCW": [80], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "6787596214171914840": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "6795874441863046431": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [12], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [24], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "6805931890748967918": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [2], + "OW": [8] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [672], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "7067180749348668692": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [200], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "7118199671802281586": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [64], + "OC": [16], + "OH": [4], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [64], + "ICW": [64], + "OC": [32], + "OCW": [32], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "7174443525992078454": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [6], + "OW": [16] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [16], + "OH": [90], + "OW": [160] + } + }, + "id": 0 + }, + "7502050776740885325": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [4], + "OW": [24] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [480], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "7639803783576501424": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [6], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [40], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "7648296221347465684": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "7656204737256184836": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [12], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [24], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "7658681339734370732": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "7752260818922033472": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [8], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [72], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "7894921693921187902": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [256], + "OH": [2], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "7899951071397421864": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [40], + "OH": [1], + "OW": [48] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [1], + "OW": [240] + } + }, + "id": 0 + }, + "7973887560263329039": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [480], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "8053843256516049804": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [200], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "8197908156936426338": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "8395564477653063847": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [104], + "OC": [24], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [200], + "ICW": [200], + "OC": [80], + "OCW": [80], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "8555647691124563120": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [240], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "8604830178297017717": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [12], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [40], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "8746170461996602897": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [256], + "OH": [1], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "8823441815797889127": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [672], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "8869778043365763113": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [96], + "OC": [24], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [184], + "ICW": [184], + "OC": [80], + "OCW": [80], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "9141387259277866585": { + "0": { + "config": {}, + "permutation": ["OW", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [8], + "OH": [4], + "OW": [24] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [672], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "9146042183721010400": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [3], + "OW": [20] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [64], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "9221809710136756139": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [12], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [24], + "OH": [45], + "OW": [80] + } + }, + "id": 0 + }, + "9278841402651738410": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [16], + "OH": [6], + "OW": [10] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [40], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "9651539258418773113": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [64] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [80], + "OC": [24], + "OH": [1], + "OW": [32] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [240], + "ICW": [240], + "OC": [80], + "OCW": [80], + "OH": [12], + "OW": [20] + } + }, + "id": 0 + }, + "9727613062865824267": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [8], + "OH": [5], + "OW": [40] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [8], + "OH": [20], + "OW": [320] + } + }, + "id": 0 + }, + "9880668269547456824": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [384], + "OH": [1], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [960], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + }, + "9884921235509989184": { + "0": { + "config": { + "config.conv_type": { + "type": "uint64", + "values": [0] + } + }, + "permutation": ["OW", "OC", "OH", "IC", "B"], + "spatial": { + "B": { + "C": [1], + "K": [1], + "R": [1] + }, + "IC": { + "C": [1], + "K": [1], + "R": [1] + }, + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "B": [1], + "IC": [16], + "OC": [16], + "OH": [3], + "OW": [48] + } + }, + "1": { + "temporal": { + "B": [1], + "IC": [176], + "ICW": [176], + "OC": [80], + "OCW": [80], + "OH": [23], + "OW": [40] + } + }, + "id": 0 + }, + "9986995421205970290": { + "0": { + "config": {}, + "permutation": ["OW", "OC", "OH"], + "spatial": { + "OC": { + "C": [1], + "K": [1], + "R": [4] + }, + "OH": { + "C": [4], + "K": [1], + "R": [1] + }, + "OW": { + "C": [1], + "K": [1], + "R": [1] + } + }, + "temporal": { + "OC": [32], + "OH": [16], + "OW": [1] + } + }, + "1": { + "temporal": { + "B": [1], + "OC": [72], + "OH": [1], + "OW": [1] + } + }, + "id": 0 + } +} diff --git a/segmentation_1_4_0_fp32_combined/context.json b/segmentation_1_4_0_fp32_combined/context.json new file mode 100644 index 0000000000000000000000000000000000000000..806bd82b0bebd5441d015c3caddb62ebfef0bca2 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/context.json @@ -0,0 +1,1052 @@ +{ + "metaDef": [ + { + "id": "vaiml_par_0", + "inputs": [ + "385", + "394", + "395", + "396", + "397" + ], + "outputs": [ + "921", + "894", + "868", + "832", + "796", + "916" + ], + "nodes": [ + "921", + "909", + "908", + "907", + "1081", + "904", + "1078", + "901", + "900", + "895", + "894", + "893", + "889", + "888", + "887", + "886", + "884", + "883", + "882", + "881", + "879", + "878", + "1075", + "875", + "874", + "869", + "868", + "867", + "863", + "862", + "861", + "860", + "858", + "857", + "856", + "855", + "853", + "852", + "1072", + "849", + "848", + "838", + "833", + "832", + "831", + "827", + "826", + "825", + "824", + "822", + "821", + "820", + "819", + "817", + "816", + "1069", + "813", + "812", + "802", + "797", + "796", + "795", + "791", + "790", + "789", + "788", + "786", + "785", + "784", + "783", + "781", + "777", + "776", + "775", + "774", + "770", + "769", + "767", + "764", + "1063", + "760", + "1060", + "757", + "756", + "754", + "751", + "749", + "748", + "747", + "746", + "745", + "744", + "742", + "739", + "1057", + "735", + "734", + "732", + "729", + "1054", + "725", + "1051", + "722", + "721", + "719", + "716", + "714", + "713", + "712", + "711", + "710", + "709", + "707", + "704", + "1048", + "700", + "699", + "697", + "694", + "1045", + "1042", + "688", + "687", + "685", + "682", + "680", + "679", + "678", + "677", + "676", + "675", + "673", + "670", + "1039", + "666", + "665", + "663", + "660", + "1036", + "656", + "1033", + "653", + "652", + "650", + "647", + "645", + "644", + "643", + "642", + "641", + "640", + "638", + "635", + "1030", + "631", + "630", + "628", + "625", + "1027", + "1024", + "619", + "618", + "616", + "613", + "611", + "610", + "609", + "608", + "607", + "606", + "604", + "601", + "1021", + "597", + "596", + "594", + "591", + "1018", + "587", + "1015", + "584", + "583", + "581", + "578", + "1012", + "574", + "573", + "571", + "568", + "1009", + "564", + "1006", + "561", + "560", + "558", + "555", + "1003", + "551", + "550", + "548", + "545", + "1000", + "541", + "997", + "538", + "537", + "535", + "532", + "994", + "528", + "527", + "525", + "522", + "991", + "988", + "516", + "515", + "513", + "510", + "985", + "506", + "505", + "503", + "500", + "982", + "496", + "979", + "493", + "492", + "490", + "487", + "485", + "484", + "483", + "482", + "481", + "976", + "478", + "973", + "475", + "970", + "472", + "471", + "469", + "466", + "464", + "463", + "462", + "461", + "460", + "967", + "457", + "964", + "961", + "452", + "451", + "449", + "446", + "444", + "443", + "442", + "441", + "440", + "958", + "437", + "955", + "434", + "952", + "431", + "949", + "428", + "946", + "943", + "423", + "940", + "420", + "937", + "417", + "934", + "414", + "931", + "411", + "410", + "408", + "405", + "928", + "401", + "399", + "393", + "392", + "387", + "773", + "1066", + "794", + "793", + "780", + "779", + "778", + "830", + "829", + "866", + "865", + "892", + "891", + "916", + "911" + ], + "constantInitializers": [ + "1001", + "1002", + "1004", + "1005", + "1007", + "1008", + "1010", + "1011", + "1013", + "1014", + "1016", + "1017", + "1019", + "1020", + "1022", + "1023", + "1025", + "1026", + "1028", + "1029", + "1031", + "1032", + "1034", + "1035", + "1037", + "1038", + "1040", + "1041", + "1043", + "1044", + "1046", + "1047", + "1049", + "1050", + "1052", + "1053", + "1055", + "1056", + "1058", + "1059", + "1061", + "1062", + "1064", + "1065", + "1067", + "1068", + "1070", + "1071", + "1073", + "1074", + "1076", + "1077", + "1079", + "1080", + "1082", + "1083", + "1086", + "1090", + "386", + "388", + "389", + "398", + "400", + "752", + "755", + "763", + "801", + "809", + "837", + "845", + "847", + "873", + "890", + "899", + "929", + "930", + "932", + "933", + "935", + "936", + "938", + "939", + "941", + "942", + "944", + "945", + "947", + "948", + "950", + "951", + "953", + "954", + "956", + "957", + "959", + "960", + "962", + "963", + "965", + "966", + "968", + "969", + "971", + "972", + "974", + "975", + "977", + "978", + "980", + "981", + "983", + "984", + "986", + "987", + "989", + "990", + "992", + "993", + "995", + "996", + "998", + "999", + "aspp.aspp2.1.weight", + "backbone.features.11.block.2.fc1.bias", + "backbone.features.11.block.2.fc1.weight", + "backbone.features.11.block.2.fc2.bias", + "backbone.features.11.block.2.fc2.weight", + "backbone.features.12.block.2.fc1.bias", + "backbone.features.12.block.2.fc1.weight", + "backbone.features.12.block.2.fc2.bias", + "backbone.features.12.block.2.fc2.weight", + "backbone.features.13.block.2.fc1.bias", + "backbone.features.13.block.2.fc1.weight", + "backbone.features.13.block.2.fc2.bias", + "backbone.features.13.block.2.fc2.weight", + "backbone.features.14.block.2.fc1.bias", + "backbone.features.14.block.2.fc1.weight", + "backbone.features.14.block.2.fc2.bias", + "backbone.features.14.block.2.fc2.weight", + "backbone.features.15.block.2.fc1.bias", + "backbone.features.15.block.2.fc1.weight", + "backbone.features.15.block.2.fc2.bias", + "backbone.features.15.block.2.fc2.weight", + "backbone.features.4.block.2.fc1.bias", + "backbone.features.4.block.2.fc1.weight", + "backbone.features.4.block.2.fc2.bias", + "backbone.features.4.block.2.fc2.weight", + "backbone.features.5.block.2.fc1.bias", + "backbone.features.5.block.2.fc1.weight", + "backbone.features.5.block.2.fc2.bias", + "backbone.features.5.block.2.fc2.weight", + "backbone.features.6.block.2.fc1.bias", + "backbone.features.6.block.2.fc1.weight", + "backbone.features.6.block.2.fc2.bias", + "backbone.features.6.block.2.fc2.weight", + "decoder.decode1.gru.hh.0.bias", + "decoder.decode1.gru.hh.0.weight", + "decoder.decode1.gru.ih.0.bias", + "decoder.decode1.gru.ih.0.weight", + "decoder.decode2.gru.hh.0.bias", + "decoder.decode2.gru.hh.0.weight", + "decoder.decode2.gru.ih.0.bias", + "decoder.decode2.gru.ih.0.weight", + "decoder.decode3.gru.hh.0.bias", + "decoder.decode3.gru.hh.0.weight", + "decoder.decode3.gru.ih.0.bias", + "decoder.decode3.gru.ih.0.weight", + "decoder.decode4.gru.hh.0.bias", + "decoder.decode4.gru.hh.0.weight", + "decoder.decode4.gru.ih.0.bias", + "decoder.decode4.gru.ih.0.weight", + "project_mat.conv.bias", + "project_mat.conv.weight" + ], + "device": "VAIML", + "vaimlParam": { + "vaimlModelPath": "./segmentation_1_4_0_fp32_combined/vaiml_par_0", + "deviceName": "stx", + "outputShapes": [ + { + "shapes": [ + "1", + "1", + "180", + "320" + ] + }, + { + "shapes": [ + "1", + "16", + "90", + "160" + ] + }, + { + "shapes": [ + "1", + "20", + "45", + "80" + ] + }, + { + "shapes": [ + "1", + "40", + "23", + "40" + ] + }, + { + "shapes": [ + "1", + "64", + "12", + "20" + ] + }, + { + "shapes": [ + "1", + "3", + "180", + "320" + ] + } + ], + "inputShapes": [ + { + "shapes": [ + "1", + "180", + "320", + "4" + ] + }, + { + "shapes": [ + "1", + "16", + "90", + "160" + ] + }, + { + "shapes": [ + "1", + "20", + "45", + "80" + ] + }, + { + "shapes": [ + "1", + "40", + "23", + "40" + ] + }, + { + "shapes": [ + "1", + "64", + "12", + "20" + ] + } + ], + "vaimlUnarchivePath": "./segmentation_1_4_0_fp32_combined", + "subgraphName": "vaiml_par_0", + "priority": "normal", + "configJson": "vitisai_config.json", + "deviceBatchSize": 1, + "inputNames": [ + "385", + "394", + "395", + "396", + "397" + ], + "outputNames": [ + "921", + "894", + "868", + "832", + "796", + "916" + ] + } + } + ], + "config": { + "passes": [ + { + "name": "init", + "plugin": "vaip-pass_init" + }, + { + "name": "vaiml_partition", + "plugin": "vaip-pass_vaiml_partition", + "vaimlConfig": { + "device": "stx", + "enableF32ToBf16Conversion": true, + "keepOutputs": true + } + } + ], + "cacheDir": "/tmp/vaip/.cache", + "cacheKey": "4e8d9ee8719e050537726a1cd8f79305", + "version": { + "versionInfos": [ + { + "packageName": "vaip", + "commit": "e4ca074a034e568a2fd44af176f112e64e845411", + "version": "vaip.1.0.0" + }, + { + "packageName": "target_factory", + "commit": "60f0780e75861ed37681e00c5eeeaf132f887c45", + "version": "target-factory.3.5.0" + }, + { + "packageName": "vart", + "commit": "01141f8d14af4be8a1229d3d5dd93f26ba608f58", + "version": "vart" + }, + { + "packageName": "xcompiler", + "commit": "8ccbd710317a738251bbddcebb919fd0e8145250", + "version": "xcompiler.3.5.0" + }, + { + "packageName": "onnxrutnime", + "commit": "5c1b7ccbff7e5141c1da7a9d963d660e5741c319", + "version": "onnxruntime.1.20.1" + }, + { + "packageName": "xir", + "commit": "bef5d269fc06f065c5f8de5ba202d6f02edfad7f", + "version": "xir.3.5.0" + }, + { + "packageName": "xrt", + "commit": "acc144998d650acbfda7e5919a1290de8f8c7735", + "version": "xrt.2.19.72" + }, + { + "packageName": "graph_engine", + "commit": "d7385f8afa5f6eb7d5d4452749b0e07bce67b75b", + "version": "graph_engine" + } + ] + }, + "onnxPath": "segmentation_1_4_0_fp32_combined.onnx", + "sessionOptions": { + "config_file": "vitisai_config.json" + }, + "enableCacheFileIoInMem": false + }, + "events": [ + { + "name": "before_compile_onnx_model_internal", + "ph": "X", + "ts": "31561", + "pid": "214", + "tid": "214", + "args": { + "memUsage": {} + }, + "dur": "30664" + }, + { + "id": "before_compile_onnx_model_internal_mem_usage_1", + "ph": "v", + "ts": "31561", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "id": "before_compile_onnx_model_internal_mem_usage_2", + "ph": "v", + "ts": "62226", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "name": "check_cache_hit", + "ph": "X", + "ts": "62248", + "pid": "214", + "tid": "214", + "args": { + "memUsage": {} + }, + "dur": "25" + }, + { + "id": "check_cache_hit_mem_usage_1", + "ph": "v", + "ts": "62248", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "id": "check_cache_hit_mem_usage_2", + "ph": "v", + "ts": "62273", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "name": "0-init@vaip-pass_init", + "ph": "X", + "ts": "65069", + "pid": "214", + "tid": "214", + "args": { + "memUsage": {} + }, + "dur": "11" + }, + { + "id": "0-init@vaip-pass_init_mem_usage_1", + "ph": "v", + "ts": "65069", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "id": "0-init@vaip-pass_init_mem_usage_2", + "ph": "v", + "ts": "65080", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "name": "1-vaiml_partition@vaip-pass_vaiml_partition", + "ph": "X", + "ts": "65090", + "pid": "214", + "tid": "214", + "args": { + "memUsage": {} + }, + "dur": "8266564167" + }, + { + "id": "1-vaiml_partition@vaip-pass_vaiml_partition_mem_usage_1", + "ph": "v", + "ts": "65090", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "id": "1-vaiml_partition@vaip-pass_vaiml_partition_mem_usage_2", + "ph": "v", + "ts": "8266629258", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "name": "update_cache", + "ph": "X", + "ts": "62287", + "pid": "214", + "tid": "214", + "args": { + "memUsage": {} + }, + "dur": "8266568873" + }, + { + "id": "update_cache_mem_usage_1", + "ph": "v", + "ts": "62287", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "id": "update_cache_mem_usage_2", + "ph": "v", + "ts": "8266631160", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "name": "read_cache", + "ph": "X", + "ts": "8266632901", + "pid": "214", + "tid": "214", + "args": { + "memUsage": {} + }, + "dur": "1033" + }, + { + "id": "read_cache_mem_usage_1", + "ph": "v", + "ts": "8266632901", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "id": "read_cache_mem_usage_2", + "ph": "v", + "ts": "8266633935", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "name": "after_compile_onnx_model_internal", + "ph": "X", + "ts": "8266633946", + "pid": "214", + "tid": "214", + "args": { + "memUsage": {} + }, + "dur": "26451" + }, + { + "id": "after_compile_onnx_model_internal_mem_usage_1", + "ph": "v", + "ts": "8266633946", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "id": "after_compile_onnx_model_internal_mem_usage_2", + "ph": "v", + "ts": "8266660397", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "name": "compile_onnx_model_internal", + "ph": "X", + "ts": "31560", + "pid": "214", + "tid": "214", + "args": { + "memUsage": {} + }, + "dur": "8266628844" + }, + { + "id": "compile_onnx_model_internal_mem_usage_1", + "ph": "v", + "ts": "31560", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "id": "compile_onnx_model_internal_mem_usage_2", + "ph": "v", + "ts": "8266660404", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "name": "compile_onnx_model_3", + "ph": "X", + "ts": "31447", + "pid": "214", + "tid": "214", + "args": { + "memUsage": {} + }, + "dur": "8266629035" + }, + { + "id": "compile_onnx_model_3_mem_usage_1", + "ph": "v", + "ts": "31447", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + }, + { + "id": "compile_onnx_model_3_mem_usage_2", + "ph": "v", + "ts": "8266660483", + "pid": "214", + "args": { + "dumps": { + "process_totals": { + "peak_memory": "0", + "current_memory": "0" + } + } + } + } + ], + "cpuUsage": [ + { + "avgCpuUtil": 75.2757797, + "memPeakWorkingSetSize": 1998.73828 + } + ], + "cacheFiles": [ + "context.json" + ] +} diff --git a/segmentation_1_4_0_fp32_combined/final-vaiml-pass-summary.txt b/segmentation_1_4_0_fp32_combined/final-vaiml-pass-summary.txt new file mode 100644 index 0000000000000000000000000000000000000000..9503c82ac5e915fbc6764d3203e91c81c86eb2be --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/final-vaiml-pass-summary.txt @@ -0,0 +1,23 @@ +--------- Final Summary of VAIML Pass ---------- +VAIP commit: e4ca074a034e568a2fd44af176f112e64e845411 +Model: segmentation_1_4_0_fp32_combined.onnx +Model signature: 6bbb891ab96ca9362e0e61024cd02778 +Device: stx +Model data type: float32 +Device data type: bfloat16 +Number of operators in the model: 317 +GOPs of the model: 3.73517 +Number of operators supported by VAIML: 306 (96.530%) +GOPs supported by VAIML: 3.735 (100.000%) +Number of subgraphs supported by VAIML: 1 +Number of operators offloaded by VAIML: 306 (96.530%) +GOPs offloaded by VAIML: 3.735 (100.000%) +Number of subgraphs offloaded by VAIML: 1 +Number of subgraphs with compilation errors (fall back to CPU): 0 +Number of subgraphs below 2% GOPs threshold (fall back to CPU): 0 +Stats for offloaded subgraphs +Subgraph vaiml_par_0 stats: + Operators: 306 (96.530%) + GOPs : 3.735 (100.000%) OPs: 3,735,173,696 + + diff --git a/segmentation_1_4_0_fp32_combined/gops.csv b/segmentation_1_4_0_fp32_combined/gops.csv new file mode 100644 index 0000000000000000000000000000000000000000..dd282b256c14b0e48337f3d6152032b0266aab16 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/gops.csv @@ -0,0 +1,318 @@ +Node,OPs,Note +Add_105,115200 +Add_115,96000 +Add_124,96000 +Add_132,38400 +Add_135,88320 +Add_144,88320 +Add_152,38400 +Add_155,88320 +Add_164,88320 +Add_172,38400 +Add_175,230400 +Add_184,230400 +Add_19,460800 +Add_196,960 +Add_206,322560 +Add_215,322560 +Add_227,1344 +Add_235,53760 +Add_238,322560 +Add_247,322560 +Add_259,1344 +Add_269,460800 +Add_278,460800 +Add_29,460800 +Add_290,1920 +Add_298,76800 +Add_301,460800 +Add_310,460800 +Add_322,1920 +Add_330,76800 +Add_333,460800 +Add_362,30720 +Add_388,73600 +Add_40,172800 +Add_414,144000 +Add_434,460800 +Add_445,345600 +Add_50,144 +Add_67,240 +Add_75,73600 +Add_85,240 +Add_93,73600 +Add_96,441600 +AveragePool_346,0 +AveragePool_347,0 +AveragePool_348,0 +Cast_0,0 +Clip_108,230400 +Clip_118,192000 +Clip_127,192000 +Clip_138,176640 +Clip_147,176640 +Clip_158,176640 +Clip_167,176640 +Clip_178,460800 +Clip_187,460800 +Clip_199,1920 +Clip_209,645120 +Clip_218,645120 +Clip_22,921600 +Clip_230,2688 +Clip_241,645120 +Clip_250,645120 +Clip_262,2688 +Clip_272,921600 +Clip_281,921600 +Clip_293,3840 +Clip_304,921600 +Clip_313,921600 +Clip_325,3840 +Clip_336,921600 +Clip_446,691200 +Clip_447,230400 +Clip_53,288 +Clip_70,480 +Clip_88,480 +Clip_99,883200 +Concat_350,30720 +Concat_355,30720 +Concat_363,30720 +Concat_372,235520 +Concat_376,73600 +Concat_381,73600 +Concat_389,73600 +Concat_398,576000 +Concat_402,144000 +Concat_407,144000 +Concat_415,144000 +Concat_418,1152000 +Concat_422,460800 +Concat_427,460800 +Concat_435,460800 +Concat_438,3686400 +Conv_103,1152000 +Conv_112,9254400 +Conv_113,7776000 +Conv_122,960000 +Conv_131,7718400 +Conv_133,7153920 +Conv_142,883200 +Conv_151,7104000 +Conv_153,7153920 +Conv_162,883200 +Conv_17,12902400 +Conv_171,7104000 +Conv_173,18662400 +Conv_182,2304000 +Conv_192,115440 +Conv_194,116160 +Conv_203,25858560 +Conv_204,36449280 +Conv_213,3225600 +Conv_223,226128 +Conv_225,227136 +Conv_234,36180480 +Conv_236,36449280 +Conv_245,8386560 +Conv_255,226128 +Conv_257,227136 +Conv_26,4608000 +Conv_266,51686400 +Conv_267,74188800 +Conv_276,11980800 +Conv_28,7833600 +Conv_286,461280 +Conv_288,462720 +Conv_297,73804800 +Conv_299,74188800 +Conv_30,31334400 +Conv_308,11980800 +Conv_318,461280 +Conv_32,4608000 +Conv_320,462720 +Conv_329,73804800 +Conv_331,74188800 +Conv_34,11232000 +Conv_340,59043840 +Conv_343,246016 +Conv_35,12960000 +Conv_351,70840320 +Conv_356,35420160 +Conv_37,5184000 +Conv_373,226688000 +Conv_377,106131200 +Conv_382,53065600 +Conv_39,12614400 +Conv_399,277632000 +Conv_403,103968000 +Conv_408,51984000 +Conv_41,12960000 +Conv_419,490291200 +Conv_423,266342400 +Conv_428,133171200 +Conv_43,3444480 +Conv_439,582451200 +Conv_441,267264000 +Conv_443,7833600 +Conv_46,3504 +Conv_48,3600 +Conv_57,5372800 +Conv_58,9052800 +Conv_60,5740800 +Conv_63,7744 +Conv_65,7920 +Conv_74,8905600 +Conv_76,9052800 +Conv_78,5740800 +Conv_81,7744 +Conv_83,7920 +Conv_92,8905600 +Conv_94,18105600 +Div_101,1766400 +Div_110,460800 +Div_120,384000 +Div_129,384000 +Div_140,353280 +Div_149,353280 +Div_16,1382400 +Div_160,353280 +Div_169,353280 +Div_180,921600 +Div_189,921600 +Div_2,1843200 +Div_201,3840 +Div_211,1290240 +Div_220,1290240 +Div_232,5376 +Div_24,1843200 +Div_243,1290240 +Div_252,1290240 +Div_264,5376 +Div_274,1843200 +Div_283,1843200 +Div_295,7680 +Div_306,1843200 +Div_315,1843200 +Div_327,7680 +Div_338,1843200 +Div_55,576 +Div_72,960 +Div_90,960 +GlobalAveragePool_191,234240 +GlobalAveragePool_222,327936 +GlobalAveragePool_254,327936 +GlobalAveragePool_285,468480 +GlobalAveragePool_317,468480 +GlobalAveragePool_342,468480 +GlobalAveragePool_45,133056 +GlobalAveragePool_62,221760 +GlobalAveragePool_80,221760 +Mul_102,441600 +Mul_111,115200 +Mul_121,96000 +Mul_130,96000 +Mul_141,88320 +Mul_150,88320 +Mul_161,88320 +Mul_170,88320 +Mul_181,230400 +Mul_190,230400 +Mul_202,230400 +Mul_212,322560 +Mul_221,322560 +Mul_233,322560 +Mul_244,322560 +Mul_25,460800 +Mul_253,322560 +Mul_265,322560 +Mul_275,460800 +Mul_284,460800 +Mul_296,460800 +Mul_307,460800 +Mul_316,460800 +Mul_328,460800 +Mul_339,460800 +Mul_345,61440 +Mul_354,30720 +Mul_360,30720 +Mul_361,30720 +Mul_380,73600 +Mul_386,73600 +Mul_387,73600 +Mul_406,144000 +Mul_412,144000 +Mul_413,144000 +Mul_426,460800 +Mul_432,460800 +Mul_433,460800 +Mul_56,132480 +Mul_73,220800 +Mul_91,220800 +Relu_193,240 +Relu_224,336 +Relu_256,336 +Relu_27,460800 +Relu_287,480 +Relu_31,1843200 +Relu_319,480 +Relu_33,460800 +Relu_341,61440 +Relu_36,518400 +Relu_374,147200 +Relu_38,518400 +Relu_400,288000 +Relu_42,518400 +Relu_420,921600 +Relu_44,132480 +Relu_440,1843200 +Relu_442,1843200 +Relu_47,48 +Relu_59,220800 +Relu_61,220800 +Relu_64,64 +Relu_77,220800 +Relu_79,220800 +Relu_82,64 +Resize_365,245760 +Resize_391,588800 +Resize_417,1152000 +Resize_437,3686400 +Sigmoid_344,0 +Sigmoid_352,0 +Sigmoid_378,0 +Sigmoid_404,0 +Sigmoid_424,0 +Slice_371,235520 +Slice_397,576000 +Slice_7,345600 +Split_349,0 +Split_353,0 +Split_375,0 +Split_379,0 +Split_401,0 +Split_405,0 +Split_421,0 +Split_425,0 +Split_444,0 +Sub_14,345600 +Sub_359,30720 +Sub_385,73600 +Sub_411,144000 +Sub_431,460800 +Tanh_357,2334720 +Tanh_383,5593600 +Tanh_409,10944000 +Tanh_429,35020800 +Transpose_10,0 +Transpose_11,0 +Transpose_12,0 +Transpose_448,0 +Transpose_449,0 +Transpose_450,0 +Transpose_451,0 +Transpose_452,0 +Transpose_453,0 +Transpose_8,0 +Transpose_9,0 diff --git a/segmentation_1_4_0_fp32_combined/graph_partition_trace.csv b/segmentation_1_4_0_fp32_combined/graph_partition_trace.csv new file mode 100644 index 0000000000000000000000000000000000000000..43a1468bd66b35bbf4ff132a73d5392dbb8e9b0d --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/graph_partition_trace.csv @@ -0,0 +1,318 @@ +Node, Type, Subgraph/CustomOp, Status +Add_105,,vaiml_par_0,Supported +Add_115,,vaiml_par_0,Supported +Add_124,,vaiml_par_0,Supported +Add_132,,vaiml_par_0,Supported +Add_135,,vaiml_par_0,Supported +Add_144,,vaiml_par_0,Supported +Add_152,,vaiml_par_0,Supported +Add_155,,vaiml_par_0,Supported +Add_164,,vaiml_par_0,Supported +Add_172,,vaiml_par_0,Supported +Add_175,,vaiml_par_0,Supported +Add_184,,vaiml_par_0,Supported +Add_19,,vaiml_par_0,Supported +Add_196,,vaiml_par_0,Supported +Add_206,,vaiml_par_0,Supported +Add_215,,vaiml_par_0,Supported +Add_227,,vaiml_par_0,Supported +Add_235,,vaiml_par_0,Supported +Add_238,,vaiml_par_0,Supported +Add_247,,vaiml_par_0,Supported +Add_259,,vaiml_par_0,Supported +Add_269,,vaiml_par_0,Supported +Add_278,,vaiml_par_0,Supported +Add_29,,vaiml_par_0,Supported +Add_290,,vaiml_par_0,Supported +Add_298,,vaiml_par_0,Supported +Add_301,,vaiml_par_0,Supported +Add_310,,vaiml_par_0,Supported +Add_322,,vaiml_par_0,Supported +Add_330,,vaiml_par_0,Supported +Add_333,,vaiml_par_0,Supported +Add_362,,vaiml_par_0,Supported +Add_388,,vaiml_par_0,Supported +Add_40,,vaiml_par_0,Supported +Add_414,,vaiml_par_0,Supported +Add_434,,vaiml_par_0,Supported +Add_445,,vaiml_par_0,Supported +Add_50,,vaiml_par_0,Supported +Add_67,,vaiml_par_0,Supported +Add_75,,vaiml_par_0,Supported +Add_85,,vaiml_par_0,Supported +Add_93,,vaiml_par_0,Supported +Add_96,,vaiml_par_0,Supported +AveragePool_346,,vaiml_par_0,Supported +AveragePool_347,,vaiml_par_0,Supported +AveragePool_348,,vaiml_par_0,Supported +Cast_0,,,Not supported. Check aie_unsupported_original_ops.json +Clip_108,,vaiml_par_0,Supported +Clip_118,,vaiml_par_0,Supported +Clip_127,,vaiml_par_0,Supported +Clip_138,,vaiml_par_0,Supported +Clip_147,,vaiml_par_0,Supported +Clip_158,,vaiml_par_0,Supported +Clip_167,,vaiml_par_0,Supported +Clip_178,,vaiml_par_0,Supported +Clip_187,,vaiml_par_0,Supported +Clip_199,,vaiml_par_0,Supported +Clip_209,,vaiml_par_0,Supported +Clip_218,,vaiml_par_0,Supported +Clip_22,,vaiml_par_0,Supported +Clip_230,,vaiml_par_0,Supported +Clip_241,,vaiml_par_0,Supported +Clip_250,,vaiml_par_0,Supported +Clip_262,,vaiml_par_0,Supported +Clip_272,,vaiml_par_0,Supported +Clip_281,,vaiml_par_0,Supported +Clip_293,,vaiml_par_0,Supported +Clip_304,,vaiml_par_0,Supported +Clip_313,,vaiml_par_0,Supported +Clip_325,,vaiml_par_0,Supported +Clip_336,,vaiml_par_0,Supported +Clip_446,,vaiml_par_0,Supported +Clip_447,,vaiml_par_0,Supported +Clip_53,,vaiml_par_0,Supported +Clip_70,,vaiml_par_0,Supported +Clip_88,,vaiml_par_0,Supported +Clip_99,,vaiml_par_0,Supported +Concat_350,,vaiml_par_0,Supported +Concat_355,,vaiml_par_0,Supported +Concat_363,,vaiml_par_0,Supported +Concat_372,,vaiml_par_0,Supported +Concat_376,,vaiml_par_0,Supported +Concat_381,,vaiml_par_0,Supported +Concat_389,,vaiml_par_0,Supported +Concat_398,,vaiml_par_0,Supported +Concat_402,,vaiml_par_0,Supported +Concat_407,,vaiml_par_0,Supported +Concat_415,,vaiml_par_0,Supported +Concat_418,,vaiml_par_0,Supported +Concat_422,,vaiml_par_0,Supported +Concat_427,,vaiml_par_0,Supported +Concat_435,,vaiml_par_0,Supported +Concat_438,,vaiml_par_0,Supported +Conv_103,,vaiml_par_0,Supported +Conv_112,,vaiml_par_0,Supported +Conv_113,,vaiml_par_0,Supported +Conv_122,,vaiml_par_0,Supported +Conv_131,,vaiml_par_0,Supported +Conv_133,,vaiml_par_0,Supported +Conv_142,,vaiml_par_0,Supported +Conv_151,,vaiml_par_0,Supported +Conv_153,,vaiml_par_0,Supported +Conv_162,,vaiml_par_0,Supported +Conv_17,,vaiml_par_0,Supported +Conv_171,,vaiml_par_0,Supported +Conv_173,,vaiml_par_0,Supported +Conv_182,,vaiml_par_0,Supported +Conv_192,,vaiml_par_0,Supported +Conv_194,,vaiml_par_0,Supported +Conv_203,,vaiml_par_0,Supported +Conv_204,,vaiml_par_0,Supported +Conv_213,,vaiml_par_0,Supported +Conv_223,,vaiml_par_0,Supported +Conv_225,,vaiml_par_0,Supported +Conv_234,,vaiml_par_0,Supported +Conv_236,,vaiml_par_0,Supported +Conv_245,,vaiml_par_0,Supported +Conv_255,,vaiml_par_0,Supported +Conv_257,,vaiml_par_0,Supported +Conv_26,,vaiml_par_0,Supported +Conv_266,,vaiml_par_0,Supported +Conv_267,,vaiml_par_0,Supported +Conv_276,,vaiml_par_0,Supported +Conv_28,,vaiml_par_0,Supported +Conv_286,,vaiml_par_0,Supported +Conv_288,,vaiml_par_0,Supported +Conv_297,,vaiml_par_0,Supported +Conv_299,,vaiml_par_0,Supported +Conv_30,,vaiml_par_0,Supported +Conv_308,,vaiml_par_0,Supported +Conv_318,,vaiml_par_0,Supported +Conv_32,,vaiml_par_0,Supported +Conv_320,,vaiml_par_0,Supported +Conv_329,,vaiml_par_0,Supported +Conv_331,,vaiml_par_0,Supported +Conv_34,,vaiml_par_0,Supported +Conv_340,,vaiml_par_0,Supported +Conv_343,,vaiml_par_0,Supported +Conv_35,,vaiml_par_0,Supported +Conv_351,,vaiml_par_0,Supported +Conv_356,,vaiml_par_0,Supported +Conv_37,,vaiml_par_0,Supported +Conv_373,,vaiml_par_0,Supported +Conv_377,,vaiml_par_0,Supported +Conv_382,,vaiml_par_0,Supported +Conv_39,,vaiml_par_0,Supported +Conv_399,,vaiml_par_0,Supported +Conv_403,,vaiml_par_0,Supported +Conv_408,,vaiml_par_0,Supported +Conv_41,,vaiml_par_0,Supported +Conv_419,,vaiml_par_0,Supported +Conv_423,,vaiml_par_0,Supported +Conv_428,,vaiml_par_0,Supported +Conv_43,,vaiml_par_0,Supported +Conv_439,,vaiml_par_0,Supported +Conv_441,,vaiml_par_0,Supported +Conv_443,,vaiml_par_0,Supported +Conv_46,,vaiml_par_0,Supported +Conv_48,,vaiml_par_0,Supported +Conv_57,,vaiml_par_0,Supported +Conv_58,,vaiml_par_0,Supported +Conv_60,,vaiml_par_0,Supported +Conv_63,,vaiml_par_0,Supported +Conv_65,,vaiml_par_0,Supported +Conv_74,,vaiml_par_0,Supported +Conv_76,,vaiml_par_0,Supported +Conv_78,,vaiml_par_0,Supported +Conv_81,,vaiml_par_0,Supported +Conv_83,,vaiml_par_0,Supported +Conv_92,,vaiml_par_0,Supported +Conv_94,,vaiml_par_0,Supported +Div_101,,vaiml_par_0,Supported +Div_110,,vaiml_par_0,Supported +Div_120,,vaiml_par_0,Supported +Div_129,,vaiml_par_0,Supported +Div_140,,vaiml_par_0,Supported +Div_149,,vaiml_par_0,Supported +Div_16,,vaiml_par_0,Supported +Div_160,,vaiml_par_0,Supported +Div_169,,vaiml_par_0,Supported +Div_180,,vaiml_par_0,Supported +Div_189,,vaiml_par_0,Supported +Div_2,,vaiml_par_0,Supported +Div_201,,vaiml_par_0,Supported +Div_211,,vaiml_par_0,Supported +Div_220,,vaiml_par_0,Supported +Div_232,,vaiml_par_0,Supported +Div_24,,vaiml_par_0,Supported +Div_243,,vaiml_par_0,Supported +Div_252,,vaiml_par_0,Supported +Div_264,,vaiml_par_0,Supported +Div_274,,vaiml_par_0,Supported +Div_283,,vaiml_par_0,Supported +Div_295,,vaiml_par_0,Supported +Div_306,,vaiml_par_0,Supported +Div_315,,vaiml_par_0,Supported +Div_327,,vaiml_par_0,Supported +Div_338,,vaiml_par_0,Supported +Div_55,,vaiml_par_0,Supported +Div_72,,vaiml_par_0,Supported +Div_90,,vaiml_par_0,Supported +GlobalAveragePool_191,,vaiml_par_0,Supported +GlobalAveragePool_222,,vaiml_par_0,Supported +GlobalAveragePool_254,,vaiml_par_0,Supported +GlobalAveragePool_285,,vaiml_par_0,Supported +GlobalAveragePool_317,,vaiml_par_0,Supported +GlobalAveragePool_342,,vaiml_par_0,Supported +GlobalAveragePool_45,,vaiml_par_0,Supported +GlobalAveragePool_62,,vaiml_par_0,Supported +GlobalAveragePool_80,,vaiml_par_0,Supported +Mul_102,,vaiml_par_0,Supported +Mul_111,,vaiml_par_0,Supported +Mul_121,,vaiml_par_0,Supported +Mul_130,,vaiml_par_0,Supported +Mul_141,,vaiml_par_0,Supported +Mul_150,,vaiml_par_0,Supported +Mul_161,,vaiml_par_0,Supported +Mul_170,,vaiml_par_0,Supported +Mul_181,,vaiml_par_0,Supported +Mul_190,,vaiml_par_0,Supported +Mul_202,,vaiml_par_0,Supported +Mul_212,,vaiml_par_0,Supported +Mul_221,,vaiml_par_0,Supported +Mul_233,,vaiml_par_0,Supported +Mul_244,,vaiml_par_0,Supported +Mul_25,,vaiml_par_0,Supported +Mul_253,,vaiml_par_0,Supported +Mul_265,,vaiml_par_0,Supported +Mul_275,,vaiml_par_0,Supported +Mul_284,,vaiml_par_0,Supported +Mul_296,,vaiml_par_0,Supported +Mul_307,,vaiml_par_0,Supported +Mul_316,,vaiml_par_0,Supported +Mul_328,,vaiml_par_0,Supported +Mul_339,,vaiml_par_0,Supported +Mul_345,,vaiml_par_0,Supported +Mul_354,,vaiml_par_0,Supported +Mul_360,,vaiml_par_0,Supported +Mul_361,,vaiml_par_0,Supported +Mul_380,,vaiml_par_0,Supported +Mul_386,,vaiml_par_0,Supported +Mul_387,,vaiml_par_0,Supported +Mul_406,,vaiml_par_0,Supported +Mul_412,,vaiml_par_0,Supported +Mul_413,,vaiml_par_0,Supported +Mul_426,,vaiml_par_0,Supported +Mul_432,,vaiml_par_0,Supported +Mul_433,,vaiml_par_0,Supported +Mul_56,,vaiml_par_0,Supported +Mul_73,,vaiml_par_0,Supported +Mul_91,,vaiml_par_0,Supported +Relu_193,,vaiml_par_0,Supported +Relu_224,,vaiml_par_0,Supported +Relu_256,,vaiml_par_0,Supported +Relu_27,,vaiml_par_0,Supported +Relu_287,,vaiml_par_0,Supported +Relu_31,,vaiml_par_0,Supported +Relu_319,,vaiml_par_0,Supported +Relu_33,,vaiml_par_0,Supported +Relu_341,,vaiml_par_0,Supported +Relu_36,,vaiml_par_0,Supported +Relu_374,,vaiml_par_0,Supported +Relu_38,,vaiml_par_0,Supported +Relu_400,,vaiml_par_0,Supported +Relu_42,,vaiml_par_0,Supported +Relu_420,,vaiml_par_0,Supported +Relu_44,,vaiml_par_0,Supported +Relu_440,,vaiml_par_0,Supported +Relu_442,,vaiml_par_0,Supported +Relu_47,,vaiml_par_0,Supported +Relu_59,,vaiml_par_0,Supported +Relu_61,,vaiml_par_0,Supported +Relu_64,,vaiml_par_0,Supported +Relu_77,,vaiml_par_0,Supported +Relu_79,,vaiml_par_0,Supported +Relu_82,,vaiml_par_0,Supported +Resize_365,,vaiml_par_0,Supported +Resize_391,,vaiml_par_0,Supported +Resize_417,,vaiml_par_0,Supported +Resize_437,,vaiml_par_0,Supported +Sigmoid_344,,vaiml_par_0,Supported +Sigmoid_352,,vaiml_par_0,Supported +Sigmoid_378,,vaiml_par_0,Supported +Sigmoid_404,,vaiml_par_0,Supported +Sigmoid_424,,vaiml_par_0,Supported +Slice_371,,vaiml_par_0,Supported +Slice_397,,vaiml_par_0,Supported +Slice_7,,vaiml_par_0,Supported +Split_349,,vaiml_par_0,Supported +Split_353,,vaiml_par_0,Supported +Split_375,,vaiml_par_0,Supported +Split_379,,vaiml_par_0,Supported +Split_401,,vaiml_par_0,Supported +Split_405,,vaiml_par_0,Supported +Split_421,,vaiml_par_0,Supported +Split_425,,vaiml_par_0,Supported +Split_444,,vaiml_par_0,Supported +Sub_14,,vaiml_par_0,Supported +Sub_359,,vaiml_par_0,Supported +Sub_385,,vaiml_par_0,Supported +Sub_411,,vaiml_par_0,Supported +Sub_431,,vaiml_par_0,Supported +Tanh_357,,vaiml_par_0,Supported +Tanh_383,,vaiml_par_0,Supported +Tanh_409,,vaiml_par_0,Supported +Tanh_429,,vaiml_par_0,Supported +Transpose_10,,,Not supported. Check aie_unsupported_original_ops.json +Transpose_11,,,Not supported. Check aie_unsupported_original_ops.json +Transpose_12,,,Not supported. Check aie_unsupported_original_ops.json +Transpose_448,,,Not supported. Check aie_unsupported_original_ops.json +Transpose_449,,,Not supported. Check aie_unsupported_original_ops.json +Transpose_450,,,Not supported. Check aie_unsupported_original_ops.json +Transpose_451,,,Not supported. Check aie_unsupported_original_ops.json +Transpose_452,,,Not supported. Check aie_unsupported_original_ops.json +Transpose_453,,,Not supported. Check aie_unsupported_original_ops.json +Transpose_8,,vaiml_par_0,Supported +Transpose_9,,,Not supported. Check aie_unsupported_original_ops.json diff --git a/segmentation_1_4_0_fp32_combined/original-info-signature.txt b/segmentation_1_4_0_fp32_combined/original-info-signature.txt new file mode 100644 index 0000000000000000000000000000000000000000..4262119974eb805ab03f9c534be3592fbbdd1a54 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/original-info-signature.txt @@ -0,0 +1 @@ +4f00fb244983f7c2158dc9333522f122 \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/original-model-signature.txt b/segmentation_1_4_0_fp32_combined/original-model-signature.txt new file mode 100644 index 0000000000000000000000000000000000000000..c662815d6b6f6f7d554bf95938c117b912e549a3 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/original-model-signature.txt @@ -0,0 +1 @@ +6bbb891ab96ca9362e0e61024cd02778 \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/preliminary-vaiml-pass-summary.txt b/segmentation_1_4_0_fp32_combined/preliminary-vaiml-pass-summary.txt new file mode 100644 index 0000000000000000000000000000000000000000..88a0917d082bd1d41d41eeea6212c70c5f742ce9 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/preliminary-vaiml-pass-summary.txt @@ -0,0 +1,14 @@ +----- Preliminary Summary of VAIML Pass ------ +Model data type: float32 +Device data type: bfloat16 +Number of operators in the model: 317 +GOPs of the model: 3.73517 +Number of operators supported by VAIML: 306(96.530%) +GOPs supported by VAIML: 3.735 (100.000%) +Number of subgraphs supported by VAIML: 1 + +Number of identified subgraphs: 1 +Stats for identified subgraphs +vaiml_par_0 stats: + Operators: 306 (96.530%) + GOPs: 3.735(100.000%) OPs: 3,735,173,696 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/AIECompiler.log b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/AIECompiler.log new file mode 100644 index 0000000000000000000000000000000000000000..60c6cf1c23762c86c68d95f748169065cacc32b6 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/AIECompiler.log @@ -0,0 +1,1759 @@ +INFO: [aiecompiler 77-297] Cmd Line : /usr/local/lib/python3.10/dist-packages/bin/unwrapped/lnx64.o/aiecompiler /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/top.cpp --part=xc10AIE2P_ML-die-0x-e-S-es1 --nodot-graph --runtime-opt=1 --disable-multirate-analysis --enable-core-processor-bus --enable-multi-layer --heapsize=1792 --stacksize=1400 --max-layer-ctrl-param-size=256 --compile-for-aiesim=false --workdir=Work --multi-layer-ctrl-pkt --aie2ipu-base-addr=0 -enable-light-cdo --Xelfgen=-j4 --multi-layer-pipelining --multi-layer-opt=3 --Xpreproc=-D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ --multi-layer-ext-buf-file=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/flexmlrt-hsi.json --enable-partition=0:4 --multi-layer-ctrl-pkt-column-span=4 --multi-layer-prebuilt-archive=/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras/data/ryzen-ai/stx/unified-overlay-4x4.json --multi-layer-prebuilt-archive-enable-elf-gen --multi-layer-init-core-elf-ctrl-pkt --multi-layer-pm-id 29006 --Xpreproc=-DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 --include=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend --include=/usr/local/lib/python3.10/site-packages/include/aie_api --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf --output-archive libadf.a --adf-api-log-level=0 --multi-layer-pm-reloading=1 --Xelfgen=-j1 +INFO: [aiecompiler 77-404] Executing Cmd: /usr/local/lib/python3.10/dist-packages/bin/../lnx64.o/tools/clang/bin/clang++ -I/usr/local/lib/python3.10/dist-packages/include -I/usr/local/lib/python3.10/dist-packages/tps/lnx64/gcc/include/c++/8.3.0 -I/usr/local/lib/python3.10/dist-packages/tps/lnx64/gcc/include/c++/8.3.0/x86_64-pc-linux-gnu -I/usr/local/lib/python3.10/dist-packages/tps/lnx64/gcc/include -Wno-error=reserved-user-defined-literal -E -std=c++17 -D__ADF_FRONTEND__ -D__AIE_ARCH__=21 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -I/usr/local/lib/python3.10/dist-packages/include -I . -I /app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/site-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/top.cpp > Work/temp/top.ii +INFO: [aiecompiler 77-404] Executing Cmd: graph_preprocessor Work/temp/top.ii -o Work/temp/top.processed.ii -report-core-dump -- -std=c++17 -ftemplate-depth=2048 -Wno-return-stack-address -Wno-missing-declarations -Wno-parentheses-equality -Wno-shift-negative-value +INFO: [aiecompiler 77-404] Executing Cmd: /usr/local/lib/python3.10/dist-packages/bin/../lnx64.o/tools/clang/bin/clang++ --gcc-install-dir=/usr/local/lib/python3.10/dist-packages/tps/lnx64/gcc/lib/gcc/x86_64-pc-linux-gnu/8.3.0 -stdlib=libstdc++ -Wno-error=reserved-user-defined-literal -std=c++17 -I . Work/temp/top.processed.ii -o Work/temp/top.out -L /usr/local/lib/python3.10/dist-packages/lib/lnx64.o -g -O0 -Wl,--unresolved-symbols=ignore-all -Wno-return-stack-address -Wno-missing-declarations -lmeir_frontend -ladf_api_frontend +INFO: [aiecompiler 77-404] Executing Cmd: Work/temp/top.out -I /app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/site-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf --workdir=Work --part=xc10AIE2P_ML-die-0x-e-S-es1 --aiearch=aie2p --log-level=1 --pl-axi-lite=0 --enable-multi-layer=1 --multi-layer-opt=3 --disable-dma-autostart=0 --enable-light-cdo=1 --large-program-memory=0 --swfifo-threshold=40 --enable-dma-fifo=1 --target=hw +INFO: [aiecompiler 77-749] Reading logical device aie2p_8x4_device.json +INFO: [aiecompiler 77-6447] Executing Cmd: aieir_be --time-passes=0 --Xelfgen="-j4" --Xelfgen="-j1" --heapsize=1792 --Xpreproc="-D__IO_BUFFER_FORCE_LIGHT_WEIGHT__" --Xpreproc="-DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1" --target=hw --include="/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend" --include="/usr/local/lib/python3.10/site-packages/include/aie_api" --include="/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common" --include="/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common" --include="/usr/local/lib/python3.10/dist-packages/vitis_mllib" --include="/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc" --include="/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf" --stacksize=1400 --enable-partition=0:4 --nodot-graph=true --adf-api-log-level=0 --event-trace-port=gmio --part=xc10AIE2P_ML-die-0x-e-S-es1 --json="Work/temp/top.json" +INFO: [aiecompiler 77-749] Reading logical device aie2p_8x4_device.json +WARNING: [aiecompiler 77-22869] Forcing --gen-elf-ctrl-pkts=true when --multi-layer-pm-reload-elf-ctrl-pkt is true. +WARNING: [aiecompiler 77-5902] Under --enable-multi-layer, disablefloorplanning is added to --Xmapper option. +WARNING: [aiecompiler 77-21917] When --enable-multi-layer=true, broadcast_enable_core is forced to false +WARNING: [aiecompiler 77-5758] When --enableMultiLayer=true, disable-dma-autostart is forced to true +WARNING: [aiecompiler 77-21918] When --enableMultiLayer=true, enable-core-processor-bus is forced to true +WARNING: [aiecompiler 77-6502] aiecompiler option broadcast_enable_core=true is not supported for part xc10AIE2P_ML-die-0x-e-S-es1; forcing it to false. +WARNING: [aiecompiler 77-23075] Forcing --multi-layer-ctrl-pkt-column-span=0 when --multi-layer-prebuilt-archive is specified. +WARNING: [aiecompiler 77-22616] Part xc10AIE2P_ML-die-0x-e-S-es1 does not support ECC Scrubbing; ECC Scrubbing is disabled. +INFO: [aiecompiler 77-757] Opening input file: Work/temp/top.json +INFO: [aiecompiler 77-6287] Emitting AIEIr in file: Work/temp/top_aieir_dump.txt +INFO: [aiecompiler 77-656] Processing Graph 'root' +INFO: [aiecompiler 77-21798] ### Entering UnPacker pass +INFO: [aiecompiler 77-21797] ### Done with UnPacker pass (72392752 secs) +INFO: [aiecompiler 77-5563] Reading multi-layer graph constraints from 'Work/temp/top_meir_constraints.json' +INFO: [aiecompiler 77-5545] Generating overlay graph information at 'Work/temp/top_folded.json' +INFO: [aiecompiler 77-22451] ### Entering External Buffer Coalesing Pass +INFO: [aiecompiler 77-22449] ### Applying external buffer coalesing file /app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/flexmlrt-hsi.json +INFO: [aiecompiler 77-23091] Parsing MeIR graph to create runtime_buffer_map and qualified_to_parent_name_map +INFO: [aiecompiler 77-23001] Buffer Coalescing Actions +runtime_buffers_map entries +Name: coalesed_spills + xrt_id: 2 + size_in_bytes: 69157168 + extra_padding_before: 0 + extra_padding_before: 0 +Name: coalesed_weights + xrt_id: 0 + size_in_bytes: 11809920 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ifm_ddr + xrt_id: 1 + size_in_bytes: 471040 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ifm_ddr_1 + xrt_id: 10 + size_in_bytes: 460800 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ifm_ddr_2 + xrt_id: 8 + size_in_bytes: 172800 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ifm_ddr_3 + xrt_id: 6 + size_in_bytes: 73600 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ifm_ddr_4 + xrt_id: 4 + size_in_bytes: 30720 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_0_l2l3_229_spill + xrt_id: 5 + size_in_bytes: 30720 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_1_l2l3_252_spill + xrt_id: 7 + size_in_bytes: 73600 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_2_l2l3_272_spill + xrt_id: 9 + size_in_bytes: 172800 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_3_l2l3_291_spill + xrt_id: 11 + size_in_bytes: 460800 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_4 + xrt_id: 12 + size_in_bytes: 921600 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_5 + xrt_id: 13 + size_in_bytes: 921600 + extra_padding_before: 0 + extra_padding_before: 0 +Name: runtime_control_packet + xrt_id: 3 + size_in_bytes: 0 + extra_padding_before: 0 + extra_padding_before: 0 + +INFO: [aiecompiler 77-22450] ### Done with External Buffer Coalesing Pass +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_8_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_8_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_13_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_13_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_14_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_14_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_15_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_15_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_16_l2_wts is 6. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_16_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_18_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_18_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_19_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_19_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_20_l2_wts is 2. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_20_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_22_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_22_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_23_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_23_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_24_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_24_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_27_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_27_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_28_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_28_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_34_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_34_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_35_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_35_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_36_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_36_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_39_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_39_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_40_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_40_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_46_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_46_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_47_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_47_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_48_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_48_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_51_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_51_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_52_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_52_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_58_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_58_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_59_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_59_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_65_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_65_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_70_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_70_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_71_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_71_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_76_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_76_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_81_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_81_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_82_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_82_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_87_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_87_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_92_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_92_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_93_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_93_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_98_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_98_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_103_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_103_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_104_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_104_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_109_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_109_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_116_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_116_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_117_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_117_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_123_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_123_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_124_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_124_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_129_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_129_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_136_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_136_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_137_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_137_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_143_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_143_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_144_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_144_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_149_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_149_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_156_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_156_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_157_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_157_wts_ddr is 1. +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-5917' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-404] Executing Cmd: cd Work/aie/ir + /usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +Wllvm,-O2,-fno-jump-tables,-femit-all-decls,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include empty.cc -o _header.ll + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i719_wrap_transpose4d_adf_wrapper.makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i802_wrap_resize_adf_wrapper.makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i852_wrap_slice_adf_wrapper.makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i897_wrap_concat_adf_wrapper.makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i1009_wrap_resize_adf_wrapper.makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i1100_superkernels.makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: rm -f Work/aie/ir/i719_wrap_transpose4d_adf_wrapper.makefile Work/aie/ir/i802_wrap_resize_adf_wrapper.makefile Work/aie/ir/i852_wrap_slice_adf_wrapper.makefile Work/aie/ir/i897_wrap_concat_adf_wrapper.makefile Work/aie/ir/i1009_wrap_resize_adf_wrapper.makefile Work/aie/ir/i1100_superkernels.makefile; + +INFO: [aiecompiler 77-404] Executing Cmd: cd Work/aie/ir + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i719_wrap_transpose4d_adf_wrapper_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i719_wrap_transpose4d_adf_wrapper.ll -spec=i719_wrap_transpose4d_adf_wrapper_spec.json > /dev/null + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i802_wrap_resize_adf_wrapper_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i802_wrap_resize_adf_wrapper.ll -spec=i802_wrap_resize_adf_wrapper_spec.json > /dev/null + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i852_wrap_slice_adf_wrapper_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i852_wrap_slice_adf_wrapper.ll -spec=i852_wrap_slice_adf_wrapper_spec.json > /dev/null + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i897_wrap_concat_adf_wrapper_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i897_wrap_concat_adf_wrapper.ll -spec=i897_wrap_concat_adf_wrapper_spec.json > /dev/null + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i1009_wrap_resize_adf_wrapper_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i1009_wrap_resize_adf_wrapper.ll -spec=i1009_wrap_resize_adf_wrapper_spec.json > /dev/null + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i1100_superkernels_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i1100_superkernels.ll -spec=i1100_superkernels_spec.json > /dev/null +INFO: [aiecompiler 77-404] Executing Cmd: cd Work/aie/ir + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i719_wrap_transpose4d_adf_wrapper_spec.json i719_wrap_transpose4d_adf_wrapper.ll -o i719_wrap_transpose4d_adf_wrapper.ll 2>/dev/null + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i802_wrap_resize_adf_wrapper_spec.json i802_wrap_resize_adf_wrapper.ll -o i802_wrap_resize_adf_wrapper.ll 2>/dev/null + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i852_wrap_slice_adf_wrapper_spec.json i852_wrap_slice_adf_wrapper.ll -o i852_wrap_slice_adf_wrapper.ll 2>/dev/null + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i897_wrap_concat_adf_wrapper_spec.json i897_wrap_concat_adf_wrapper.ll -o i897_wrap_concat_adf_wrapper.ll 2>/dev/null + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i1009_wrap_resize_adf_wrapper_spec.json i1009_wrap_resize_adf_wrapper.ll -o i1009_wrap_resize_adf_wrapper.ll 2>/dev/null + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i1100_superkernels_spec.json i1100_superkernels.ll -o i1100_superkernels.ll 2>/dev/null +WARNING: [aiecompiler 77-4232] All nodes in partition PT0[0.01]:[i4] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT0[0.01]:[i4] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT1[0.01]:[i5] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT1[0.01]:[i5] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT2[0.01]:[i6] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT2[0.01]:[i6] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT3[0.01]:[i7] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT3[0.01]:[i7] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT4[0.01]:[i8] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT4[0.01]:[i8] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT5[0.01]:[i9] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT5[0.01]:[i9] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT6[0.01]:[i10] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT6[0.01]:[i10] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT7[0.01]:[i11] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT7[0.01]:[i11] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT8[0.01]:[i12] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT8[0.01]:[i12] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT9[0.01]:[i13] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT9[0.01]:[i13] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT10[0.01]:[i14] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT10[0.01]:[i14] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT11[0.01]:[i15] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT11[0.01]:[i15] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT12[0.01]:[i16] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT12[0.01]:[i16] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT13[0.01]:[i17] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT13[0.01]:[i17] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT14[0.01]:[i18] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT14[0.01]:[i18] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT15[0.01]:[i19] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT15[0.01]:[i19] have heapsize constraint. Using the sum as the partition's heapsize +INFO: [aiecompiler 77-281] ###Writing Partition Data To JSON File Work/temp/top_partition.json +INFO: [aiecompiler 77-6291] Entering MAPPING ANALYSIS pass +INFO: [aiecompiler 77-6284] Done with MAPPING ANALYSIS pass +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/pm_reload_analysis0/timestamped_log +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f pm_reload_analysis0.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f pm_reload_analysis0.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-6292] Entering Scheduler pass +INFO: [aiecompiler 77-22516] Done with mergeKernelLayerNodes() for cc0 +INFO: [aiecompiler 77-22516] Done with SchedulerUtils::identifyOfmDDRSpillingScheduleWithKernelLayer() for cc0 +INFO: [aiecompiler 77-22516] Done with SchedulerUtils::identifyBufferToBufferScheduleWithKernelLayer() for cc0 +INFO: [#UNDEF] Mask used: 0x0000ffff, Adjusted PM Id: 0x714e0000 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1012): mode(0), layer(0): {compute_graph.ifm_ddr.out[0], compute_graph.ifm_ddr.out[1], compute_graph.L2_IFM_Buffer_for_input0_0_port0.in[0], compute_graph.L2_IFM_Buffer_for_input0_0_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1328), layer(0): compute_graph.flexml_layers[0].compute_node[0][0], compute_graph.flexml_layers[0].compute_node[0][1], compute_graph.flexml_layers[0].compute_node[0][2], compute_graph.flexml_layers[0].compute_node[0][3], compute_graph.flexml_layers[0].compute_node[1][0], compute_graph.flexml_layers[0].compute_node[1][1], compute_graph.flexml_layers[0].compute_node[1][2], compute_graph.flexml_layers[0].compute_node[1][3], compute_graph.flexml_layers[0].compute_node[2][0], compute_graph.flexml_layers[0].compute_node[2][1], compute_graph.flexml_layers[0].compute_node[2][2], compute_graph.flexml_layers[0].compute_node[2][3], compute_graph.flexml_layers[0].compute_node[3][0], compute_graph.flexml_layers[0].compute_node[3][1], compute_graph.flexml_layers[0].compute_node[3][2], compute_graph.flexml_layers[0].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[0], compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[1], compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[2], compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[3], compute_graph.l2_1.in[0], compute_graph.l2_1.in[1], compute_graph.l2_1.in[2], compute_graph.l2_1.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(897): mode(0), layer(0): {compute_graph.l2_1.out[0], compute_graph.l2_1.out[1], compute_graph.l2l3_1_spill.in[0], compute_graph.l2l3_1_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1013): mode(0), layer(1): {compute_graph.l2l3_1_spill.out[0], compute_graph.l2l3_1_spill.out[1], compute_graph.templated_graph_2.ifm.in[0], compute_graph.templated_graph_2.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1225): mode(0), layer(1): {compute_graph.templated_graph_2.ifm.out[0], compute_graph.templated_graph_2.ifm.out[1], compute_graph.l2l3_2_spill.in[0], compute_graph.l2l3_2_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1014): mode(0), layer(2): {compute_graph.l2l3_2_spill.out[0], compute_graph.templated_graph_3.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1226): mode(0), layer(2): {compute_graph.templated_graph_3.ifm.out[0], compute_graph.templated_graph_3.ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1227): mode(0), layer(2): {compute_graph.templated_graph_3.ofm.out[0], compute_graph.l2l3_scratch_0_3_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1015): mode(0), layer(3): {compute_graph.l2l3_scratch_0_3_spill.out[0], compute_graph.templated_graph_3.ifm2.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1228): mode(0), layer(3): {compute_graph.templated_graph_3.ifm2.out[0], compute_graph.l2l3_3_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1016): mode(0), layer(4): {compute_graph.l2l3_3_spill.out[0], compute_graph.templated_graph_4.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1230), layer(4): compute_graph.templated_graph_4.trans_comp_nd[0][0], compute_graph.templated_graph_4.trans_comp_nd[0][1], compute_graph.templated_graph_4.trans_comp_nd[0][2], compute_graph.templated_graph_4.trans_comp_nd[0][3], compute_graph.templated_graph_4.trans_comp_nd[1][0], compute_graph.templated_graph_4.trans_comp_nd[1][1], compute_graph.templated_graph_4.trans_comp_nd[1][2], compute_graph.templated_graph_4.trans_comp_nd[1][3], compute_graph.templated_graph_4.trans_comp_nd[2][0], compute_graph.templated_graph_4.trans_comp_nd[2][1], compute_graph.templated_graph_4.trans_comp_nd[2][2], compute_graph.templated_graph_4.trans_comp_nd[2][3], compute_graph.templated_graph_4.trans_comp_nd[3][0], compute_graph.templated_graph_4.trans_comp_nd[3][1], compute_graph.templated_graph_4.trans_comp_nd[3][2], compute_graph.templated_graph_4.trans_comp_nd[3][3], {compute_graph.templated_graph_4.trans_mt_ifm.out[0], compute_graph.templated_graph_4.trans_mt_ifm.out[1], compute_graph.templated_graph_4.trans_mt_ifm.out[2], compute_graph.templated_graph_4.trans_mt_ifm.out[3], compute_graph.templated_graph_4.trans_mt_ofm.in[0], compute_graph.templated_graph_4.trans_mt_ofm.in[1], compute_graph.templated_graph_4.trans_mt_ofm.in[2], compute_graph.templated_graph_4.trans_mt_ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1229): mode(3), layer(4): {compute_graph.templated_graph_4.trans_mt_ofm.out[0], compute_graph.l2l3_4_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1229) is pipelined with KernelLayerNode(1230) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1017): mode(0), layer(5): {compute_graph.l2l3_4_spill.out[0], compute_graph.templated_graph_5.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1231): mode(0), layer(5): {compute_graph.templated_graph_5.ifm.out[0], compute_graph.templated_graph_5.ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1232): mode(0), layer(5): {compute_graph.templated_graph_5.ofm.out[0], compute_graph.l2l3_scratch_0_5_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1018): mode(0), layer(6): {compute_graph.l2l3_scratch_0_5_spill.out[0], compute_graph.templated_graph_5.ifm2.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1233): mode(0), layer(6): {compute_graph.templated_graph_5.ifm2.out[0], compute_graph.l2l3_5_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 898 will not be pipelined because it's in the same layer 7 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1565), layer(7): {compute_graph.l2l3_5_spill.out[0], compute_graph.const_ifm_ddr_5.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1566), layer(7): {compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1567), layer(7): {compute_graph.l2l3_5_spill.out[1], compute_graph.const_ifm_ddr_5.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1568), layer(7): {compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1329), layer(7): compute_graph.flexml_layers[1].compute_node[0][0], compute_graph.flexml_layers[1].compute_node[0][1], compute_graph.flexml_layers[1].compute_node[0][2], compute_graph.flexml_layers[1].compute_node[0][3], compute_graph.flexml_layers[1].compute_node[1][0], compute_graph.flexml_layers[1].compute_node[1][1], compute_graph.flexml_layers[1].compute_node[1][2], compute_graph.flexml_layers[1].compute_node[1][3], compute_graph.flexml_layers[1].compute_node[2][0], compute_graph.flexml_layers[1].compute_node[2][1], compute_graph.flexml_layers[1].compute_node[2][2], compute_graph.flexml_layers[1].compute_node[2][3], compute_graph.flexml_layers[1].compute_node[3][0], compute_graph.flexml_layers[1].compute_node[3][1], compute_graph.flexml_layers[1].compute_node[3][2], compute_graph.flexml_layers[1].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[3], compute_graph.l2_6.in[0], compute_graph.l2_6.in[1], compute_graph.l2_6.in[2], compute_graph.l2_6.in[3]}, Scheduler computes number of sub-layer phases to be 9. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(898): mode(3), layer(7): {compute_graph.l2_6.out[0], compute_graph.l2_6.out[1], compute_graph.l2l3_6_spill.in[0], compute_graph.l2l3_6_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 899 will not be pipelined because it's in the same layer 8 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1569), layer(8): {compute_graph.l2l3_6_spill.out[0], compute_graph.const_ifm_ddr_4.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1570), layer(8): {compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1571), layer(8): {compute_graph.l2l3_6_spill.out[1], compute_graph.const_ifm_ddr_4.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1572), layer(8): {compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1330), layer(8): compute_graph.flexml_layers[2].compute_node[0][0], compute_graph.flexml_layers[2].compute_node[0][1], compute_graph.flexml_layers[2].compute_node[0][2], compute_graph.flexml_layers[2].compute_node[0][3], compute_graph.flexml_layers[2].compute_node[1][0], compute_graph.flexml_layers[2].compute_node[1][1], compute_graph.flexml_layers[2].compute_node[1][2], compute_graph.flexml_layers[2].compute_node[1][3], compute_graph.flexml_layers[2].compute_node[2][0], compute_graph.flexml_layers[2].compute_node[2][1], compute_graph.flexml_layers[2].compute_node[2][2], compute_graph.flexml_layers[2].compute_node[2][3], compute_graph.flexml_layers[2].compute_node[3][0], compute_graph.flexml_layers[2].compute_node[3][1], compute_graph.flexml_layers[2].compute_node[3][2], compute_graph.flexml_layers[2].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[3], compute_graph.l2_7.in[0], compute_graph.l2_7.in[1], compute_graph.l2_7.in[2], compute_graph.l2_7.in[3]}, Scheduler computes number of sub-layer phases to be 9. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(819): mode(3), layer(9): {compute_graph.Layer_8_wts_ddr.out[0], compute_graph.Layer_8_wts_ddr.out[1], compute_graph.Layer_8_l2_wts.in[0], compute_graph.Layer_8_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(819) is pipelined with KernelLayerNode(1330) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(899): mode(3), layer(8): {compute_graph.l2_7.out[0], compute_graph.l2_7.out[1], compute_graph.l2l3_7_spill.in[0], compute_graph.l2l3_7_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1026): mode(0), layer(9): {compute_graph.l2l3_7_spill.out[0], compute_graph.l2l3_7_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[3], compute_graph.l2_8.in[0], compute_graph.l2_8.in[1], compute_graph.l2_8.in[2], compute_graph.l2_8.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1331), layer(9): compute_graph.flexml_layers[3].compute_node[0][0], compute_graph.flexml_layers[3].compute_node[0][1], compute_graph.flexml_layers[3].compute_node[0][2], compute_graph.flexml_layers[3].compute_node[0][3], compute_graph.flexml_layers[3].compute_node[1][0], compute_graph.flexml_layers[3].compute_node[1][1], compute_graph.flexml_layers[3].compute_node[1][2], compute_graph.flexml_layers[3].compute_node[1][3], compute_graph.flexml_layers[3].compute_node[2][0], compute_graph.flexml_layers[3].compute_node[2][1], compute_graph.flexml_layers[3].compute_node[2][2], compute_graph.flexml_layers[3].compute_node[2][3], compute_graph.flexml_layers[3].compute_node[3][0], compute_graph.flexml_layers[3].compute_node[3][1], compute_graph.flexml_layers[3].compute_node[3][2], compute_graph.flexml_layers[3].compute_node[3][3], {compute_graph.Layer_8_l2_wts.out[0], compute_graph.Layer_8_l2_wts.out[1], compute_graph.Layer_8_l2_wts.out[2], compute_graph.Layer_8_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[3], compute_graph.l2_8.in[0], compute_graph.l2_8.in[1], compute_graph.l2_8.in[2], compute_graph.l2_8.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(900): mode(3), layer(9): {compute_graph.l2_8.out[0], compute_graph.l2_8.out[1], compute_graph.l2l3_8_spill.in[0], compute_graph.l2l3_8_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(900) is pipelined with KernelLayerNode(1331) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1027): mode(0), layer(10): {compute_graph.l2l3_8_spill.out[0], compute_graph.l2l3_8_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1332), layer(10): compute_graph.flexml_layers[4].compute_node[0][0], compute_graph.flexml_layers[4].compute_node[0][1], compute_graph.flexml_layers[4].compute_node[0][2], compute_graph.flexml_layers[4].compute_node[0][3], compute_graph.flexml_layers[4].compute_node[1][0], compute_graph.flexml_layers[4].compute_node[1][1], compute_graph.flexml_layers[4].compute_node[1][2], compute_graph.flexml_layers[4].compute_node[1][3], compute_graph.flexml_layers[4].compute_node[2][0], compute_graph.flexml_layers[4].compute_node[2][1], compute_graph.flexml_layers[4].compute_node[2][2], compute_graph.flexml_layers[4].compute_node[2][3], compute_graph.flexml_layers[4].compute_node[3][0], compute_graph.flexml_layers[4].compute_node[3][1], compute_graph.flexml_layers[4].compute_node[3][2], compute_graph.flexml_layers[4].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[3], compute_graph.l2_9.in[0], compute_graph.l2_9.in[1], compute_graph.l2_9.in[2], compute_graph.l2_9.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(901): mode(0), layer(10): {compute_graph.l2_9.out[0], compute_graph.l2_9.out[1], compute_graph.l2l3_9_spill.in[0], compute_graph.l2l3_9_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1029): mode(0), layer(11): {compute_graph.l2l3_9_spill.out[0], compute_graph.l2l3_9_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1333), layer(11): compute_graph.flexml_layers[5].compute_node[0][0], compute_graph.flexml_layers[5].compute_node[0][1], compute_graph.flexml_layers[5].compute_node[0][2], compute_graph.flexml_layers[5].compute_node[0][3], compute_graph.flexml_layers[5].compute_node[1][0], compute_graph.flexml_layers[5].compute_node[1][1], compute_graph.flexml_layers[5].compute_node[1][2], compute_graph.flexml_layers[5].compute_node[1][3], compute_graph.flexml_layers[5].compute_node[2][0], compute_graph.flexml_layers[5].compute_node[2][1], compute_graph.flexml_layers[5].compute_node[2][2], compute_graph.flexml_layers[5].compute_node[2][3], compute_graph.flexml_layers[5].compute_node[3][0], compute_graph.flexml_layers[5].compute_node[3][1], compute_graph.flexml_layers[5].compute_node[3][2], compute_graph.flexml_layers[5].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[3], compute_graph.l2_10.in[0], compute_graph.l2_10.in[1], compute_graph.l2_10.in[2], compute_graph.l2_10.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(902): mode(3), layer(11): {compute_graph.l2_10.out[0], compute_graph.l2_10.out[1], compute_graph.l2l3_10_spill.in[0], compute_graph.l2l3_10_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(902) is pipelined with KernelLayerNode(1333) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1030): mode(0), layer(12): {compute_graph.l2l3_10_spill.out[0], compute_graph.l2l3_10_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1334), layer(12): compute_graph.flexml_layers[6].compute_node[0][0], compute_graph.flexml_layers[6].compute_node[0][1], compute_graph.flexml_layers[6].compute_node[0][2], compute_graph.flexml_layers[6].compute_node[0][3], compute_graph.flexml_layers[6].compute_node[1][0], compute_graph.flexml_layers[6].compute_node[1][1], compute_graph.flexml_layers[6].compute_node[1][2], compute_graph.flexml_layers[6].compute_node[1][3], compute_graph.flexml_layers[6].compute_node[2][0], compute_graph.flexml_layers[6].compute_node[2][1], compute_graph.flexml_layers[6].compute_node[2][2], compute_graph.flexml_layers[6].compute_node[2][3], compute_graph.flexml_layers[6].compute_node[3][0], compute_graph.flexml_layers[6].compute_node[3][1], compute_graph.flexml_layers[6].compute_node[3][2], compute_graph.flexml_layers[6].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[3], compute_graph.l2_11.in[0], compute_graph.l2_11.in[1], compute_graph.l2_11.in[2], compute_graph.l2_11.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(903): mode(0), layer(12): {compute_graph.l2_11.out[0], compute_graph.l2_11.out[1], compute_graph.l2l3_11_spill.in[0], compute_graph.l2l3_11_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 904 will not be pipelined because it's in the same layer 13 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1573), layer(13): {compute_graph.l2l3_8_spill.out[2], compute_graph.l2l3_11_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1574), layer(13): {compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1575), layer(13): {compute_graph.l2l3_8_spill.out[3], compute_graph.l2l3_11_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1576), layer(13): {compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1335), layer(13): compute_graph.flexml_layers[7].compute_node[0][0], compute_graph.flexml_layers[7].compute_node[0][1], compute_graph.flexml_layers[7].compute_node[0][2], compute_graph.flexml_layers[7].compute_node[0][3], compute_graph.flexml_layers[7].compute_node[1][0], compute_graph.flexml_layers[7].compute_node[1][1], compute_graph.flexml_layers[7].compute_node[1][2], compute_graph.flexml_layers[7].compute_node[1][3], compute_graph.flexml_layers[7].compute_node[2][0], compute_graph.flexml_layers[7].compute_node[2][1], compute_graph.flexml_layers[7].compute_node[2][2], compute_graph.flexml_layers[7].compute_node[2][3], compute_graph.flexml_layers[7].compute_node[3][0], compute_graph.flexml_layers[7].compute_node[3][1], compute_graph.flexml_layers[7].compute_node[3][2], compute_graph.flexml_layers[7].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[3], compute_graph.l2_12.in[0], compute_graph.l2_12.in[1], compute_graph.l2_12.in[2], compute_graph.l2_12.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(820): mode(3), layer(14): {compute_graph.Layer_13_wts_ddr.out[0], compute_graph.Layer_13_wts_ddr.out[1], compute_graph.Layer_13_l2_wts.in[0], compute_graph.Layer_13_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(820) is pipelined with KernelLayerNode(1335) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(904): mode(3), layer(13): {compute_graph.l2_12.out[0], compute_graph.l2_12.out[1], compute_graph.l2l3_12_spill.in[0], compute_graph.l2l3_12_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1032): mode(0), layer(14): {compute_graph.l2l3_12_spill.out[0], compute_graph.l2l3_12_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1336), layer(14): compute_graph.flexml_layers[8].compute_node[0][0], compute_graph.flexml_layers[8].compute_node[0][1], compute_graph.flexml_layers[8].compute_node[0][2], compute_graph.flexml_layers[8].compute_node[0][3], compute_graph.flexml_layers[8].compute_node[1][0], compute_graph.flexml_layers[8].compute_node[1][1], compute_graph.flexml_layers[8].compute_node[1][2], compute_graph.flexml_layers[8].compute_node[1][3], compute_graph.flexml_layers[8].compute_node[2][0], compute_graph.flexml_layers[8].compute_node[2][1], compute_graph.flexml_layers[8].compute_node[2][2], compute_graph.flexml_layers[8].compute_node[2][3], compute_graph.flexml_layers[8].compute_node[3][0], compute_graph.flexml_layers[8].compute_node[3][1], compute_graph.flexml_layers[8].compute_node[3][2], compute_graph.flexml_layers[8].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[3], compute_graph.Layer_13_l2_wts.out[0], compute_graph.Layer_13_l2_wts.out[1], compute_graph.Layer_13_l2_wts.out[2], compute_graph.Layer_13_l2_wts.out[3], compute_graph.l2_13.in[0], compute_graph.l2_13.in[1], compute_graph.l2_13.in[2], compute_graph.l2_13.in[3]}, Scheduler computes number of sub-layer phases to be 8. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(821): mode(3), layer(15): {compute_graph.Layer_14_wts_ddr.out[0], compute_graph.Layer_14_wts_ddr.out[1], compute_graph.Layer_14_l2_wts.in[0], compute_graph.Layer_14_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(821) is pipelined with KernelLayerNode(1336) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(905): mode(0), layer(14): {compute_graph.l2_13.out[0], compute_graph.l2_13.out[1], compute_graph.l2l3_13_spill.in[0], compute_graph.l2l3_13_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1577), layer(15): {compute_graph.l2l3_12_spill.out[2], compute_graph.l2l3_13_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1578), layer(15): {compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1579), layer(15): {compute_graph.l2l3_12_spill.out[3], compute_graph.l2l3_13_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1580), layer(15): {compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[3], compute_graph.l2_14.in[0], compute_graph.l2_14.in[1], compute_graph.l2_14.in[2], compute_graph.l2_14.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1337), layer(15): compute_graph.flexml_layers[9].compute_node[0][0], compute_graph.flexml_layers[9].compute_node[0][1], compute_graph.flexml_layers[9].compute_node[0][2], compute_graph.flexml_layers[9].compute_node[0][3], compute_graph.flexml_layers[9].compute_node[1][0], compute_graph.flexml_layers[9].compute_node[1][1], compute_graph.flexml_layers[9].compute_node[1][2], compute_graph.flexml_layers[9].compute_node[1][3], compute_graph.flexml_layers[9].compute_node[2][0], compute_graph.flexml_layers[9].compute_node[2][1], compute_graph.flexml_layers[9].compute_node[2][2], compute_graph.flexml_layers[9].compute_node[2][3], compute_graph.flexml_layers[9].compute_node[3][0], compute_graph.flexml_layers[9].compute_node[3][1], compute_graph.flexml_layers[9].compute_node[3][2], compute_graph.flexml_layers[9].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[3], compute_graph.Layer_14_l2_wts.out[0], compute_graph.Layer_14_l2_wts.out[1], compute_graph.Layer_14_l2_wts.out[2], compute_graph.Layer_14_l2_wts.out[3], compute_graph.l2_14.in[0], compute_graph.l2_14.in[1], compute_graph.l2_14.in[2], compute_graph.l2_14.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(822): mode(3), layer(16): {compute_graph.Layer_15_wts_ddr.out[0], compute_graph.Layer_15_wts_ddr.out[1], compute_graph.Layer_15_l2_wts.in[0], compute_graph.Layer_15_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(822) is pipelined with KernelLayerNode(1337) +INFO: [aiecompiler 77-6295] For BufferSenderNode(1581), layer(15): {compute_graph.l2_14.out[0], compute_graph.l2_14.out[2]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1582), layer(15): {compute_graph.l2l3_14_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[2]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1583), layer(15): {compute_graph.l2_14.out[1], compute_graph.l2_14.out[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1584), layer(15): {compute_graph.l2l3_14_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1035): mode(0), layer(16): {compute_graph.l2l3_14_spill.out[0], compute_graph.l2l3_14_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[3], compute_graph.l2_15.in[0], compute_graph.l2_15.in[1], compute_graph.l2_15.in[2], compute_graph.l2_15.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1338), layer(16): compute_graph.flexml_layers[10].compute_node[0][0], compute_graph.flexml_layers[10].compute_node[0][1], compute_graph.flexml_layers[10].compute_node[0][2], compute_graph.flexml_layers[10].compute_node[0][3], compute_graph.flexml_layers[10].compute_node[1][0], compute_graph.flexml_layers[10].compute_node[1][1], compute_graph.flexml_layers[10].compute_node[1][2], compute_graph.flexml_layers[10].compute_node[1][3], compute_graph.flexml_layers[10].compute_node[2][0], compute_graph.flexml_layers[10].compute_node[2][1], compute_graph.flexml_layers[10].compute_node[2][2], compute_graph.flexml_layers[10].compute_node[2][3], compute_graph.flexml_layers[10].compute_node[3][0], compute_graph.flexml_layers[10].compute_node[3][1], compute_graph.flexml_layers[10].compute_node[3][2], compute_graph.flexml_layers[10].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[3], compute_graph.Layer_15_l2_wts.out[0], compute_graph.Layer_15_l2_wts.out[1], compute_graph.Layer_15_l2_wts.out[2], compute_graph.Layer_15_l2_wts.out[3], compute_graph.l2_15.in[0], compute_graph.l2_15.in[1], compute_graph.l2_15.in[2], compute_graph.l2_15.in[3]}, Scheduler computes number of sub-layer phases to be 5. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(908): mode(3), layer(16): {compute_graph.l2_15.out[0], compute_graph.l2_15.out[1], compute_graph.l2l3_15_spill.in[0], compute_graph.l2l3_15_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(908) is pipelined with KernelLayerNode(1338) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(823): mode(3), layer(17): {compute_graph.Layer_16_wts_ddr.out[0], compute_graph.Layer_16_wts_ddr.out[1], compute_graph.Layer_16_l2_wts.in[0], compute_graph.Layer_16_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(823) is pipelined with KernelLayerNode(1338) +INFO: [aiecompiler 77-23129] BufferToBufferNode 909 will not be pipelined because it's in the same layer 17 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2l3_15_spill.out[0], compute_graph.l2l3_15_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1036): mode(0), layer(17): {compute_graph.l2l3_15_spill.out[0], compute_graph.l2l3_15_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[1]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.Layer_16_l2_wts.out[0], compute_graph.Layer_16_l2_wts.out[1], compute_graph.Layer_16_l2_wts.out[2], compute_graph.Layer_16_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1339), layer(17): compute_graph.flexml_layers[11].compute_node[0][0], compute_graph.flexml_layers[11].compute_node[0][1], compute_graph.flexml_layers[11].compute_node[0][2], compute_graph.flexml_layers[11].compute_node[0][3], compute_graph.flexml_layers[11].compute_node[1][0], compute_graph.flexml_layers[11].compute_node[1][1], compute_graph.flexml_layers[11].compute_node[1][2], compute_graph.flexml_layers[11].compute_node[1][3], compute_graph.flexml_layers[11].compute_node[2][0], compute_graph.flexml_layers[11].compute_node[2][1], compute_graph.flexml_layers[11].compute_node[2][2], compute_graph.flexml_layers[11].compute_node[2][3], compute_graph.flexml_layers[11].compute_node[3][0], compute_graph.flexml_layers[11].compute_node[3][1], compute_graph.flexml_layers[11].compute_node[3][2], compute_graph.flexml_layers[11].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[3], compute_graph.Layer_16_l2_wts.out[0], compute_graph.Layer_16_l2_wts.out[1], compute_graph.Layer_16_l2_wts.out[2], compute_graph.Layer_16_l2_wts.out[3], compute_graph.l2_16.in[0], compute_graph.l2_16.in[1], compute_graph.l2_16.in[2], compute_graph.l2_16.in[3]}, Scheduler computes number of sub-layer phases to be 6. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(909): mode(3), layer(17): {compute_graph.l2_16.out[0], compute_graph.l2_16.out[1], compute_graph.l2l3_16_spill.in[0], compute_graph.l2l3_16_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1037): mode(0), layer(18): {compute_graph.l2l3_16_spill.out[0], compute_graph.templated_graph_17.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1234): mode(0), layer(18): {compute_graph.templated_graph_17.ifm.out[0], compute_graph.templated_graph_17.ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1235): mode(0), layer(18): {compute_graph.templated_graph_17.ofm.out[0], compute_graph.l2l3_scratch_0_17_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1038): mode(0), layer(19): {compute_graph.l2l3_scratch_0_17_spill.out[0], compute_graph.templated_graph_17.ifm2.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1236): mode(0), layer(19): {compute_graph.templated_graph_17.ifm2.out[0], compute_graph.l2l3_17_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1039): mode(0), layer(20): {compute_graph.l2l3_17_spill.out[0], compute_graph.l2l3_17_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(824): mode(4), layer(20): {compute_graph.Layer_18_wts_ddr.out[0], compute_graph.Layer_18_wts_ddr.out[1], compute_graph.Layer_18_l2_wts.in[0], compute_graph.Layer_18_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(824) is pipelined with BufferToBufferNode(1039) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1340), layer(20): compute_graph.flexml_layers[12].compute_node[0][0], compute_graph.flexml_layers[12].compute_node[0][1], compute_graph.flexml_layers[12].compute_node[0][2], compute_graph.flexml_layers[12].compute_node[0][3], compute_graph.flexml_layers[12].compute_node[1][0], compute_graph.flexml_layers[12].compute_node[1][1], compute_graph.flexml_layers[12].compute_node[1][2], compute_graph.flexml_layers[12].compute_node[1][3], compute_graph.flexml_layers[12].compute_node[2][0], compute_graph.flexml_layers[12].compute_node[2][1], compute_graph.flexml_layers[12].compute_node[2][2], compute_graph.flexml_layers[12].compute_node[2][3], compute_graph.flexml_layers[12].compute_node[3][0], compute_graph.flexml_layers[12].compute_node[3][1], compute_graph.flexml_layers[12].compute_node[3][2], compute_graph.flexml_layers[12].compute_node[3][3], {compute_graph.Layer_18_l2_wts.out[0], compute_graph.Layer_18_l2_wts.out[1], compute_graph.Layer_18_l2_wts.out[2], compute_graph.Layer_18_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[3], compute_graph.l2_18.in[0], compute_graph.l2_18.in[1], compute_graph.l2_18.in[2], compute_graph.l2_18.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(825): mode(3), layer(21): {compute_graph.Layer_19_wts_ddr.out[0], compute_graph.Layer_19_wts_ddr.out[1], compute_graph.Layer_19_l2_wts.in[0], compute_graph.Layer_19_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(825) is pipelined with KernelLayerNode(1340) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2_18.out[0], compute_graph.l2_18.out[1], compute_graph.l2_18.out[2], compute_graph.l2_18.out[3], compute_graph.l2_19.in[0], compute_graph.l2_19.in[1], compute_graph.l2_19.in[2], compute_graph.l2_19.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1341), layer(21): compute_graph.flexml_layers[13].compute_node[0][0], compute_graph.flexml_layers[13].compute_node[0][1], compute_graph.flexml_layers[13].compute_node[0][2], compute_graph.flexml_layers[13].compute_node[0][3], compute_graph.flexml_layers[13].compute_node[1][0], compute_graph.flexml_layers[13].compute_node[1][1], compute_graph.flexml_layers[13].compute_node[1][2], compute_graph.flexml_layers[13].compute_node[1][3], compute_graph.flexml_layers[13].compute_node[2][0], compute_graph.flexml_layers[13].compute_node[2][1], compute_graph.flexml_layers[13].compute_node[2][2], compute_graph.flexml_layers[13].compute_node[2][3], compute_graph.flexml_layers[13].compute_node[3][0], compute_graph.flexml_layers[13].compute_node[3][1], compute_graph.flexml_layers[13].compute_node[3][2], compute_graph.flexml_layers[13].compute_node[3][3], {compute_graph.Layer_19_l2_wts.out[0], compute_graph.Layer_19_l2_wts.out[1], compute_graph.Layer_19_l2_wts.out[2], compute_graph.Layer_19_l2_wts.out[3], compute_graph.l2_18.out[0], compute_graph.l2_18.out[1], compute_graph.l2_18.out[2], compute_graph.l2_18.out[3], compute_graph.l2_19.in[0], compute_graph.l2_19.in[1], compute_graph.l2_19.in[2], compute_graph.l2_19.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(910): mode(3), layer(21): {compute_graph.l2_18.out[4], compute_graph.l2_18.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(910) is pipelined with KernelLayerNode(1341) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(826): mode(3), layer(22): {compute_graph.Layer_20_wts_ddr.out[0], compute_graph.Layer_20_wts_ddr.out[1], compute_graph.Layer_20_l2_wts.in[0], compute_graph.Layer_20_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(826) is pipelined with KernelLayerNode(1341) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(911): mode(0), layer(21): {compute_graph.l2_19.out[0], compute_graph.l2_19.out[1], compute_graph.l2l3_19_spill.in[0], compute_graph.l2l3_19_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1041): mode(0), layer(22): {compute_graph.l2l3_19_spill.out[0], compute_graph.l2l3_19_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.Layer_20_l2_wts.out[0], compute_graph.Layer_20_l2_wts.out[1], compute_graph.Layer_20_l2_wts.out[2], compute_graph.Layer_20_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1342), layer(22): compute_graph.flexml_layers[14].compute_node[0][0], compute_graph.flexml_layers[14].compute_node[0][1], compute_graph.flexml_layers[14].compute_node[0][2], compute_graph.flexml_layers[14].compute_node[0][3], compute_graph.flexml_layers[14].compute_node[1][0], compute_graph.flexml_layers[14].compute_node[1][1], compute_graph.flexml_layers[14].compute_node[1][2], compute_graph.flexml_layers[14].compute_node[1][3], compute_graph.flexml_layers[14].compute_node[2][0], compute_graph.flexml_layers[14].compute_node[2][1], compute_graph.flexml_layers[14].compute_node[2][2], compute_graph.flexml_layers[14].compute_node[2][3], compute_graph.flexml_layers[14].compute_node[3][0], compute_graph.flexml_layers[14].compute_node[3][1], compute_graph.flexml_layers[14].compute_node[3][2], compute_graph.flexml_layers[14].compute_node[3][3], {compute_graph.Layer_20_l2_wts.out[0], compute_graph.Layer_20_l2_wts.out[1], compute_graph.Layer_20_l2_wts.out[2], compute_graph.Layer_20_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[3], compute_graph.l2_20.in[0], compute_graph.l2_20.in[1], compute_graph.l2_20.in[2], compute_graph.l2_20.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(912): mode(3), layer(22): {compute_graph.l2_20.out[0], compute_graph.l2_20.out[1], compute_graph.l2l3_20_spill.in[0], compute_graph.l2l3_20_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(912) is pipelined with KernelLayerNode(1342) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1042): mode(0), layer(23): {compute_graph.l2l3_20_spill.out[0], compute_graph.templated_graph_21.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1237): mode(0), layer(23): {compute_graph.templated_graph_21.ifm.out[0], compute_graph.templated_graph_21.ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1238): mode(0), layer(23): {compute_graph.templated_graph_21.ofm.out[0], compute_graph.l2l3_scratch_0_21_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1043): mode(0), layer(24): {compute_graph.l2l3_scratch_0_21_spill.out[0], compute_graph.templated_graph_21.ifm2.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1239): mode(0), layer(24): {compute_graph.templated_graph_21.ifm2.out[0], compute_graph.l2l3_21_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818 overlaps with memory space used by compute_graph.l2_22. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1040): mode(0), layer(25): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(827): mode(4), layer(25): {compute_graph.Layer_22_wts_ddr.out[0], compute_graph.Layer_22_wts_ddr.out[1], compute_graph.Layer_22_l2_wts.in[0], compute_graph.Layer_22_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(827) is pipelined with BufferToBufferNode(1040) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1044): mode(0), layer(25): {compute_graph.l2l3_21_spill.out[0], compute_graph.l2l3_21_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[3], compute_graph.l2_22.in[0], compute_graph.l2_22.in[1], compute_graph.l2_22.in[2], compute_graph.l2_22.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1343), layer(25): compute_graph.flexml_layers[15].compute_node[0][0], compute_graph.flexml_layers[15].compute_node[0][1], compute_graph.flexml_layers[15].compute_node[0][2], compute_graph.flexml_layers[15].compute_node[0][3], compute_graph.flexml_layers[15].compute_node[1][0], compute_graph.flexml_layers[15].compute_node[1][1], compute_graph.flexml_layers[15].compute_node[1][2], compute_graph.flexml_layers[15].compute_node[1][3], compute_graph.flexml_layers[15].compute_node[2][0], compute_graph.flexml_layers[15].compute_node[2][1], compute_graph.flexml_layers[15].compute_node[2][2], compute_graph.flexml_layers[15].compute_node[2][3], compute_graph.flexml_layers[15].compute_node[3][0], compute_graph.flexml_layers[15].compute_node[3][1], compute_graph.flexml_layers[15].compute_node[3][2], compute_graph.flexml_layers[15].compute_node[3][3], {compute_graph.Layer_22_l2_wts.out[0], compute_graph.Layer_22_l2_wts.out[1], compute_graph.Layer_22_l2_wts.out[2], compute_graph.Layer_22_l2_wts.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[3], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[3], compute_graph.l2_22.in[0], compute_graph.l2_22.in[1], compute_graph.l2_22.in[2], compute_graph.l2_22.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(828): mode(3), layer(26): {compute_graph.Layer_23_wts_ddr.out[0], compute_graph.Layer_23_wts_ddr.out[1], compute_graph.Layer_23_l2_wts.in[0], compute_graph.Layer_23_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(828) is pipelined with KernelLayerNode(1343) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2_22.out[0], compute_graph.l2_22.out[1], compute_graph.l2_22.out[2], compute_graph.l2_22.out[3], compute_graph.l2_23.in[0], compute_graph.l2_23.in[1], compute_graph.l2_23.in[2], compute_graph.l2_23.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1344), layer(26): compute_graph.flexml_layers[16].compute_node[0][0], compute_graph.flexml_layers[16].compute_node[0][1], compute_graph.flexml_layers[16].compute_node[0][2], compute_graph.flexml_layers[16].compute_node[0][3], compute_graph.flexml_layers[16].compute_node[1][0], compute_graph.flexml_layers[16].compute_node[1][1], compute_graph.flexml_layers[16].compute_node[1][2], compute_graph.flexml_layers[16].compute_node[1][3], compute_graph.flexml_layers[16].compute_node[2][0], compute_graph.flexml_layers[16].compute_node[2][1], compute_graph.flexml_layers[16].compute_node[2][2], compute_graph.flexml_layers[16].compute_node[2][3], compute_graph.flexml_layers[16].compute_node[3][0], compute_graph.flexml_layers[16].compute_node[3][1], compute_graph.flexml_layers[16].compute_node[3][2], compute_graph.flexml_layers[16].compute_node[3][3], {compute_graph.Layer_23_l2_wts.out[0], compute_graph.Layer_23_l2_wts.out[1], compute_graph.Layer_23_l2_wts.out[2], compute_graph.Layer_23_l2_wts.out[3], compute_graph.l2_22.out[0], compute_graph.l2_22.out[1], compute_graph.l2_22.out[2], compute_graph.l2_22.out[3], compute_graph.l2_23.in[0], compute_graph.l2_23.in[1], compute_graph.l2_23.in[2], compute_graph.l2_23.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(913): mode(3), layer(26): {compute_graph.l2_22.out[4], compute_graph.l2_22.out[5], compute_graph.spill_L3_Concat_Buffer_layer_256.in[2], compute_graph.spill_L3_Concat_Buffer_layer_256.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(913) is pipelined with KernelLayerNode(1344) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(829): mode(3), layer(27): {compute_graph.Layer_24_wts_ddr.out[0], compute_graph.Layer_24_wts_ddr.out[1], compute_graph.Layer_24_l2_wts.in[0], compute_graph.Layer_24_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(829) is pipelined with KernelLayerNode(1344) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2_23.out[0], compute_graph.l2_23.out[1], compute_graph.l2_23.out[2], compute_graph.l2_23.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1345), layer(27): compute_graph.flexml_layers[17].compute_node[0][0], compute_graph.flexml_layers[17].compute_node[0][1], compute_graph.flexml_layers[17].compute_node[0][2], compute_graph.flexml_layers[17].compute_node[0][3], compute_graph.flexml_layers[17].compute_node[1][0], compute_graph.flexml_layers[17].compute_node[1][1], compute_graph.flexml_layers[17].compute_node[1][2], compute_graph.flexml_layers[17].compute_node[1][3], compute_graph.flexml_layers[17].compute_node[2][0], compute_graph.flexml_layers[17].compute_node[2][1], compute_graph.flexml_layers[17].compute_node[2][2], compute_graph.flexml_layers[17].compute_node[2][3], compute_graph.flexml_layers[17].compute_node[3][0], compute_graph.flexml_layers[17].compute_node[3][1], compute_graph.flexml_layers[17].compute_node[3][2], compute_graph.flexml_layers[17].compute_node[3][3], {compute_graph.l2_23.out[0], compute_graph.l2_23.out[1], compute_graph.l2_23.out[2], compute_graph.l2_23.out[3], compute_graph.Layer_24_l2_wts.out[0], compute_graph.Layer_24_l2_wts.out[1], compute_graph.Layer_24_l2_wts.out[2], compute_graph.Layer_24_l2_wts.out[3], compute_graph.l2_24.in[0], compute_graph.l2_24.in[1], compute_graph.l2_24.in[2], compute_graph.l2_24.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(915): mode(0), layer(27): {compute_graph.l2_24.out[2], compute_graph.l2_24.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(914): mode(0), layer(27): {compute_graph.l2_24.out[0], compute_graph.l2_24.out[1], compute_graph.l2l3_24_spill.in[0], compute_graph.l2l3_24_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1046): mode(0), layer(28): {compute_graph.l2l3_24_spill.out[0], compute_graph.templated_graph_25.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1240): mode(0), layer(28): {compute_graph.templated_graph_25.trans_mt_ifm.out[0], compute_graph.templated_graph_25.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1241): mode(0), layer(28): {compute_graph.templated_graph_25.trans_mt_ofm.out[0], compute_graph.l2l3_25_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1047): mode(0), layer(29): {compute_graph.l2l3_25_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1346), layer(29): compute_graph.flexml_layers[18].compute_node[0][0], compute_graph.flexml_layers[18].compute_node[0][1], compute_graph.flexml_layers[18].compute_node[0][2], compute_graph.flexml_layers[18].compute_node[0][3], compute_graph.flexml_layers[18].compute_node[1][0], compute_graph.flexml_layers[18].compute_node[1][1], compute_graph.flexml_layers[18].compute_node[1][2], compute_graph.flexml_layers[18].compute_node[1][3], compute_graph.flexml_layers[18].compute_node[2][0], compute_graph.flexml_layers[18].compute_node[2][1], compute_graph.flexml_layers[18].compute_node[2][2], compute_graph.flexml_layers[18].compute_node[2][3], compute_graph.flexml_layers[18].compute_node[3][0], compute_graph.flexml_layers[18].compute_node[3][1], compute_graph.flexml_layers[18].compute_node[3][2], compute_graph.flexml_layers[18].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[3], compute_graph.l2_26.in[0], compute_graph.l2_26.in[1], compute_graph.l2_26.in[2], compute_graph.l2_26.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(830): mode(3), layer(30): {compute_graph.Layer_27_wts_ddr.out[0], compute_graph.Layer_27_wts_ddr.out[1], compute_graph.Layer_27_l2_wts.in[0], compute_graph.Layer_27_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(830) is pipelined with KernelLayerNode(1346) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1347), layer(30): compute_graph.flexml_layers[19].compute_node[0][0], compute_graph.flexml_layers[19].compute_node[0][1], compute_graph.flexml_layers[19].compute_node[0][2], compute_graph.flexml_layers[19].compute_node[0][3], compute_graph.flexml_layers[19].compute_node[1][0], compute_graph.flexml_layers[19].compute_node[1][1], compute_graph.flexml_layers[19].compute_node[1][2], compute_graph.flexml_layers[19].compute_node[1][3], compute_graph.flexml_layers[19].compute_node[2][0], compute_graph.flexml_layers[19].compute_node[2][1], compute_graph.flexml_layers[19].compute_node[2][2], compute_graph.flexml_layers[19].compute_node[2][3], compute_graph.flexml_layers[19].compute_node[3][0], compute_graph.flexml_layers[19].compute_node[3][1], compute_graph.flexml_layers[19].compute_node[3][2], compute_graph.flexml_layers[19].compute_node[3][3], {compute_graph.l2_26.out[0], compute_graph.l2_26.out[1], compute_graph.l2_26.out[2], compute_graph.l2_26.out[3], compute_graph.Layer_27_l2_wts.out[0], compute_graph.Layer_27_l2_wts.out[1], compute_graph.Layer_27_l2_wts.out[2], compute_graph.Layer_27_l2_wts.out[3], compute_graph.l2_27.in[0], compute_graph.l2_27.in[1], compute_graph.l2_27.in[2], compute_graph.l2_27.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(831): mode(3), layer(31): {compute_graph.Layer_28_wts_ddr.out[0], compute_graph.Layer_28_wts_ddr.out[1], compute_graph.Layer_28_l2_wts.in[0], compute_graph.Layer_28_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(831) is pipelined with KernelLayerNode(1347) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1348), layer(31): compute_graph.flexml_layers[20].compute_node[0][0], compute_graph.flexml_layers[20].compute_node[0][1], compute_graph.flexml_layers[20].compute_node[0][2], compute_graph.flexml_layers[20].compute_node[0][3], compute_graph.flexml_layers[20].compute_node[1][0], compute_graph.flexml_layers[20].compute_node[1][1], compute_graph.flexml_layers[20].compute_node[1][2], compute_graph.flexml_layers[20].compute_node[1][3], compute_graph.flexml_layers[20].compute_node[2][0], compute_graph.flexml_layers[20].compute_node[2][1], compute_graph.flexml_layers[20].compute_node[2][2], compute_graph.flexml_layers[20].compute_node[2][3], compute_graph.flexml_layers[20].compute_node[3][0], compute_graph.flexml_layers[20].compute_node[3][1], compute_graph.flexml_layers[20].compute_node[3][2], compute_graph.flexml_layers[20].compute_node[3][3], {compute_graph.l2_27.out[0], compute_graph.l2_27.out[1], compute_graph.l2_27.out[2], compute_graph.l2_27.out[3], compute_graph.Layer_28_l2_wts.out[0], compute_graph.Layer_28_l2_wts.out[1], compute_graph.Layer_28_l2_wts.out[2], compute_graph.Layer_28_l2_wts.out[3], compute_graph.l2_28.in[0], compute_graph.l2_28.in[1], compute_graph.l2_28.in[2], compute_graph.l2_28.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1349), layer(32): compute_graph.flexml_layers[21].compute_node[0][0], compute_graph.flexml_layers[21].compute_node[0][1], compute_graph.flexml_layers[21].compute_node[0][2], compute_graph.flexml_layers[21].compute_node[0][3], compute_graph.flexml_layers[21].compute_node[1][0], compute_graph.flexml_layers[21].compute_node[1][1], compute_graph.flexml_layers[21].compute_node[1][2], compute_graph.flexml_layers[21].compute_node[1][3], compute_graph.flexml_layers[21].compute_node[2][0], compute_graph.flexml_layers[21].compute_node[2][1], compute_graph.flexml_layers[21].compute_node[2][2], compute_graph.flexml_layers[21].compute_node[2][3], compute_graph.flexml_layers[21].compute_node[3][0], compute_graph.flexml_layers[21].compute_node[3][1], compute_graph.flexml_layers[21].compute_node[3][2], compute_graph.flexml_layers[21].compute_node[3][3], {compute_graph.l2_28.out[0], compute_graph.l2_28.out[1], compute_graph.l2_28.out[2], compute_graph.l2_28.out[3], compute_graph.l2_29.in[0], compute_graph.l2_29.in[1], compute_graph.l2_29.in[2], compute_graph.l2_29.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1350), layer(33): compute_graph.flexml_layers[22].compute_node[0][0], compute_graph.flexml_layers[22].compute_node[0][1], compute_graph.flexml_layers[22].compute_node[0][2], compute_graph.flexml_layers[22].compute_node[0][3], compute_graph.flexml_layers[22].compute_node[1][0], compute_graph.flexml_layers[22].compute_node[1][1], compute_graph.flexml_layers[22].compute_node[1][2], compute_graph.flexml_layers[22].compute_node[1][3], compute_graph.flexml_layers[22].compute_node[2][0], compute_graph.flexml_layers[22].compute_node[2][1], compute_graph.flexml_layers[22].compute_node[2][2], compute_graph.flexml_layers[22].compute_node[2][3], compute_graph.flexml_layers[22].compute_node[3][0], compute_graph.flexml_layers[22].compute_node[3][1], compute_graph.flexml_layers[22].compute_node[3][2], compute_graph.flexml_layers[22].compute_node[3][3], {compute_graph.l2_29.out[0], compute_graph.l2_29.out[1], compute_graph.l2_29.out[2], compute_graph.l2_29.out[3], compute_graph.l2_30.in[0], compute_graph.l2_30.in[1], compute_graph.l2_30.in[2], compute_graph.l2_30.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1351), layer(34): compute_graph.flexml_layers[23].compute_node[0][0], compute_graph.flexml_layers[23].compute_node[0][1], compute_graph.flexml_layers[23].compute_node[0][2], compute_graph.flexml_layers[23].compute_node[0][3], compute_graph.flexml_layers[23].compute_node[1][0], compute_graph.flexml_layers[23].compute_node[1][1], compute_graph.flexml_layers[23].compute_node[1][2], compute_graph.flexml_layers[23].compute_node[1][3], compute_graph.flexml_layers[23].compute_node[2][0], compute_graph.flexml_layers[23].compute_node[2][1], compute_graph.flexml_layers[23].compute_node[2][2], compute_graph.flexml_layers[23].compute_node[2][3], compute_graph.flexml_layers[23].compute_node[3][0], compute_graph.flexml_layers[23].compute_node[3][1], compute_graph.flexml_layers[23].compute_node[3][2], compute_graph.flexml_layers[23].compute_node[3][3], {compute_graph.l2_30.out[0], compute_graph.l2_30.out[1], compute_graph.l2_30.out[2], compute_graph.l2_30.out[3], compute_graph.l2_31.in[0], compute_graph.l2_31.in[1], compute_graph.l2_31.in[2], compute_graph.l2_31.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(916): mode(0), layer(34): {compute_graph.l2_31.out[1], compute_graph.l2l3_31_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1048): mode(0), layer(35): {compute_graph.l2l3_31_spill.out[0], compute_graph.templated_graph_32.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1242): mode(0), layer(35): {compute_graph.templated_graph_32.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_32_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1049): mode(0), layer(36): {compute_graph.l2l3_scratch_0_32_spill.out[0], compute_graph.templated_graph_32.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1243): mode(0), layer(36): {compute_graph.templated_graph_32.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_32_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1050): mode(0), layer(37): {compute_graph.l2l3_scratch_1_32_spill.out[0], compute_graph.templated_graph_32.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1244): mode(0), layer(37): {compute_graph.templated_graph_32.g2.ifm_mem.out[0], compute_graph.l2l3_32_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1045): mode(0), layer(38): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1051): mode(0), layer(38): {compute_graph.l2l3_32_spill.out[0], compute_graph.l2l3_32_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1352), layer(38): compute_graph.flexml_layers[24].compute_node[0][0], compute_graph.flexml_layers[24].compute_node[0][1], compute_graph.flexml_layers[24].compute_node[0][2], compute_graph.flexml_layers[24].compute_node[0][3], compute_graph.flexml_layers[24].compute_node[1][0], compute_graph.flexml_layers[24].compute_node[1][1], compute_graph.flexml_layers[24].compute_node[1][2], compute_graph.flexml_layers[24].compute_node[1][3], compute_graph.flexml_layers[24].compute_node[2][0], compute_graph.flexml_layers[24].compute_node[2][1], compute_graph.flexml_layers[24].compute_node[2][2], compute_graph.flexml_layers[24].compute_node[2][3], compute_graph.flexml_layers[24].compute_node[3][0], compute_graph.flexml_layers[24].compute_node[3][1], compute_graph.flexml_layers[24].compute_node[3][2], compute_graph.flexml_layers[24].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[3], compute_graph.l2_33.in[0], compute_graph.l2_33.in[1], compute_graph.l2_33.in[2], compute_graph.l2_33.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(832): mode(3), layer(39): {compute_graph.Layer_34_wts_ddr.out[0], compute_graph.Layer_34_wts_ddr.out[1], compute_graph.Layer_34_l2_wts.in[0], compute_graph.Layer_34_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(832) is pipelined with KernelLayerNode(1352) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1353), layer(39): compute_graph.flexml_layers[25].compute_node[0][0], compute_graph.flexml_layers[25].compute_node[0][1], compute_graph.flexml_layers[25].compute_node[0][2], compute_graph.flexml_layers[25].compute_node[0][3], compute_graph.flexml_layers[25].compute_node[1][0], compute_graph.flexml_layers[25].compute_node[1][1], compute_graph.flexml_layers[25].compute_node[1][2], compute_graph.flexml_layers[25].compute_node[1][3], compute_graph.flexml_layers[25].compute_node[2][0], compute_graph.flexml_layers[25].compute_node[2][1], compute_graph.flexml_layers[25].compute_node[2][2], compute_graph.flexml_layers[25].compute_node[2][3], compute_graph.flexml_layers[25].compute_node[3][0], compute_graph.flexml_layers[25].compute_node[3][1], compute_graph.flexml_layers[25].compute_node[3][2], compute_graph.flexml_layers[25].compute_node[3][3], {compute_graph.l2_33.out[0], compute_graph.l2_33.out[1], compute_graph.l2_33.out[2], compute_graph.l2_33.out[3], compute_graph.Layer_34_l2_wts.out[0], compute_graph.Layer_34_l2_wts.out[1], compute_graph.Layer_34_l2_wts.out[2], compute_graph.Layer_34_l2_wts.out[3], compute_graph.l2_34.in[0], compute_graph.l2_34.in[1], compute_graph.l2_34.in[2], compute_graph.l2_34.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(833): mode(3), layer(40): {compute_graph.Layer_35_wts_ddr.out[0], compute_graph.Layer_35_wts_ddr.out[1], compute_graph.Layer_35_l2_wts.in[0], compute_graph.Layer_35_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(833) is pipelined with KernelLayerNode(1353) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1354), layer(40): compute_graph.flexml_layers[26].compute_node[0][0], compute_graph.flexml_layers[26].compute_node[0][1], compute_graph.flexml_layers[26].compute_node[0][2], compute_graph.flexml_layers[26].compute_node[0][3], compute_graph.flexml_layers[26].compute_node[1][0], compute_graph.flexml_layers[26].compute_node[1][1], compute_graph.flexml_layers[26].compute_node[1][2], compute_graph.flexml_layers[26].compute_node[1][3], compute_graph.flexml_layers[26].compute_node[2][0], compute_graph.flexml_layers[26].compute_node[2][1], compute_graph.flexml_layers[26].compute_node[2][2], compute_graph.flexml_layers[26].compute_node[2][3], compute_graph.flexml_layers[26].compute_node[3][0], compute_graph.flexml_layers[26].compute_node[3][1], compute_graph.flexml_layers[26].compute_node[3][2], compute_graph.flexml_layers[26].compute_node[3][3], {compute_graph.l2_34.out[0], compute_graph.l2_34.out[1], compute_graph.l2_34.out[2], compute_graph.l2_34.out[3], compute_graph.Layer_35_l2_wts.out[0], compute_graph.Layer_35_l2_wts.out[1], compute_graph.Layer_35_l2_wts.out[2], compute_graph.Layer_35_l2_wts.out[3], compute_graph.l2_35.in[0], compute_graph.l2_35.in[1], compute_graph.l2_35.in[2], compute_graph.l2_35.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(917): mode(3), layer(40): {compute_graph.l2_34.out[4], compute_graph.l2_34.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(917) is pipelined with KernelLayerNode(1354) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(834): mode(3), layer(41): {compute_graph.Layer_36_wts_ddr.out[0], compute_graph.Layer_36_wts_ddr.out[1], compute_graph.Layer_36_l2_wts.in[0], compute_graph.Layer_36_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(834) is pipelined with KernelLayerNode(1354) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1355), layer(41): compute_graph.flexml_layers[27].compute_node[0][0], compute_graph.flexml_layers[27].compute_node[0][1], compute_graph.flexml_layers[27].compute_node[0][2], compute_graph.flexml_layers[27].compute_node[0][3], compute_graph.flexml_layers[27].compute_node[1][0], compute_graph.flexml_layers[27].compute_node[1][1], compute_graph.flexml_layers[27].compute_node[1][2], compute_graph.flexml_layers[27].compute_node[1][3], compute_graph.flexml_layers[27].compute_node[2][0], compute_graph.flexml_layers[27].compute_node[2][1], compute_graph.flexml_layers[27].compute_node[2][2], compute_graph.flexml_layers[27].compute_node[2][3], compute_graph.flexml_layers[27].compute_node[3][0], compute_graph.flexml_layers[27].compute_node[3][1], compute_graph.flexml_layers[27].compute_node[3][2], compute_graph.flexml_layers[27].compute_node[3][3], {compute_graph.l2_35.out[0], compute_graph.l2_35.out[1], compute_graph.l2_35.out[2], compute_graph.l2_35.out[3], compute_graph.Layer_36_l2_wts.out[0], compute_graph.Layer_36_l2_wts.out[1], compute_graph.Layer_36_l2_wts.out[2], compute_graph.Layer_36_l2_wts.out[3], compute_graph.l2_36.in[0], compute_graph.l2_36.in[1], compute_graph.l2_36.in[2], compute_graph.l2_36.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(918): mode(0), layer(41): {compute_graph.l2_36.out[0], compute_graph.l2_36.out[1], compute_graph.l2l3_36_spill.in[0], compute_graph.l2l3_36_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(919): mode(0), layer(41): {compute_graph.l2_36.out[2], compute_graph.l2_36.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1054): mode(0), layer(42): {compute_graph.l2l3_36_spill.out[0], compute_graph.templated_graph_37.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1245): mode(0), layer(42): {compute_graph.templated_graph_37.trans_mt_ifm.out[0], compute_graph.templated_graph_37.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1246): mode(0), layer(42): {compute_graph.templated_graph_37.trans_mt_ofm.out[0], compute_graph.l2l3_37_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1055): mode(0), layer(43): {compute_graph.l2l3_37_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1356), layer(43): compute_graph.flexml_layers[28].compute_node[0][0], compute_graph.flexml_layers[28].compute_node[0][1], compute_graph.flexml_layers[28].compute_node[0][2], compute_graph.flexml_layers[28].compute_node[0][3], compute_graph.flexml_layers[28].compute_node[1][0], compute_graph.flexml_layers[28].compute_node[1][1], compute_graph.flexml_layers[28].compute_node[1][2], compute_graph.flexml_layers[28].compute_node[1][3], compute_graph.flexml_layers[28].compute_node[2][0], compute_graph.flexml_layers[28].compute_node[2][1], compute_graph.flexml_layers[28].compute_node[2][2], compute_graph.flexml_layers[28].compute_node[2][3], compute_graph.flexml_layers[28].compute_node[3][0], compute_graph.flexml_layers[28].compute_node[3][1], compute_graph.flexml_layers[28].compute_node[3][2], compute_graph.flexml_layers[28].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[3], compute_graph.l2_38.in[0], compute_graph.l2_38.in[1], compute_graph.l2_38.in[2], compute_graph.l2_38.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(835): mode(3), layer(44): {compute_graph.Layer_39_wts_ddr.out[0], compute_graph.Layer_39_wts_ddr.out[1], compute_graph.Layer_39_l2_wts.in[0], compute_graph.Layer_39_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(835) is pipelined with KernelLayerNode(1356) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1357), layer(44): compute_graph.flexml_layers[29].compute_node[0][0], compute_graph.flexml_layers[29].compute_node[0][1], compute_graph.flexml_layers[29].compute_node[0][2], compute_graph.flexml_layers[29].compute_node[0][3], compute_graph.flexml_layers[29].compute_node[1][0], compute_graph.flexml_layers[29].compute_node[1][1], compute_graph.flexml_layers[29].compute_node[1][2], compute_graph.flexml_layers[29].compute_node[1][3], compute_graph.flexml_layers[29].compute_node[2][0], compute_graph.flexml_layers[29].compute_node[2][1], compute_graph.flexml_layers[29].compute_node[2][2], compute_graph.flexml_layers[29].compute_node[2][3], compute_graph.flexml_layers[29].compute_node[3][0], compute_graph.flexml_layers[29].compute_node[3][1], compute_graph.flexml_layers[29].compute_node[3][2], compute_graph.flexml_layers[29].compute_node[3][3], {compute_graph.l2_38.out[0], compute_graph.l2_38.out[1], compute_graph.l2_38.out[2], compute_graph.l2_38.out[3], compute_graph.Layer_39_l2_wts.out[0], compute_graph.Layer_39_l2_wts.out[1], compute_graph.Layer_39_l2_wts.out[2], compute_graph.Layer_39_l2_wts.out[3], compute_graph.l2_39.in[0], compute_graph.l2_39.in[1], compute_graph.l2_39.in[2], compute_graph.l2_39.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(836): mode(3), layer(45): {compute_graph.Layer_40_wts_ddr.out[0], compute_graph.Layer_40_wts_ddr.out[1], compute_graph.Layer_40_l2_wts.in[0], compute_graph.Layer_40_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(836) is pipelined with KernelLayerNode(1357) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1358), layer(45): compute_graph.flexml_layers[30].compute_node[0][0], compute_graph.flexml_layers[30].compute_node[0][1], compute_graph.flexml_layers[30].compute_node[0][2], compute_graph.flexml_layers[30].compute_node[0][3], compute_graph.flexml_layers[30].compute_node[1][0], compute_graph.flexml_layers[30].compute_node[1][1], compute_graph.flexml_layers[30].compute_node[1][2], compute_graph.flexml_layers[30].compute_node[1][3], compute_graph.flexml_layers[30].compute_node[2][0], compute_graph.flexml_layers[30].compute_node[2][1], compute_graph.flexml_layers[30].compute_node[2][2], compute_graph.flexml_layers[30].compute_node[2][3], compute_graph.flexml_layers[30].compute_node[3][0], compute_graph.flexml_layers[30].compute_node[3][1], compute_graph.flexml_layers[30].compute_node[3][2], compute_graph.flexml_layers[30].compute_node[3][3], {compute_graph.l2_39.out[0], compute_graph.l2_39.out[1], compute_graph.l2_39.out[2], compute_graph.l2_39.out[3], compute_graph.Layer_40_l2_wts.out[0], compute_graph.Layer_40_l2_wts.out[1], compute_graph.Layer_40_l2_wts.out[2], compute_graph.Layer_40_l2_wts.out[3], compute_graph.l2_40.in[0], compute_graph.l2_40.in[1], compute_graph.l2_40.in[2], compute_graph.l2_40.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1359), layer(46): compute_graph.flexml_layers[31].compute_node[0][0], compute_graph.flexml_layers[31].compute_node[0][1], compute_graph.flexml_layers[31].compute_node[0][2], compute_graph.flexml_layers[31].compute_node[0][3], compute_graph.flexml_layers[31].compute_node[1][0], compute_graph.flexml_layers[31].compute_node[1][1], compute_graph.flexml_layers[31].compute_node[1][2], compute_graph.flexml_layers[31].compute_node[1][3], compute_graph.flexml_layers[31].compute_node[2][0], compute_graph.flexml_layers[31].compute_node[2][1], compute_graph.flexml_layers[31].compute_node[2][2], compute_graph.flexml_layers[31].compute_node[2][3], compute_graph.flexml_layers[31].compute_node[3][0], compute_graph.flexml_layers[31].compute_node[3][1], compute_graph.flexml_layers[31].compute_node[3][2], compute_graph.flexml_layers[31].compute_node[3][3], {compute_graph.l2_40.out[0], compute_graph.l2_40.out[1], compute_graph.l2_40.out[2], compute_graph.l2_40.out[3], compute_graph.l2_41.in[0], compute_graph.l2_41.in[1], compute_graph.l2_41.in[2], compute_graph.l2_41.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1360), layer(47): compute_graph.flexml_layers[32].compute_node[0][0], compute_graph.flexml_layers[32].compute_node[0][1], compute_graph.flexml_layers[32].compute_node[0][2], compute_graph.flexml_layers[32].compute_node[0][3], compute_graph.flexml_layers[32].compute_node[1][0], compute_graph.flexml_layers[32].compute_node[1][1], compute_graph.flexml_layers[32].compute_node[1][2], compute_graph.flexml_layers[32].compute_node[1][3], compute_graph.flexml_layers[32].compute_node[2][0], compute_graph.flexml_layers[32].compute_node[2][1], compute_graph.flexml_layers[32].compute_node[2][2], compute_graph.flexml_layers[32].compute_node[2][3], compute_graph.flexml_layers[32].compute_node[3][0], compute_graph.flexml_layers[32].compute_node[3][1], compute_graph.flexml_layers[32].compute_node[3][2], compute_graph.flexml_layers[32].compute_node[3][3], {compute_graph.l2_41.out[0], compute_graph.l2_41.out[1], compute_graph.l2_41.out[2], compute_graph.l2_41.out[3], compute_graph.l2_42.in[0], compute_graph.l2_42.in[1], compute_graph.l2_42.in[2], compute_graph.l2_42.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1361), layer(48): compute_graph.flexml_layers[33].compute_node[0][0], compute_graph.flexml_layers[33].compute_node[0][1], compute_graph.flexml_layers[33].compute_node[0][2], compute_graph.flexml_layers[33].compute_node[0][3], compute_graph.flexml_layers[33].compute_node[1][0], compute_graph.flexml_layers[33].compute_node[1][1], compute_graph.flexml_layers[33].compute_node[1][2], compute_graph.flexml_layers[33].compute_node[1][3], compute_graph.flexml_layers[33].compute_node[2][0], compute_graph.flexml_layers[33].compute_node[2][1], compute_graph.flexml_layers[33].compute_node[2][2], compute_graph.flexml_layers[33].compute_node[2][3], compute_graph.flexml_layers[33].compute_node[3][0], compute_graph.flexml_layers[33].compute_node[3][1], compute_graph.flexml_layers[33].compute_node[3][2], compute_graph.flexml_layers[33].compute_node[3][3], {compute_graph.l2_42.out[0], compute_graph.l2_42.out[1], compute_graph.l2_42.out[2], compute_graph.l2_42.out[3], compute_graph.l2_43.in[0], compute_graph.l2_43.in[1], compute_graph.l2_43.in[2], compute_graph.l2_43.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(920): mode(0), layer(48): {compute_graph.l2_43.out[1], compute_graph.l2l3_43_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1056): mode(0), layer(49): {compute_graph.l2l3_43_spill.out[0], compute_graph.templated_graph_44.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1247): mode(0), layer(49): {compute_graph.templated_graph_44.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_44_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1057): mode(0), layer(50): {compute_graph.l2l3_scratch_0_44_spill.out[0], compute_graph.templated_graph_44.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1248): mode(0), layer(50): {compute_graph.templated_graph_44.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_44_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1058): mode(0), layer(51): {compute_graph.l2l3_scratch_1_44_spill.out[0], compute_graph.templated_graph_44.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1249): mode(0), layer(51): {compute_graph.templated_graph_44.g2.ifm_mem.out[0], compute_graph.l2l3_44_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1053): mode(0), layer(52): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1059): mode(0), layer(52): {compute_graph.l2l3_44_spill.out[0], compute_graph.l2l3_44_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1362), layer(52): compute_graph.flexml_layers[34].compute_node[0][0], compute_graph.flexml_layers[34].compute_node[0][1], compute_graph.flexml_layers[34].compute_node[0][2], compute_graph.flexml_layers[34].compute_node[0][3], compute_graph.flexml_layers[34].compute_node[1][0], compute_graph.flexml_layers[34].compute_node[1][1], compute_graph.flexml_layers[34].compute_node[1][2], compute_graph.flexml_layers[34].compute_node[1][3], compute_graph.flexml_layers[34].compute_node[2][0], compute_graph.flexml_layers[34].compute_node[2][1], compute_graph.flexml_layers[34].compute_node[2][2], compute_graph.flexml_layers[34].compute_node[2][3], compute_graph.flexml_layers[34].compute_node[3][0], compute_graph.flexml_layers[34].compute_node[3][1], compute_graph.flexml_layers[34].compute_node[3][2], compute_graph.flexml_layers[34].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[3], compute_graph.l2_45.in[0], compute_graph.l2_45.in[1], compute_graph.l2_45.in[2], compute_graph.l2_45.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(837): mode(3), layer(53): {compute_graph.Layer_46_wts_ddr.out[0], compute_graph.Layer_46_wts_ddr.out[1], compute_graph.Layer_46_l2_wts.in[0], compute_graph.Layer_46_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(837) is pipelined with KernelLayerNode(1362) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434 overlaps with memory space used by compute_graph.l2_46. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1052): mode(0), layer(53): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1363), layer(53): compute_graph.flexml_layers[35].compute_node[0][0], compute_graph.flexml_layers[35].compute_node[0][1], compute_graph.flexml_layers[35].compute_node[0][2], compute_graph.flexml_layers[35].compute_node[0][3], compute_graph.flexml_layers[35].compute_node[1][0], compute_graph.flexml_layers[35].compute_node[1][1], compute_graph.flexml_layers[35].compute_node[1][2], compute_graph.flexml_layers[35].compute_node[1][3], compute_graph.flexml_layers[35].compute_node[2][0], compute_graph.flexml_layers[35].compute_node[2][1], compute_graph.flexml_layers[35].compute_node[2][2], compute_graph.flexml_layers[35].compute_node[2][3], compute_graph.flexml_layers[35].compute_node[3][0], compute_graph.flexml_layers[35].compute_node[3][1], compute_graph.flexml_layers[35].compute_node[3][2], compute_graph.flexml_layers[35].compute_node[3][3], {compute_graph.l2_45.out[0], compute_graph.l2_45.out[1], compute_graph.l2_45.out[2], compute_graph.l2_45.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[3], compute_graph.Layer_46_l2_wts.out[0], compute_graph.Layer_46_l2_wts.out[1], compute_graph.Layer_46_l2_wts.out[2], compute_graph.Layer_46_l2_wts.out[3], compute_graph.l2_46.in[0], compute_graph.l2_46.in[1], compute_graph.l2_46.in[2], compute_graph.l2_46.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(838): mode(3), layer(54): {compute_graph.Layer_47_wts_ddr.out[0], compute_graph.Layer_47_wts_ddr.out[1], compute_graph.Layer_47_l2_wts.in[0], compute_graph.Layer_47_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(838) is pipelined with KernelLayerNode(1363) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1364), layer(54): compute_graph.flexml_layers[36].compute_node[0][0], compute_graph.flexml_layers[36].compute_node[0][1], compute_graph.flexml_layers[36].compute_node[0][2], compute_graph.flexml_layers[36].compute_node[0][3], compute_graph.flexml_layers[36].compute_node[1][0], compute_graph.flexml_layers[36].compute_node[1][1], compute_graph.flexml_layers[36].compute_node[1][2], compute_graph.flexml_layers[36].compute_node[1][3], compute_graph.flexml_layers[36].compute_node[2][0], compute_graph.flexml_layers[36].compute_node[2][1], compute_graph.flexml_layers[36].compute_node[2][2], compute_graph.flexml_layers[36].compute_node[2][3], compute_graph.flexml_layers[36].compute_node[3][0], compute_graph.flexml_layers[36].compute_node[3][1], compute_graph.flexml_layers[36].compute_node[3][2], compute_graph.flexml_layers[36].compute_node[3][3], {compute_graph.l2_46.out[0], compute_graph.l2_46.out[1], compute_graph.l2_46.out[2], compute_graph.l2_46.out[3], compute_graph.Layer_47_l2_wts.out[0], compute_graph.Layer_47_l2_wts.out[1], compute_graph.Layer_47_l2_wts.out[2], compute_graph.Layer_47_l2_wts.out[3], compute_graph.l2_47.in[0], compute_graph.l2_47.in[1], compute_graph.l2_47.in[2], compute_graph.l2_47.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(921): mode(3), layer(54): {compute_graph.l2_46.out[4], compute_graph.l2_46.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(921) is pipelined with KernelLayerNode(1364) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(839): mode(3), layer(55): {compute_graph.Layer_48_wts_ddr.out[0], compute_graph.Layer_48_wts_ddr.out[1], compute_graph.Layer_48_l2_wts.in[0], compute_graph.Layer_48_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(839) is pipelined with KernelLayerNode(1364) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1365), layer(55): compute_graph.flexml_layers[37].compute_node[0][0], compute_graph.flexml_layers[37].compute_node[0][1], compute_graph.flexml_layers[37].compute_node[0][2], compute_graph.flexml_layers[37].compute_node[0][3], compute_graph.flexml_layers[37].compute_node[1][0], compute_graph.flexml_layers[37].compute_node[1][1], compute_graph.flexml_layers[37].compute_node[1][2], compute_graph.flexml_layers[37].compute_node[1][3], compute_graph.flexml_layers[37].compute_node[2][0], compute_graph.flexml_layers[37].compute_node[2][1], compute_graph.flexml_layers[37].compute_node[2][2], compute_graph.flexml_layers[37].compute_node[2][3], compute_graph.flexml_layers[37].compute_node[3][0], compute_graph.flexml_layers[37].compute_node[3][1], compute_graph.flexml_layers[37].compute_node[3][2], compute_graph.flexml_layers[37].compute_node[3][3], {compute_graph.l2_47.out[0], compute_graph.l2_47.out[1], compute_graph.l2_47.out[2], compute_graph.l2_47.out[3], compute_graph.Layer_48_l2_wts.out[0], compute_graph.Layer_48_l2_wts.out[1], compute_graph.Layer_48_l2_wts.out[2], compute_graph.Layer_48_l2_wts.out[3], compute_graph.l2_48.in[0], compute_graph.l2_48.in[1], compute_graph.l2_48.in[2], compute_graph.l2_48.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(923): mode(0), layer(55): {compute_graph.l2_48.out[2], compute_graph.l2_48.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(922): mode(0), layer(55): {compute_graph.l2_48.out[0], compute_graph.l2_48.out[1], compute_graph.l2l3_48_spill.in[0], compute_graph.l2l3_48_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1062): mode(0), layer(56): {compute_graph.l2l3_48_spill.out[0], compute_graph.templated_graph_49.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1250): mode(0), layer(56): {compute_graph.templated_graph_49.trans_mt_ifm.out[0], compute_graph.templated_graph_49.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1251): mode(0), layer(56): {compute_graph.templated_graph_49.trans_mt_ofm.out[0], compute_graph.l2l3_49_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1063): mode(0), layer(57): {compute_graph.l2l3_49_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1366), layer(57): compute_graph.flexml_layers[38].compute_node[0][0], compute_graph.flexml_layers[38].compute_node[0][1], compute_graph.flexml_layers[38].compute_node[0][2], compute_graph.flexml_layers[38].compute_node[0][3], compute_graph.flexml_layers[38].compute_node[1][0], compute_graph.flexml_layers[38].compute_node[1][1], compute_graph.flexml_layers[38].compute_node[1][2], compute_graph.flexml_layers[38].compute_node[1][3], compute_graph.flexml_layers[38].compute_node[2][0], compute_graph.flexml_layers[38].compute_node[2][1], compute_graph.flexml_layers[38].compute_node[2][2], compute_graph.flexml_layers[38].compute_node[2][3], compute_graph.flexml_layers[38].compute_node[3][0], compute_graph.flexml_layers[38].compute_node[3][1], compute_graph.flexml_layers[38].compute_node[3][2], compute_graph.flexml_layers[38].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[3], compute_graph.l2_50.in[0], compute_graph.l2_50.in[1], compute_graph.l2_50.in[2], compute_graph.l2_50.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(840): mode(3), layer(58): {compute_graph.Layer_51_wts_ddr.out[0], compute_graph.Layer_51_wts_ddr.out[1], compute_graph.Layer_51_l2_wts.in[0], compute_graph.Layer_51_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(840) is pipelined with KernelLayerNode(1366) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1367), layer(58): compute_graph.flexml_layers[39].compute_node[0][0], compute_graph.flexml_layers[39].compute_node[0][1], compute_graph.flexml_layers[39].compute_node[0][2], compute_graph.flexml_layers[39].compute_node[0][3], compute_graph.flexml_layers[39].compute_node[1][0], compute_graph.flexml_layers[39].compute_node[1][1], compute_graph.flexml_layers[39].compute_node[1][2], compute_graph.flexml_layers[39].compute_node[1][3], compute_graph.flexml_layers[39].compute_node[2][0], compute_graph.flexml_layers[39].compute_node[2][1], compute_graph.flexml_layers[39].compute_node[2][2], compute_graph.flexml_layers[39].compute_node[2][3], compute_graph.flexml_layers[39].compute_node[3][0], compute_graph.flexml_layers[39].compute_node[3][1], compute_graph.flexml_layers[39].compute_node[3][2], compute_graph.flexml_layers[39].compute_node[3][3], {compute_graph.l2_50.out[0], compute_graph.l2_50.out[1], compute_graph.l2_50.out[2], compute_graph.l2_50.out[3], compute_graph.Layer_51_l2_wts.out[0], compute_graph.Layer_51_l2_wts.out[1], compute_graph.Layer_51_l2_wts.out[2], compute_graph.Layer_51_l2_wts.out[3], compute_graph.l2_51.in[0], compute_graph.l2_51.in[1], compute_graph.l2_51.in[2], compute_graph.l2_51.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(841): mode(3), layer(59): {compute_graph.Layer_52_wts_ddr.out[0], compute_graph.Layer_52_wts_ddr.out[1], compute_graph.Layer_52_l2_wts.in[0], compute_graph.Layer_52_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(841) is pipelined with KernelLayerNode(1367) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1368), layer(59): compute_graph.flexml_layers[40].compute_node[0][0], compute_graph.flexml_layers[40].compute_node[0][1], compute_graph.flexml_layers[40].compute_node[0][2], compute_graph.flexml_layers[40].compute_node[0][3], compute_graph.flexml_layers[40].compute_node[1][0], compute_graph.flexml_layers[40].compute_node[1][1], compute_graph.flexml_layers[40].compute_node[1][2], compute_graph.flexml_layers[40].compute_node[1][3], compute_graph.flexml_layers[40].compute_node[2][0], compute_graph.flexml_layers[40].compute_node[2][1], compute_graph.flexml_layers[40].compute_node[2][2], compute_graph.flexml_layers[40].compute_node[2][3], compute_graph.flexml_layers[40].compute_node[3][0], compute_graph.flexml_layers[40].compute_node[3][1], compute_graph.flexml_layers[40].compute_node[3][2], compute_graph.flexml_layers[40].compute_node[3][3], {compute_graph.l2_51.out[0], compute_graph.l2_51.out[1], compute_graph.l2_51.out[2], compute_graph.l2_51.out[3], compute_graph.Layer_52_l2_wts.out[0], compute_graph.Layer_52_l2_wts.out[1], compute_graph.Layer_52_l2_wts.out[2], compute_graph.Layer_52_l2_wts.out[3], compute_graph.l2_52.in[0], compute_graph.l2_52.in[1], compute_graph.l2_52.in[2], compute_graph.l2_52.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1369), layer(60): compute_graph.flexml_layers[41].compute_node[0][0], compute_graph.flexml_layers[41].compute_node[0][1], compute_graph.flexml_layers[41].compute_node[0][2], compute_graph.flexml_layers[41].compute_node[0][3], compute_graph.flexml_layers[41].compute_node[1][0], compute_graph.flexml_layers[41].compute_node[1][1], compute_graph.flexml_layers[41].compute_node[1][2], compute_graph.flexml_layers[41].compute_node[1][3], compute_graph.flexml_layers[41].compute_node[2][0], compute_graph.flexml_layers[41].compute_node[2][1], compute_graph.flexml_layers[41].compute_node[2][2], compute_graph.flexml_layers[41].compute_node[2][3], compute_graph.flexml_layers[41].compute_node[3][0], compute_graph.flexml_layers[41].compute_node[3][1], compute_graph.flexml_layers[41].compute_node[3][2], compute_graph.flexml_layers[41].compute_node[3][3], {compute_graph.l2_52.out[0], compute_graph.l2_52.out[1], compute_graph.l2_52.out[2], compute_graph.l2_52.out[3], compute_graph.l2_53.in[0], compute_graph.l2_53.in[1], compute_graph.l2_53.in[2], compute_graph.l2_53.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1370), layer(61): compute_graph.flexml_layers[42].compute_node[0][0], compute_graph.flexml_layers[42].compute_node[0][1], compute_graph.flexml_layers[42].compute_node[0][2], compute_graph.flexml_layers[42].compute_node[0][3], compute_graph.flexml_layers[42].compute_node[1][0], compute_graph.flexml_layers[42].compute_node[1][1], compute_graph.flexml_layers[42].compute_node[1][2], compute_graph.flexml_layers[42].compute_node[1][3], compute_graph.flexml_layers[42].compute_node[2][0], compute_graph.flexml_layers[42].compute_node[2][1], compute_graph.flexml_layers[42].compute_node[2][2], compute_graph.flexml_layers[42].compute_node[2][3], compute_graph.flexml_layers[42].compute_node[3][0], compute_graph.flexml_layers[42].compute_node[3][1], compute_graph.flexml_layers[42].compute_node[3][2], compute_graph.flexml_layers[42].compute_node[3][3], {compute_graph.l2_53.out[0], compute_graph.l2_53.out[1], compute_graph.l2_53.out[2], compute_graph.l2_53.out[3], compute_graph.l2_54.in[0], compute_graph.l2_54.in[1], compute_graph.l2_54.in[2], compute_graph.l2_54.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1371), layer(62): compute_graph.flexml_layers[43].compute_node[0][0], compute_graph.flexml_layers[43].compute_node[0][1], compute_graph.flexml_layers[43].compute_node[0][2], compute_graph.flexml_layers[43].compute_node[0][3], compute_graph.flexml_layers[43].compute_node[1][0], compute_graph.flexml_layers[43].compute_node[1][1], compute_graph.flexml_layers[43].compute_node[1][2], compute_graph.flexml_layers[43].compute_node[1][3], compute_graph.flexml_layers[43].compute_node[2][0], compute_graph.flexml_layers[43].compute_node[2][1], compute_graph.flexml_layers[43].compute_node[2][2], compute_graph.flexml_layers[43].compute_node[2][3], compute_graph.flexml_layers[43].compute_node[3][0], compute_graph.flexml_layers[43].compute_node[3][1], compute_graph.flexml_layers[43].compute_node[3][2], compute_graph.flexml_layers[43].compute_node[3][3], {compute_graph.l2_54.out[0], compute_graph.l2_54.out[1], compute_graph.l2_54.out[2], compute_graph.l2_54.out[3], compute_graph.l2_55.in[0], compute_graph.l2_55.in[1], compute_graph.l2_55.in[2], compute_graph.l2_55.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(924): mode(0), layer(62): {compute_graph.l2_55.out[1], compute_graph.l2l3_55_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1064): mode(0), layer(63): {compute_graph.l2l3_55_spill.out[0], compute_graph.templated_graph_56.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1252): mode(0), layer(63): {compute_graph.templated_graph_56.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_56_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1065): mode(0), layer(64): {compute_graph.l2l3_scratch_0_56_spill.out[0], compute_graph.templated_graph_56.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1253): mode(0), layer(64): {compute_graph.templated_graph_56.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_56_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1066): mode(0), layer(65): {compute_graph.l2l3_scratch_1_56_spill.out[0], compute_graph.templated_graph_56.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1254): mode(0), layer(65): {compute_graph.templated_graph_56.g2.ifm_mem.out[0], compute_graph.l2l3_56_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1061): mode(0), layer(66): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1067): mode(0), layer(66): {compute_graph.l2l3_56_spill.out[0], compute_graph.l2l3_56_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1372), layer(66): compute_graph.flexml_layers[44].compute_node[0][0], compute_graph.flexml_layers[44].compute_node[0][1], compute_graph.flexml_layers[44].compute_node[0][2], compute_graph.flexml_layers[44].compute_node[0][3], compute_graph.flexml_layers[44].compute_node[1][0], compute_graph.flexml_layers[44].compute_node[1][1], compute_graph.flexml_layers[44].compute_node[1][2], compute_graph.flexml_layers[44].compute_node[1][3], compute_graph.flexml_layers[44].compute_node[2][0], compute_graph.flexml_layers[44].compute_node[2][1], compute_graph.flexml_layers[44].compute_node[2][2], compute_graph.flexml_layers[44].compute_node[2][3], compute_graph.flexml_layers[44].compute_node[3][0], compute_graph.flexml_layers[44].compute_node[3][1], compute_graph.flexml_layers[44].compute_node[3][2], compute_graph.flexml_layers[44].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[3], compute_graph.l2_57.in[0], compute_graph.l2_57.in[1], compute_graph.l2_57.in[2], compute_graph.l2_57.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(842): mode(3), layer(67): {compute_graph.Layer_58_wts_ddr.out[0], compute_graph.Layer_58_wts_ddr.out[1], compute_graph.Layer_58_l2_wts.in[0], compute_graph.Layer_58_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(842) is pipelined with KernelLayerNode(1372) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646 overlaps with memory space used by compute_graph.l2_58. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1060): mode(0), layer(67): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1373), layer(67): compute_graph.flexml_layers[45].compute_node[0][0], compute_graph.flexml_layers[45].compute_node[0][1], compute_graph.flexml_layers[45].compute_node[0][2], compute_graph.flexml_layers[45].compute_node[0][3], compute_graph.flexml_layers[45].compute_node[1][0], compute_graph.flexml_layers[45].compute_node[1][1], compute_graph.flexml_layers[45].compute_node[1][2], compute_graph.flexml_layers[45].compute_node[1][3], compute_graph.flexml_layers[45].compute_node[2][0], compute_graph.flexml_layers[45].compute_node[2][1], compute_graph.flexml_layers[45].compute_node[2][2], compute_graph.flexml_layers[45].compute_node[2][3], compute_graph.flexml_layers[45].compute_node[3][0], compute_graph.flexml_layers[45].compute_node[3][1], compute_graph.flexml_layers[45].compute_node[3][2], compute_graph.flexml_layers[45].compute_node[3][3], {compute_graph.l2_57.out[0], compute_graph.l2_57.out[1], compute_graph.l2_57.out[2], compute_graph.l2_57.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[3], compute_graph.Layer_58_l2_wts.out[0], compute_graph.Layer_58_l2_wts.out[1], compute_graph.Layer_58_l2_wts.out[2], compute_graph.Layer_58_l2_wts.out[3], compute_graph.l2_58.in[0], compute_graph.l2_58.in[1], compute_graph.l2_58.in[2], compute_graph.l2_58.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(843): mode(3), layer(68): {compute_graph.Layer_59_wts_ddr.out[0], compute_graph.Layer_59_wts_ddr.out[1], compute_graph.Layer_59_l2_wts.in[0], compute_graph.Layer_59_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(843) is pipelined with KernelLayerNode(1373) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2_58.out[0], compute_graph.l2_58.out[1], compute_graph.l2_58.out[2], compute_graph.l2_58.out[3], compute_graph.l2_59.in[0], compute_graph.l2_59.in[1], compute_graph.l2_59.in[2], compute_graph.l2_59.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1374), layer(68): compute_graph.flexml_layers[46].compute_node[0][0], compute_graph.flexml_layers[46].compute_node[0][1], compute_graph.flexml_layers[46].compute_node[0][2], compute_graph.flexml_layers[46].compute_node[0][3], compute_graph.flexml_layers[46].compute_node[1][0], compute_graph.flexml_layers[46].compute_node[1][1], compute_graph.flexml_layers[46].compute_node[1][2], compute_graph.flexml_layers[46].compute_node[1][3], compute_graph.flexml_layers[46].compute_node[2][0], compute_graph.flexml_layers[46].compute_node[2][1], compute_graph.flexml_layers[46].compute_node[2][2], compute_graph.flexml_layers[46].compute_node[2][3], compute_graph.flexml_layers[46].compute_node[3][0], compute_graph.flexml_layers[46].compute_node[3][1], compute_graph.flexml_layers[46].compute_node[3][2], compute_graph.flexml_layers[46].compute_node[3][3], {compute_graph.l2_58.out[0], compute_graph.l2_58.out[1], compute_graph.l2_58.out[2], compute_graph.l2_58.out[3], compute_graph.Layer_59_l2_wts.out[0], compute_graph.Layer_59_l2_wts.out[1], compute_graph.Layer_59_l2_wts.out[2], compute_graph.Layer_59_l2_wts.out[3], compute_graph.l2_59.in[0], compute_graph.l2_59.in[1], compute_graph.l2_59.in[2], compute_graph.l2_59.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(925): mode(3), layer(68): {compute_graph.l2_58.out[4], compute_graph.l2_58.out[5], compute_graph.spill_L3_Concat_Buffer_layer_236.in[2], compute_graph.spill_L3_Concat_Buffer_layer_236.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(925) is pipelined with KernelLayerNode(1374) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1375), layer(69): compute_graph.flexml_layers[47].compute_node[0][0], compute_graph.flexml_layers[47].compute_node[0][1], compute_graph.flexml_layers[47].compute_node[0][2], compute_graph.flexml_layers[47].compute_node[0][3], compute_graph.flexml_layers[47].compute_node[1][0], compute_graph.flexml_layers[47].compute_node[1][1], compute_graph.flexml_layers[47].compute_node[1][2], compute_graph.flexml_layers[47].compute_node[1][3], compute_graph.flexml_layers[47].compute_node[2][0], compute_graph.flexml_layers[47].compute_node[2][1], compute_graph.flexml_layers[47].compute_node[2][2], compute_graph.flexml_layers[47].compute_node[2][3], compute_graph.flexml_layers[47].compute_node[3][0], compute_graph.flexml_layers[47].compute_node[3][1], compute_graph.flexml_layers[47].compute_node[3][2], compute_graph.flexml_layers[47].compute_node[3][3], {compute_graph.l2_59.out[0], compute_graph.l2_59.out[1], compute_graph.l2_59.out[2], compute_graph.l2_59.out[3], compute_graph.l2_60.in[0], compute_graph.l2_60.in[1], compute_graph.l2_60.in[2], compute_graph.l2_60.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(926): mode(3), layer(69): {compute_graph.l2_59.out[4], compute_graph.l2_59.out[5], compute_graph.l2l3_59_spill.in[0], compute_graph.l2l3_59_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(926) is pipelined with KernelLayerNode(1375) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1376), layer(70): compute_graph.flexml_layers[48].compute_node[0][0], compute_graph.flexml_layers[48].compute_node[0][1], compute_graph.flexml_layers[48].compute_node[0][2], compute_graph.flexml_layers[48].compute_node[0][3], compute_graph.flexml_layers[48].compute_node[1][0], compute_graph.flexml_layers[48].compute_node[1][1], compute_graph.flexml_layers[48].compute_node[1][2], compute_graph.flexml_layers[48].compute_node[1][3], compute_graph.flexml_layers[48].compute_node[2][0], compute_graph.flexml_layers[48].compute_node[2][1], compute_graph.flexml_layers[48].compute_node[2][2], compute_graph.flexml_layers[48].compute_node[2][3], compute_graph.flexml_layers[48].compute_node[3][0], compute_graph.flexml_layers[48].compute_node[3][1], compute_graph.flexml_layers[48].compute_node[3][2], compute_graph.flexml_layers[48].compute_node[3][3], {compute_graph.l2_60.out[0], compute_graph.l2_60.out[1], compute_graph.l2_60.out[2], compute_graph.l2_60.out[3], compute_graph.l2_61.in[0], compute_graph.l2_61.in[1], compute_graph.l2_61.in[2], compute_graph.l2_61.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1377), layer(71): compute_graph.flexml_layers[49].compute_node[0][0], compute_graph.flexml_layers[49].compute_node[0][1], compute_graph.flexml_layers[49].compute_node[0][2], compute_graph.flexml_layers[49].compute_node[0][3], compute_graph.flexml_layers[49].compute_node[1][0], compute_graph.flexml_layers[49].compute_node[1][1], compute_graph.flexml_layers[49].compute_node[1][2], compute_graph.flexml_layers[49].compute_node[1][3], compute_graph.flexml_layers[49].compute_node[2][0], compute_graph.flexml_layers[49].compute_node[2][1], compute_graph.flexml_layers[49].compute_node[2][2], compute_graph.flexml_layers[49].compute_node[2][3], compute_graph.flexml_layers[49].compute_node[3][0], compute_graph.flexml_layers[49].compute_node[3][1], compute_graph.flexml_layers[49].compute_node[3][2], compute_graph.flexml_layers[49].compute_node[3][3], {compute_graph.l2_61.out[0], compute_graph.l2_61.out[1], compute_graph.l2_61.out[2], compute_graph.l2_61.out[3], compute_graph.l2_62.in[0], compute_graph.l2_62.in[1], compute_graph.l2_62.in[2], compute_graph.l2_62.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(927): mode(0), layer(71): {compute_graph.l2_62.out[0], compute_graph.l2_62.out[1], compute_graph.l2l3_62_spill.in[0], compute_graph.l2l3_62_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 928 will not be pipelined because it's in the same layer 72 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1585), layer(72): {compute_graph.l2l3_59_spill.out[0], compute_graph.l2l3_62_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1586), layer(72): {compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1587), layer(72): {compute_graph.l2l3_59_spill.out[1], compute_graph.l2l3_62_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1588), layer(72): {compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[3], compute_graph.l2_63.in[0], compute_graph.l2_63.in[1], compute_graph.l2_63.in[2], compute_graph.l2_63.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1378), layer(72): compute_graph.flexml_layers[50].compute_node[0][0], compute_graph.flexml_layers[50].compute_node[0][1], compute_graph.flexml_layers[50].compute_node[0][2], compute_graph.flexml_layers[50].compute_node[0][3], compute_graph.flexml_layers[50].compute_node[1][0], compute_graph.flexml_layers[50].compute_node[1][1], compute_graph.flexml_layers[50].compute_node[1][2], compute_graph.flexml_layers[50].compute_node[1][3], compute_graph.flexml_layers[50].compute_node[2][0], compute_graph.flexml_layers[50].compute_node[2][1], compute_graph.flexml_layers[50].compute_node[2][2], compute_graph.flexml_layers[50].compute_node[2][3], compute_graph.flexml_layers[50].compute_node[3][0], compute_graph.flexml_layers[50].compute_node[3][1], compute_graph.flexml_layers[50].compute_node[3][2], compute_graph.flexml_layers[50].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[3], compute_graph.l2_63.in[0], compute_graph.l2_63.in[1], compute_graph.l2_63.in[2], compute_graph.l2_63.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(928): mode(3), layer(72): {compute_graph.l2_63.out[0], compute_graph.l2_63.out[1], compute_graph.l2l3_63_spill.in[0], compute_graph.l2l3_63_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1070): mode(0), layer(73): {compute_graph.l2l3_63_spill.out[0], compute_graph.templated_graph_64.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1255): mode(0), layer(73): {compute_graph.templated_graph_64.ifm.out[0], compute_graph.templated_graph_64.ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1256): mode(0), layer(73): {compute_graph.templated_graph_64.ofm.out[0], compute_graph.l2l3_scratch_0_64_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1071): mode(0), layer(74): {compute_graph.l2l3_scratch_0_64_spill.out[0], compute_graph.templated_graph_64.ifm2.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1257): mode(0), layer(74): {compute_graph.templated_graph_64.ifm2.out[0], compute_graph.l2l3_64_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1072): mode(0), layer(75): {compute_graph.l2l3_64_spill.out[0], compute_graph.l2l3_64_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(844): mode(4), layer(75): {compute_graph.Layer_65_wts_ddr.out[0], compute_graph.Layer_65_wts_ddr.out[1], compute_graph.Layer_65_l2_wts.in[0], compute_graph.Layer_65_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(844) is pipelined with BufferToBufferNode(1072) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1379), layer(75): compute_graph.flexml_layers[51].compute_node[0][0], compute_graph.flexml_layers[51].compute_node[0][1], compute_graph.flexml_layers[51].compute_node[0][2], compute_graph.flexml_layers[51].compute_node[0][3], compute_graph.flexml_layers[51].compute_node[1][0], compute_graph.flexml_layers[51].compute_node[1][1], compute_graph.flexml_layers[51].compute_node[1][2], compute_graph.flexml_layers[51].compute_node[1][3], compute_graph.flexml_layers[51].compute_node[2][0], compute_graph.flexml_layers[51].compute_node[2][1], compute_graph.flexml_layers[51].compute_node[2][2], compute_graph.flexml_layers[51].compute_node[2][3], compute_graph.flexml_layers[51].compute_node[3][0], compute_graph.flexml_layers[51].compute_node[3][1], compute_graph.flexml_layers[51].compute_node[3][2], compute_graph.flexml_layers[51].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[3], compute_graph.Layer_65_l2_wts.out[0], compute_graph.Layer_65_l2_wts.out[1], compute_graph.Layer_65_l2_wts.out[2], compute_graph.Layer_65_l2_wts.out[3], compute_graph.l2_65.in[0], compute_graph.l2_65.in[1], compute_graph.l2_65.in[2], compute_graph.l2_65.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1380), layer(76): compute_graph.flexml_layers[52].compute_node[0][0], compute_graph.flexml_layers[52].compute_node[0][1], compute_graph.flexml_layers[52].compute_node[0][2], compute_graph.flexml_layers[52].compute_node[0][3], compute_graph.flexml_layers[52].compute_node[1][0], compute_graph.flexml_layers[52].compute_node[1][1], compute_graph.flexml_layers[52].compute_node[1][2], compute_graph.flexml_layers[52].compute_node[1][3], compute_graph.flexml_layers[52].compute_node[2][0], compute_graph.flexml_layers[52].compute_node[2][1], compute_graph.flexml_layers[52].compute_node[2][2], compute_graph.flexml_layers[52].compute_node[2][3], compute_graph.flexml_layers[52].compute_node[3][0], compute_graph.flexml_layers[52].compute_node[3][1], compute_graph.flexml_layers[52].compute_node[3][2], compute_graph.flexml_layers[52].compute_node[3][3], {compute_graph.l2_65.out[0], compute_graph.l2_65.out[1], compute_graph.l2_65.out[2], compute_graph.l2_65.out[3], compute_graph.l2_66.in[0], compute_graph.l2_66.in[1], compute_graph.l2_66.in[2], compute_graph.l2_66.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1381), layer(77): compute_graph.flexml_layers[53].compute_node[0][0], compute_graph.flexml_layers[53].compute_node[0][1], compute_graph.flexml_layers[53].compute_node[0][2], compute_graph.flexml_layers[53].compute_node[0][3], compute_graph.flexml_layers[53].compute_node[1][0], compute_graph.flexml_layers[53].compute_node[1][1], compute_graph.flexml_layers[53].compute_node[1][2], compute_graph.flexml_layers[53].compute_node[1][3], compute_graph.flexml_layers[53].compute_node[2][0], compute_graph.flexml_layers[53].compute_node[2][1], compute_graph.flexml_layers[53].compute_node[2][2], compute_graph.flexml_layers[53].compute_node[2][3], compute_graph.flexml_layers[53].compute_node[3][0], compute_graph.flexml_layers[53].compute_node[3][1], compute_graph.flexml_layers[53].compute_node[3][2], compute_graph.flexml_layers[53].compute_node[3][3], {compute_graph.l2_66.out[0], compute_graph.l2_66.out[1], compute_graph.l2_66.out[2], compute_graph.l2_66.out[3], compute_graph.l2_67.in[0], compute_graph.l2_67.in[1], compute_graph.l2_67.in[2], compute_graph.l2_67.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1382), layer(78): compute_graph.flexml_layers[54].compute_node[0][0], compute_graph.flexml_layers[54].compute_node[0][1], compute_graph.flexml_layers[54].compute_node[0][2], compute_graph.flexml_layers[54].compute_node[0][3], compute_graph.flexml_layers[54].compute_node[1][0], compute_graph.flexml_layers[54].compute_node[1][1], compute_graph.flexml_layers[54].compute_node[1][2], compute_graph.flexml_layers[54].compute_node[1][3], compute_graph.flexml_layers[54].compute_node[2][0], compute_graph.flexml_layers[54].compute_node[2][1], compute_graph.flexml_layers[54].compute_node[2][2], compute_graph.flexml_layers[54].compute_node[2][3], compute_graph.flexml_layers[54].compute_node[3][0], compute_graph.flexml_layers[54].compute_node[3][1], compute_graph.flexml_layers[54].compute_node[3][2], compute_graph.flexml_layers[54].compute_node[3][3], {compute_graph.l2_67.out[0], compute_graph.l2_67.out[1], compute_graph.l2_67.out[2], compute_graph.l2_67.out[3], compute_graph.l2_68.in[0], compute_graph.l2_68.in[1], compute_graph.l2_68.in[2], compute_graph.l2_68.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1383), layer(79): compute_graph.flexml_layers[55].compute_node[0][0], compute_graph.flexml_layers[55].compute_node[0][1], compute_graph.flexml_layers[55].compute_node[0][2], compute_graph.flexml_layers[55].compute_node[0][3], compute_graph.flexml_layers[55].compute_node[1][0], compute_graph.flexml_layers[55].compute_node[1][1], compute_graph.flexml_layers[55].compute_node[1][2], compute_graph.flexml_layers[55].compute_node[1][3], compute_graph.flexml_layers[55].compute_node[2][0], compute_graph.flexml_layers[55].compute_node[2][1], compute_graph.flexml_layers[55].compute_node[2][2], compute_graph.flexml_layers[55].compute_node[2][3], compute_graph.flexml_layers[55].compute_node[3][0], compute_graph.flexml_layers[55].compute_node[3][1], compute_graph.flexml_layers[55].compute_node[3][2], compute_graph.flexml_layers[55].compute_node[3][3], {compute_graph.l2_65.out[4], compute_graph.l2_65.out[5], compute_graph.l2_65.out[6], compute_graph.l2_65.out[7], compute_graph.l2_68.out[0], compute_graph.l2_68.out[1], compute_graph.l2_68.out[2], compute_graph.l2_68.out[3], compute_graph.l2_69.in[0], compute_graph.l2_69.in[1], compute_graph.l2_69.in[2], compute_graph.l2_69.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(845): mode(3), layer(80): {compute_graph.Layer_70_wts_ddr.out[0], compute_graph.Layer_70_wts_ddr.out[1], compute_graph.Layer_70_l2_wts.in[0], compute_graph.Layer_70_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(845) is pipelined with KernelLayerNode(1383) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1384), layer(80): compute_graph.flexml_layers[56].compute_node[0][0], compute_graph.flexml_layers[56].compute_node[0][1], compute_graph.flexml_layers[56].compute_node[0][2], compute_graph.flexml_layers[56].compute_node[0][3], compute_graph.flexml_layers[56].compute_node[1][0], compute_graph.flexml_layers[56].compute_node[1][1], compute_graph.flexml_layers[56].compute_node[1][2], compute_graph.flexml_layers[56].compute_node[1][3], compute_graph.flexml_layers[56].compute_node[2][0], compute_graph.flexml_layers[56].compute_node[2][1], compute_graph.flexml_layers[56].compute_node[2][2], compute_graph.flexml_layers[56].compute_node[2][3], compute_graph.flexml_layers[56].compute_node[3][0], compute_graph.flexml_layers[56].compute_node[3][1], compute_graph.flexml_layers[56].compute_node[3][2], compute_graph.flexml_layers[56].compute_node[3][3], {compute_graph.l2_69.out[0], compute_graph.l2_69.out[1], compute_graph.l2_69.out[2], compute_graph.l2_69.out[3], compute_graph.Layer_70_l2_wts.out[0], compute_graph.Layer_70_l2_wts.out[1], compute_graph.Layer_70_l2_wts.out[2], compute_graph.Layer_70_l2_wts.out[3], compute_graph.l2_70.in[0], compute_graph.l2_70.in[1], compute_graph.l2_70.in[2], compute_graph.l2_70.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(846): mode(3), layer(81): {compute_graph.Layer_71_wts_ddr.out[0], compute_graph.Layer_71_wts_ddr.out[1], compute_graph.Layer_71_l2_wts.in[0], compute_graph.Layer_71_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(846) is pipelined with KernelLayerNode(1384) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1385), layer(81): compute_graph.flexml_layers[57].compute_node[0][0], compute_graph.flexml_layers[57].compute_node[0][1], compute_graph.flexml_layers[57].compute_node[0][2], compute_graph.flexml_layers[57].compute_node[0][3], compute_graph.flexml_layers[57].compute_node[1][0], compute_graph.flexml_layers[57].compute_node[1][1], compute_graph.flexml_layers[57].compute_node[1][2], compute_graph.flexml_layers[57].compute_node[1][3], compute_graph.flexml_layers[57].compute_node[2][0], compute_graph.flexml_layers[57].compute_node[2][1], compute_graph.flexml_layers[57].compute_node[2][2], compute_graph.flexml_layers[57].compute_node[2][3], compute_graph.flexml_layers[57].compute_node[3][0], compute_graph.flexml_layers[57].compute_node[3][1], compute_graph.flexml_layers[57].compute_node[3][2], compute_graph.flexml_layers[57].compute_node[3][3], {compute_graph.l2_70.out[0], compute_graph.l2_70.out[1], compute_graph.l2_70.out[2], compute_graph.l2_70.out[3], compute_graph.Layer_71_l2_wts.out[0], compute_graph.Layer_71_l2_wts.out[1], compute_graph.Layer_71_l2_wts.out[2], compute_graph.Layer_71_l2_wts.out[3], compute_graph.l2_71.in[0], compute_graph.l2_71.in[1], compute_graph.l2_71.in[2], compute_graph.l2_71.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1386), layer(82): compute_graph.flexml_layers[58].compute_node[0][0], compute_graph.flexml_layers[58].compute_node[0][1], compute_graph.flexml_layers[58].compute_node[0][2], compute_graph.flexml_layers[58].compute_node[0][3], compute_graph.flexml_layers[58].compute_node[1][0], compute_graph.flexml_layers[58].compute_node[1][1], compute_graph.flexml_layers[58].compute_node[1][2], compute_graph.flexml_layers[58].compute_node[1][3], compute_graph.flexml_layers[58].compute_node[2][0], compute_graph.flexml_layers[58].compute_node[2][1], compute_graph.flexml_layers[58].compute_node[2][2], compute_graph.flexml_layers[58].compute_node[2][3], compute_graph.flexml_layers[58].compute_node[3][0], compute_graph.flexml_layers[58].compute_node[3][1], compute_graph.flexml_layers[58].compute_node[3][2], compute_graph.flexml_layers[58].compute_node[3][3], {compute_graph.l2_71.out[0], compute_graph.l2_71.out[1], compute_graph.l2_71.out[2], compute_graph.l2_71.out[3], compute_graph.l2_72.in[0], compute_graph.l2_72.in[1], compute_graph.l2_72.in[2], compute_graph.l2_72.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1387), layer(83): compute_graph.flexml_layers[59].compute_node[0][0], compute_graph.flexml_layers[59].compute_node[0][1], compute_graph.flexml_layers[59].compute_node[0][2], compute_graph.flexml_layers[59].compute_node[0][3], compute_graph.flexml_layers[59].compute_node[1][0], compute_graph.flexml_layers[59].compute_node[1][1], compute_graph.flexml_layers[59].compute_node[1][2], compute_graph.flexml_layers[59].compute_node[1][3], compute_graph.flexml_layers[59].compute_node[2][0], compute_graph.flexml_layers[59].compute_node[2][1], compute_graph.flexml_layers[59].compute_node[2][2], compute_graph.flexml_layers[59].compute_node[2][3], compute_graph.flexml_layers[59].compute_node[3][0], compute_graph.flexml_layers[59].compute_node[3][1], compute_graph.flexml_layers[59].compute_node[3][2], compute_graph.flexml_layers[59].compute_node[3][3], {compute_graph.l2_72.out[0], compute_graph.l2_72.out[1], compute_graph.l2_72.out[2], compute_graph.l2_72.out[3], compute_graph.l2_73.in[0], compute_graph.l2_73.in[1], compute_graph.l2_73.in[2], compute_graph.l2_73.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1388), layer(84): compute_graph.flexml_layers[60].compute_node[0][0], compute_graph.flexml_layers[60].compute_node[0][1], compute_graph.flexml_layers[60].compute_node[0][2], compute_graph.flexml_layers[60].compute_node[0][3], compute_graph.flexml_layers[60].compute_node[1][0], compute_graph.flexml_layers[60].compute_node[1][1], compute_graph.flexml_layers[60].compute_node[1][2], compute_graph.flexml_layers[60].compute_node[1][3], compute_graph.flexml_layers[60].compute_node[2][0], compute_graph.flexml_layers[60].compute_node[2][1], compute_graph.flexml_layers[60].compute_node[2][2], compute_graph.flexml_layers[60].compute_node[2][3], compute_graph.flexml_layers[60].compute_node[3][0], compute_graph.flexml_layers[60].compute_node[3][1], compute_graph.flexml_layers[60].compute_node[3][2], compute_graph.flexml_layers[60].compute_node[3][3], {compute_graph.l2_73.out[0], compute_graph.l2_73.out[1], compute_graph.l2_73.out[2], compute_graph.l2_73.out[3], compute_graph.l2_74.in[0], compute_graph.l2_74.in[1], compute_graph.l2_74.in[2], compute_graph.l2_74.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1389), layer(85): compute_graph.flexml_layers[61].compute_node[0][0], compute_graph.flexml_layers[61].compute_node[0][1], compute_graph.flexml_layers[61].compute_node[0][2], compute_graph.flexml_layers[61].compute_node[0][3], compute_graph.flexml_layers[61].compute_node[1][0], compute_graph.flexml_layers[61].compute_node[1][1], compute_graph.flexml_layers[61].compute_node[1][2], compute_graph.flexml_layers[61].compute_node[1][3], compute_graph.flexml_layers[61].compute_node[2][0], compute_graph.flexml_layers[61].compute_node[2][1], compute_graph.flexml_layers[61].compute_node[2][2], compute_graph.flexml_layers[61].compute_node[2][3], compute_graph.flexml_layers[61].compute_node[3][0], compute_graph.flexml_layers[61].compute_node[3][1], compute_graph.flexml_layers[61].compute_node[3][2], compute_graph.flexml_layers[61].compute_node[3][3], {compute_graph.l2_71.out[4], compute_graph.l2_71.out[5], compute_graph.l2_71.out[6], compute_graph.l2_71.out[7], compute_graph.l2_74.out[0], compute_graph.l2_74.out[1], compute_graph.l2_74.out[2], compute_graph.l2_74.out[3], compute_graph.l2_75.in[0], compute_graph.l2_75.in[1], compute_graph.l2_75.in[2], compute_graph.l2_75.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(847): mode(3), layer(86): {compute_graph.Layer_76_wts_ddr.out[0], compute_graph.Layer_76_wts_ddr.out[1], compute_graph.Layer_76_l2_wts.in[0], compute_graph.Layer_76_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(847) is pipelined with KernelLayerNode(1389) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1390), layer(86): compute_graph.flexml_layers[62].compute_node[0][0], compute_graph.flexml_layers[62].compute_node[0][1], compute_graph.flexml_layers[62].compute_node[0][2], compute_graph.flexml_layers[62].compute_node[0][3], compute_graph.flexml_layers[62].compute_node[1][0], compute_graph.flexml_layers[62].compute_node[1][1], compute_graph.flexml_layers[62].compute_node[1][2], compute_graph.flexml_layers[62].compute_node[1][3], compute_graph.flexml_layers[62].compute_node[2][0], compute_graph.flexml_layers[62].compute_node[2][1], compute_graph.flexml_layers[62].compute_node[2][2], compute_graph.flexml_layers[62].compute_node[2][3], compute_graph.flexml_layers[62].compute_node[3][0], compute_graph.flexml_layers[62].compute_node[3][1], compute_graph.flexml_layers[62].compute_node[3][2], compute_graph.flexml_layers[62].compute_node[3][3], {compute_graph.l2_75.out[0], compute_graph.l2_75.out[1], compute_graph.l2_75.out[2], compute_graph.l2_75.out[3], compute_graph.Layer_76_l2_wts.out[0], compute_graph.Layer_76_l2_wts.out[1], compute_graph.Layer_76_l2_wts.out[2], compute_graph.Layer_76_l2_wts.out[3], compute_graph.l2_76.in[0], compute_graph.l2_76.in[1], compute_graph.l2_76.in[2], compute_graph.l2_76.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1391), layer(87): compute_graph.flexml_layers[63].compute_node[0][0], compute_graph.flexml_layers[63].compute_node[0][1], compute_graph.flexml_layers[63].compute_node[0][2], compute_graph.flexml_layers[63].compute_node[0][3], compute_graph.flexml_layers[63].compute_node[1][0], compute_graph.flexml_layers[63].compute_node[1][1], compute_graph.flexml_layers[63].compute_node[1][2], compute_graph.flexml_layers[63].compute_node[1][3], compute_graph.flexml_layers[63].compute_node[2][0], compute_graph.flexml_layers[63].compute_node[2][1], compute_graph.flexml_layers[63].compute_node[2][2], compute_graph.flexml_layers[63].compute_node[2][3], compute_graph.flexml_layers[63].compute_node[3][0], compute_graph.flexml_layers[63].compute_node[3][1], compute_graph.flexml_layers[63].compute_node[3][2], compute_graph.flexml_layers[63].compute_node[3][3], {compute_graph.l2_76.out[0], compute_graph.l2_76.out[1], compute_graph.l2_76.out[2], compute_graph.l2_76.out[3], compute_graph.l2_77.in[0], compute_graph.l2_77.in[1], compute_graph.l2_77.in[2], compute_graph.l2_77.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1392), layer(88): compute_graph.flexml_layers[64].compute_node[0][0], compute_graph.flexml_layers[64].compute_node[0][1], compute_graph.flexml_layers[64].compute_node[0][2], compute_graph.flexml_layers[64].compute_node[0][3], compute_graph.flexml_layers[64].compute_node[1][0], compute_graph.flexml_layers[64].compute_node[1][1], compute_graph.flexml_layers[64].compute_node[1][2], compute_graph.flexml_layers[64].compute_node[1][3], compute_graph.flexml_layers[64].compute_node[2][0], compute_graph.flexml_layers[64].compute_node[2][1], compute_graph.flexml_layers[64].compute_node[2][2], compute_graph.flexml_layers[64].compute_node[2][3], compute_graph.flexml_layers[64].compute_node[3][0], compute_graph.flexml_layers[64].compute_node[3][1], compute_graph.flexml_layers[64].compute_node[3][2], compute_graph.flexml_layers[64].compute_node[3][3], {compute_graph.l2_77.out[0], compute_graph.l2_77.out[1], compute_graph.l2_77.out[2], compute_graph.l2_77.out[3], compute_graph.l2_78.in[0], compute_graph.l2_78.in[1], compute_graph.l2_78.in[2], compute_graph.l2_78.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1393), layer(89): compute_graph.flexml_layers[65].compute_node[0][0], compute_graph.flexml_layers[65].compute_node[0][1], compute_graph.flexml_layers[65].compute_node[0][2], compute_graph.flexml_layers[65].compute_node[0][3], compute_graph.flexml_layers[65].compute_node[1][0], compute_graph.flexml_layers[65].compute_node[1][1], compute_graph.flexml_layers[65].compute_node[1][2], compute_graph.flexml_layers[65].compute_node[1][3], compute_graph.flexml_layers[65].compute_node[2][0], compute_graph.flexml_layers[65].compute_node[2][1], compute_graph.flexml_layers[65].compute_node[2][2], compute_graph.flexml_layers[65].compute_node[2][3], compute_graph.flexml_layers[65].compute_node[3][0], compute_graph.flexml_layers[65].compute_node[3][1], compute_graph.flexml_layers[65].compute_node[3][2], compute_graph.flexml_layers[65].compute_node[3][3], {compute_graph.l2_78.out[0], compute_graph.l2_78.out[1], compute_graph.l2_78.out[2], compute_graph.l2_78.out[3], compute_graph.l2_79.in[0], compute_graph.l2_79.in[1], compute_graph.l2_79.in[2], compute_graph.l2_79.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1394), layer(90): compute_graph.flexml_layers[66].compute_node[0][0], compute_graph.flexml_layers[66].compute_node[0][1], compute_graph.flexml_layers[66].compute_node[0][2], compute_graph.flexml_layers[66].compute_node[0][3], compute_graph.flexml_layers[66].compute_node[1][0], compute_graph.flexml_layers[66].compute_node[1][1], compute_graph.flexml_layers[66].compute_node[1][2], compute_graph.flexml_layers[66].compute_node[1][3], compute_graph.flexml_layers[66].compute_node[2][0], compute_graph.flexml_layers[66].compute_node[2][1], compute_graph.flexml_layers[66].compute_node[2][2], compute_graph.flexml_layers[66].compute_node[2][3], compute_graph.flexml_layers[66].compute_node[3][0], compute_graph.flexml_layers[66].compute_node[3][1], compute_graph.flexml_layers[66].compute_node[3][2], compute_graph.flexml_layers[66].compute_node[3][3], {compute_graph.l2_76.out[4], compute_graph.l2_76.out[5], compute_graph.l2_76.out[6], compute_graph.l2_76.out[7], compute_graph.l2_79.out[0], compute_graph.l2_79.out[1], compute_graph.l2_79.out[2], compute_graph.l2_79.out[3], compute_graph.l2_80.in[0], compute_graph.l2_80.in[1], compute_graph.l2_80.in[2], compute_graph.l2_80.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(848): mode(3), layer(91): {compute_graph.Layer_81_wts_ddr.out[0], compute_graph.Layer_81_wts_ddr.out[1], compute_graph.Layer_81_l2_wts.in[0], compute_graph.Layer_81_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(848) is pipelined with KernelLayerNode(1394) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.l2_70 overlaps with memory space used by compute_graph.l2_81. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For KernelLayerNode(1395), layer(91): compute_graph.flexml_layers[67].compute_node[0][0], compute_graph.flexml_layers[67].compute_node[0][1], compute_graph.flexml_layers[67].compute_node[0][2], compute_graph.flexml_layers[67].compute_node[0][3], compute_graph.flexml_layers[67].compute_node[1][0], compute_graph.flexml_layers[67].compute_node[1][1], compute_graph.flexml_layers[67].compute_node[1][2], compute_graph.flexml_layers[67].compute_node[1][3], compute_graph.flexml_layers[67].compute_node[2][0], compute_graph.flexml_layers[67].compute_node[2][1], compute_graph.flexml_layers[67].compute_node[2][2], compute_graph.flexml_layers[67].compute_node[2][3], compute_graph.flexml_layers[67].compute_node[3][0], compute_graph.flexml_layers[67].compute_node[3][1], compute_graph.flexml_layers[67].compute_node[3][2], compute_graph.flexml_layers[67].compute_node[3][3], {compute_graph.l2_70.out[4], compute_graph.l2_70.out[5], compute_graph.l2_70.out[6], compute_graph.l2_70.out[7], compute_graph.l2_80.out[0], compute_graph.l2_80.out[1], compute_graph.l2_80.out[2], compute_graph.l2_80.out[3], compute_graph.Layer_81_l2_wts.out[0], compute_graph.Layer_81_l2_wts.out[1], compute_graph.Layer_81_l2_wts.out[2], compute_graph.Layer_81_l2_wts.out[3], compute_graph.l2_81.in[0], compute_graph.l2_81.in[1], compute_graph.l2_81.in[2], compute_graph.l2_81.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(849): mode(3), layer(92): {compute_graph.Layer_82_wts_ddr.out[0], compute_graph.Layer_82_wts_ddr.out[1], compute_graph.Layer_82_l2_wts.in[0], compute_graph.Layer_82_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(849) is pipelined with KernelLayerNode(1395) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1396), layer(92): compute_graph.flexml_layers[68].compute_node[0][0], compute_graph.flexml_layers[68].compute_node[0][1], compute_graph.flexml_layers[68].compute_node[0][2], compute_graph.flexml_layers[68].compute_node[0][3], compute_graph.flexml_layers[68].compute_node[1][0], compute_graph.flexml_layers[68].compute_node[1][1], compute_graph.flexml_layers[68].compute_node[1][2], compute_graph.flexml_layers[68].compute_node[1][3], compute_graph.flexml_layers[68].compute_node[2][0], compute_graph.flexml_layers[68].compute_node[2][1], compute_graph.flexml_layers[68].compute_node[2][2], compute_graph.flexml_layers[68].compute_node[2][3], compute_graph.flexml_layers[68].compute_node[3][0], compute_graph.flexml_layers[68].compute_node[3][1], compute_graph.flexml_layers[68].compute_node[3][2], compute_graph.flexml_layers[68].compute_node[3][3], {compute_graph.l2_81.out[0], compute_graph.l2_81.out[1], compute_graph.l2_81.out[2], compute_graph.l2_81.out[3], compute_graph.Layer_82_l2_wts.out[0], compute_graph.Layer_82_l2_wts.out[1], compute_graph.Layer_82_l2_wts.out[2], compute_graph.Layer_82_l2_wts.out[3], compute_graph.l2_82.in[0], compute_graph.l2_82.in[1], compute_graph.l2_82.in[2], compute_graph.l2_82.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1397), layer(93): compute_graph.flexml_layers[69].compute_node[0][0], compute_graph.flexml_layers[69].compute_node[0][1], compute_graph.flexml_layers[69].compute_node[0][2], compute_graph.flexml_layers[69].compute_node[0][3], compute_graph.flexml_layers[69].compute_node[1][0], compute_graph.flexml_layers[69].compute_node[1][1], compute_graph.flexml_layers[69].compute_node[1][2], compute_graph.flexml_layers[69].compute_node[1][3], compute_graph.flexml_layers[69].compute_node[2][0], compute_graph.flexml_layers[69].compute_node[2][1], compute_graph.flexml_layers[69].compute_node[2][2], compute_graph.flexml_layers[69].compute_node[2][3], compute_graph.flexml_layers[69].compute_node[3][0], compute_graph.flexml_layers[69].compute_node[3][1], compute_graph.flexml_layers[69].compute_node[3][2], compute_graph.flexml_layers[69].compute_node[3][3], {compute_graph.l2_82.out[0], compute_graph.l2_82.out[1], compute_graph.l2_82.out[2], compute_graph.l2_82.out[3], compute_graph.l2_83.in[0], compute_graph.l2_83.in[1], compute_graph.l2_83.in[2], compute_graph.l2_83.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1398), layer(94): compute_graph.flexml_layers[70].compute_node[0][0], compute_graph.flexml_layers[70].compute_node[0][1], compute_graph.flexml_layers[70].compute_node[0][2], compute_graph.flexml_layers[70].compute_node[0][3], compute_graph.flexml_layers[70].compute_node[1][0], compute_graph.flexml_layers[70].compute_node[1][1], compute_graph.flexml_layers[70].compute_node[1][2], compute_graph.flexml_layers[70].compute_node[1][3], compute_graph.flexml_layers[70].compute_node[2][0], compute_graph.flexml_layers[70].compute_node[2][1], compute_graph.flexml_layers[70].compute_node[2][2], compute_graph.flexml_layers[70].compute_node[2][3], compute_graph.flexml_layers[70].compute_node[3][0], compute_graph.flexml_layers[70].compute_node[3][1], compute_graph.flexml_layers[70].compute_node[3][2], compute_graph.flexml_layers[70].compute_node[3][3], {compute_graph.l2_83.out[0], compute_graph.l2_83.out[1], compute_graph.l2_83.out[2], compute_graph.l2_83.out[3], compute_graph.l2_84.in[0], compute_graph.l2_84.in[1], compute_graph.l2_84.in[2], compute_graph.l2_84.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1399), layer(95): compute_graph.flexml_layers[71].compute_node[0][0], compute_graph.flexml_layers[71].compute_node[0][1], compute_graph.flexml_layers[71].compute_node[0][2], compute_graph.flexml_layers[71].compute_node[0][3], compute_graph.flexml_layers[71].compute_node[1][0], compute_graph.flexml_layers[71].compute_node[1][1], compute_graph.flexml_layers[71].compute_node[1][2], compute_graph.flexml_layers[71].compute_node[1][3], compute_graph.flexml_layers[71].compute_node[2][0], compute_graph.flexml_layers[71].compute_node[2][1], compute_graph.flexml_layers[71].compute_node[2][2], compute_graph.flexml_layers[71].compute_node[2][3], compute_graph.flexml_layers[71].compute_node[3][0], compute_graph.flexml_layers[71].compute_node[3][1], compute_graph.flexml_layers[71].compute_node[3][2], compute_graph.flexml_layers[71].compute_node[3][3], {compute_graph.l2_84.out[0], compute_graph.l2_84.out[1], compute_graph.l2_84.out[2], compute_graph.l2_84.out[3], compute_graph.l2_85.in[0], compute_graph.l2_85.in[1], compute_graph.l2_85.in[2], compute_graph.l2_85.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1400), layer(96): compute_graph.flexml_layers[72].compute_node[0][0], compute_graph.flexml_layers[72].compute_node[0][1], compute_graph.flexml_layers[72].compute_node[0][2], compute_graph.flexml_layers[72].compute_node[0][3], compute_graph.flexml_layers[72].compute_node[1][0], compute_graph.flexml_layers[72].compute_node[1][1], compute_graph.flexml_layers[72].compute_node[1][2], compute_graph.flexml_layers[72].compute_node[1][3], compute_graph.flexml_layers[72].compute_node[2][0], compute_graph.flexml_layers[72].compute_node[2][1], compute_graph.flexml_layers[72].compute_node[2][2], compute_graph.flexml_layers[72].compute_node[2][3], compute_graph.flexml_layers[72].compute_node[3][0], compute_graph.flexml_layers[72].compute_node[3][1], compute_graph.flexml_layers[72].compute_node[3][2], compute_graph.flexml_layers[72].compute_node[3][3], {compute_graph.l2_82.out[4], compute_graph.l2_82.out[5], compute_graph.l2_82.out[6], compute_graph.l2_82.out[7], compute_graph.l2_85.out[0], compute_graph.l2_85.out[1], compute_graph.l2_85.out[2], compute_graph.l2_85.out[3], compute_graph.l2_86.in[0], compute_graph.l2_86.in[1], compute_graph.l2_86.in[2], compute_graph.l2_86.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(850): mode(3), layer(97): {compute_graph.Layer_87_wts_ddr.out[0], compute_graph.Layer_87_wts_ddr.out[1], compute_graph.Layer_87_l2_wts.in[0], compute_graph.Layer_87_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(850) is pipelined with KernelLayerNode(1400) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1401), layer(97): compute_graph.flexml_layers[73].compute_node[0][0], compute_graph.flexml_layers[73].compute_node[0][1], compute_graph.flexml_layers[73].compute_node[0][2], compute_graph.flexml_layers[73].compute_node[0][3], compute_graph.flexml_layers[73].compute_node[1][0], compute_graph.flexml_layers[73].compute_node[1][1], compute_graph.flexml_layers[73].compute_node[1][2], compute_graph.flexml_layers[73].compute_node[1][3], compute_graph.flexml_layers[73].compute_node[2][0], compute_graph.flexml_layers[73].compute_node[2][1], compute_graph.flexml_layers[73].compute_node[2][2], compute_graph.flexml_layers[73].compute_node[2][3], compute_graph.flexml_layers[73].compute_node[3][0], compute_graph.flexml_layers[73].compute_node[3][1], compute_graph.flexml_layers[73].compute_node[3][2], compute_graph.flexml_layers[73].compute_node[3][3], {compute_graph.l2_86.out[0], compute_graph.l2_86.out[1], compute_graph.l2_86.out[2], compute_graph.l2_86.out[3], compute_graph.Layer_87_l2_wts.out[0], compute_graph.Layer_87_l2_wts.out[1], compute_graph.Layer_87_l2_wts.out[2], compute_graph.Layer_87_l2_wts.out[3], compute_graph.l2_87.in[0], compute_graph.l2_87.in[1], compute_graph.l2_87.in[2], compute_graph.l2_87.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1402), layer(98): compute_graph.flexml_layers[74].compute_node[0][0], compute_graph.flexml_layers[74].compute_node[0][1], compute_graph.flexml_layers[74].compute_node[0][2], compute_graph.flexml_layers[74].compute_node[0][3], compute_graph.flexml_layers[74].compute_node[1][0], compute_graph.flexml_layers[74].compute_node[1][1], compute_graph.flexml_layers[74].compute_node[1][2], compute_graph.flexml_layers[74].compute_node[1][3], compute_graph.flexml_layers[74].compute_node[2][0], compute_graph.flexml_layers[74].compute_node[2][1], compute_graph.flexml_layers[74].compute_node[2][2], compute_graph.flexml_layers[74].compute_node[2][3], compute_graph.flexml_layers[74].compute_node[3][0], compute_graph.flexml_layers[74].compute_node[3][1], compute_graph.flexml_layers[74].compute_node[3][2], compute_graph.flexml_layers[74].compute_node[3][3], {compute_graph.l2_87.out[0], compute_graph.l2_87.out[1], compute_graph.l2_87.out[2], compute_graph.l2_87.out[3], compute_graph.l2_88.in[0], compute_graph.l2_88.in[1], compute_graph.l2_88.in[2], compute_graph.l2_88.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1403), layer(99): compute_graph.flexml_layers[75].compute_node[0][0], compute_graph.flexml_layers[75].compute_node[0][1], compute_graph.flexml_layers[75].compute_node[0][2], compute_graph.flexml_layers[75].compute_node[0][3], compute_graph.flexml_layers[75].compute_node[1][0], compute_graph.flexml_layers[75].compute_node[1][1], compute_graph.flexml_layers[75].compute_node[1][2], compute_graph.flexml_layers[75].compute_node[1][3], compute_graph.flexml_layers[75].compute_node[2][0], compute_graph.flexml_layers[75].compute_node[2][1], compute_graph.flexml_layers[75].compute_node[2][2], compute_graph.flexml_layers[75].compute_node[2][3], compute_graph.flexml_layers[75].compute_node[3][0], compute_graph.flexml_layers[75].compute_node[3][1], compute_graph.flexml_layers[75].compute_node[3][2], compute_graph.flexml_layers[75].compute_node[3][3], {compute_graph.l2_88.out[0], compute_graph.l2_88.out[1], compute_graph.l2_88.out[2], compute_graph.l2_88.out[3], compute_graph.l2_89.in[0], compute_graph.l2_89.in[1], compute_graph.l2_89.in[2], compute_graph.l2_89.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1404), layer(100): compute_graph.flexml_layers[76].compute_node[0][0], compute_graph.flexml_layers[76].compute_node[0][1], compute_graph.flexml_layers[76].compute_node[0][2], compute_graph.flexml_layers[76].compute_node[0][3], compute_graph.flexml_layers[76].compute_node[1][0], compute_graph.flexml_layers[76].compute_node[1][1], compute_graph.flexml_layers[76].compute_node[1][2], compute_graph.flexml_layers[76].compute_node[1][3], compute_graph.flexml_layers[76].compute_node[2][0], compute_graph.flexml_layers[76].compute_node[2][1], compute_graph.flexml_layers[76].compute_node[2][2], compute_graph.flexml_layers[76].compute_node[2][3], compute_graph.flexml_layers[76].compute_node[3][0], compute_graph.flexml_layers[76].compute_node[3][1], compute_graph.flexml_layers[76].compute_node[3][2], compute_graph.flexml_layers[76].compute_node[3][3], {compute_graph.l2_89.out[0], compute_graph.l2_89.out[1], compute_graph.l2_89.out[2], compute_graph.l2_89.out[3], compute_graph.l2_90.in[0], compute_graph.l2_90.in[1], compute_graph.l2_90.in[2], compute_graph.l2_90.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1405), layer(101): compute_graph.flexml_layers[77].compute_node[0][0], compute_graph.flexml_layers[77].compute_node[0][1], compute_graph.flexml_layers[77].compute_node[0][2], compute_graph.flexml_layers[77].compute_node[0][3], compute_graph.flexml_layers[77].compute_node[1][0], compute_graph.flexml_layers[77].compute_node[1][1], compute_graph.flexml_layers[77].compute_node[1][2], compute_graph.flexml_layers[77].compute_node[1][3], compute_graph.flexml_layers[77].compute_node[2][0], compute_graph.flexml_layers[77].compute_node[2][1], compute_graph.flexml_layers[77].compute_node[2][2], compute_graph.flexml_layers[77].compute_node[2][3], compute_graph.flexml_layers[77].compute_node[3][0], compute_graph.flexml_layers[77].compute_node[3][1], compute_graph.flexml_layers[77].compute_node[3][2], compute_graph.flexml_layers[77].compute_node[3][3], {compute_graph.l2_87.out[4], compute_graph.l2_87.out[5], compute_graph.l2_87.out[6], compute_graph.l2_87.out[7], compute_graph.l2_90.out[0], compute_graph.l2_90.out[1], compute_graph.l2_90.out[2], compute_graph.l2_90.out[3], compute_graph.l2_91.in[0], compute_graph.l2_91.in[1], compute_graph.l2_91.in[2], compute_graph.l2_91.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(851): mode(3), layer(102): {compute_graph.Layer_92_wts_ddr.out[0], compute_graph.Layer_92_wts_ddr.out[1], compute_graph.Layer_92_l2_wts.in[0], compute_graph.Layer_92_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(851) is pipelined with KernelLayerNode(1405) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.l2_81 overlaps with memory space used by compute_graph.l2_92. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For KernelLayerNode(1406), layer(102): compute_graph.flexml_layers[78].compute_node[0][0], compute_graph.flexml_layers[78].compute_node[0][1], compute_graph.flexml_layers[78].compute_node[0][2], compute_graph.flexml_layers[78].compute_node[0][3], compute_graph.flexml_layers[78].compute_node[1][0], compute_graph.flexml_layers[78].compute_node[1][1], compute_graph.flexml_layers[78].compute_node[1][2], compute_graph.flexml_layers[78].compute_node[1][3], compute_graph.flexml_layers[78].compute_node[2][0], compute_graph.flexml_layers[78].compute_node[2][1], compute_graph.flexml_layers[78].compute_node[2][2], compute_graph.flexml_layers[78].compute_node[2][3], compute_graph.flexml_layers[78].compute_node[3][0], compute_graph.flexml_layers[78].compute_node[3][1], compute_graph.flexml_layers[78].compute_node[3][2], compute_graph.flexml_layers[78].compute_node[3][3], {compute_graph.l2_81.out[4], compute_graph.l2_81.out[5], compute_graph.l2_81.out[6], compute_graph.l2_81.out[7], compute_graph.l2_91.out[0], compute_graph.l2_91.out[1], compute_graph.l2_91.out[2], compute_graph.l2_91.out[3], compute_graph.Layer_92_l2_wts.out[0], compute_graph.Layer_92_l2_wts.out[1], compute_graph.Layer_92_l2_wts.out[2], compute_graph.Layer_92_l2_wts.out[3], compute_graph.l2_92.in[0], compute_graph.l2_92.in[1], compute_graph.l2_92.in[2], compute_graph.l2_92.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(852): mode(3), layer(103): {compute_graph.Layer_93_wts_ddr.out[0], compute_graph.Layer_93_wts_ddr.out[1], compute_graph.Layer_93_l2_wts.in[0], compute_graph.Layer_93_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(852) is pipelined with KernelLayerNode(1406) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1407), layer(103): compute_graph.flexml_layers[79].compute_node[0][0], compute_graph.flexml_layers[79].compute_node[0][1], compute_graph.flexml_layers[79].compute_node[0][2], compute_graph.flexml_layers[79].compute_node[0][3], compute_graph.flexml_layers[79].compute_node[1][0], compute_graph.flexml_layers[79].compute_node[1][1], compute_graph.flexml_layers[79].compute_node[1][2], compute_graph.flexml_layers[79].compute_node[1][3], compute_graph.flexml_layers[79].compute_node[2][0], compute_graph.flexml_layers[79].compute_node[2][1], compute_graph.flexml_layers[79].compute_node[2][2], compute_graph.flexml_layers[79].compute_node[2][3], compute_graph.flexml_layers[79].compute_node[3][0], compute_graph.flexml_layers[79].compute_node[3][1], compute_graph.flexml_layers[79].compute_node[3][2], compute_graph.flexml_layers[79].compute_node[3][3], {compute_graph.l2_92.out[0], compute_graph.l2_92.out[1], compute_graph.l2_92.out[2], compute_graph.l2_92.out[3], compute_graph.Layer_93_l2_wts.out[0], compute_graph.Layer_93_l2_wts.out[1], compute_graph.Layer_93_l2_wts.out[2], compute_graph.Layer_93_l2_wts.out[3], compute_graph.l2_93.in[0], compute_graph.l2_93.in[1], compute_graph.l2_93.in[2], compute_graph.l2_93.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1408), layer(104): compute_graph.flexml_layers[80].compute_node[0][0], compute_graph.flexml_layers[80].compute_node[0][1], compute_graph.flexml_layers[80].compute_node[0][2], compute_graph.flexml_layers[80].compute_node[0][3], compute_graph.flexml_layers[80].compute_node[1][0], compute_graph.flexml_layers[80].compute_node[1][1], compute_graph.flexml_layers[80].compute_node[1][2], compute_graph.flexml_layers[80].compute_node[1][3], compute_graph.flexml_layers[80].compute_node[2][0], compute_graph.flexml_layers[80].compute_node[2][1], compute_graph.flexml_layers[80].compute_node[2][2], compute_graph.flexml_layers[80].compute_node[2][3], compute_graph.flexml_layers[80].compute_node[3][0], compute_graph.flexml_layers[80].compute_node[3][1], compute_graph.flexml_layers[80].compute_node[3][2], compute_graph.flexml_layers[80].compute_node[3][3], {compute_graph.l2_93.out[0], compute_graph.l2_93.out[1], compute_graph.l2_93.out[2], compute_graph.l2_93.out[3], compute_graph.l2_94.in[0], compute_graph.l2_94.in[1], compute_graph.l2_94.in[2], compute_graph.l2_94.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1409), layer(105): compute_graph.flexml_layers[81].compute_node[0][0], compute_graph.flexml_layers[81].compute_node[0][1], compute_graph.flexml_layers[81].compute_node[0][2], compute_graph.flexml_layers[81].compute_node[0][3], compute_graph.flexml_layers[81].compute_node[1][0], compute_graph.flexml_layers[81].compute_node[1][1], compute_graph.flexml_layers[81].compute_node[1][2], compute_graph.flexml_layers[81].compute_node[1][3], compute_graph.flexml_layers[81].compute_node[2][0], compute_graph.flexml_layers[81].compute_node[2][1], compute_graph.flexml_layers[81].compute_node[2][2], compute_graph.flexml_layers[81].compute_node[2][3], compute_graph.flexml_layers[81].compute_node[3][0], compute_graph.flexml_layers[81].compute_node[3][1], compute_graph.flexml_layers[81].compute_node[3][2], compute_graph.flexml_layers[81].compute_node[3][3], {compute_graph.l2_94.out[0], compute_graph.l2_94.out[1], compute_graph.l2_94.out[2], compute_graph.l2_94.out[3], compute_graph.l2_95.in[0], compute_graph.l2_95.in[1], compute_graph.l2_95.in[2], compute_graph.l2_95.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1410), layer(106): compute_graph.flexml_layers[82].compute_node[0][0], compute_graph.flexml_layers[82].compute_node[0][1], compute_graph.flexml_layers[82].compute_node[0][2], compute_graph.flexml_layers[82].compute_node[0][3], compute_graph.flexml_layers[82].compute_node[1][0], compute_graph.flexml_layers[82].compute_node[1][1], compute_graph.flexml_layers[82].compute_node[1][2], compute_graph.flexml_layers[82].compute_node[1][3], compute_graph.flexml_layers[82].compute_node[2][0], compute_graph.flexml_layers[82].compute_node[2][1], compute_graph.flexml_layers[82].compute_node[2][2], compute_graph.flexml_layers[82].compute_node[2][3], compute_graph.flexml_layers[82].compute_node[3][0], compute_graph.flexml_layers[82].compute_node[3][1], compute_graph.flexml_layers[82].compute_node[3][2], compute_graph.flexml_layers[82].compute_node[3][3], {compute_graph.l2_95.out[0], compute_graph.l2_95.out[1], compute_graph.l2_95.out[2], compute_graph.l2_95.out[3], compute_graph.l2_96.in[0], compute_graph.l2_96.in[1], compute_graph.l2_96.in[2], compute_graph.l2_96.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1411), layer(107): compute_graph.flexml_layers[83].compute_node[0][0], compute_graph.flexml_layers[83].compute_node[0][1], compute_graph.flexml_layers[83].compute_node[0][2], compute_graph.flexml_layers[83].compute_node[0][3], compute_graph.flexml_layers[83].compute_node[1][0], compute_graph.flexml_layers[83].compute_node[1][1], compute_graph.flexml_layers[83].compute_node[1][2], compute_graph.flexml_layers[83].compute_node[1][3], compute_graph.flexml_layers[83].compute_node[2][0], compute_graph.flexml_layers[83].compute_node[2][1], compute_graph.flexml_layers[83].compute_node[2][2], compute_graph.flexml_layers[83].compute_node[2][3], compute_graph.flexml_layers[83].compute_node[3][0], compute_graph.flexml_layers[83].compute_node[3][1], compute_graph.flexml_layers[83].compute_node[3][2], compute_graph.flexml_layers[83].compute_node[3][3], {compute_graph.l2_93.out[4], compute_graph.l2_93.out[5], compute_graph.l2_93.out[6], compute_graph.l2_93.out[7], compute_graph.l2_96.out[0], compute_graph.l2_96.out[1], compute_graph.l2_96.out[2], compute_graph.l2_96.out[3], compute_graph.l2_97.in[0], compute_graph.l2_97.in[1], compute_graph.l2_97.in[2], compute_graph.l2_97.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(853): mode(3), layer(108): {compute_graph.Layer_98_wts_ddr.out[0], compute_graph.Layer_98_wts_ddr.out[1], compute_graph.Layer_98_l2_wts.in[0], compute_graph.Layer_98_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(853) is pipelined with KernelLayerNode(1411) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1412), layer(108): compute_graph.flexml_layers[84].compute_node[0][0], compute_graph.flexml_layers[84].compute_node[0][1], compute_graph.flexml_layers[84].compute_node[0][2], compute_graph.flexml_layers[84].compute_node[0][3], compute_graph.flexml_layers[84].compute_node[1][0], compute_graph.flexml_layers[84].compute_node[1][1], compute_graph.flexml_layers[84].compute_node[1][2], compute_graph.flexml_layers[84].compute_node[1][3], compute_graph.flexml_layers[84].compute_node[2][0], compute_graph.flexml_layers[84].compute_node[2][1], compute_graph.flexml_layers[84].compute_node[2][2], compute_graph.flexml_layers[84].compute_node[2][3], compute_graph.flexml_layers[84].compute_node[3][0], compute_graph.flexml_layers[84].compute_node[3][1], compute_graph.flexml_layers[84].compute_node[3][2], compute_graph.flexml_layers[84].compute_node[3][3], {compute_graph.l2_97.out[0], compute_graph.l2_97.out[1], compute_graph.l2_97.out[2], compute_graph.l2_97.out[3], compute_graph.Layer_98_l2_wts.out[0], compute_graph.Layer_98_l2_wts.out[1], compute_graph.Layer_98_l2_wts.out[2], compute_graph.Layer_98_l2_wts.out[3], compute_graph.l2_98.in[0], compute_graph.l2_98.in[1], compute_graph.l2_98.in[2], compute_graph.l2_98.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1413), layer(109): compute_graph.flexml_layers[85].compute_node[0][0], compute_graph.flexml_layers[85].compute_node[0][1], compute_graph.flexml_layers[85].compute_node[0][2], compute_graph.flexml_layers[85].compute_node[0][3], compute_graph.flexml_layers[85].compute_node[1][0], compute_graph.flexml_layers[85].compute_node[1][1], compute_graph.flexml_layers[85].compute_node[1][2], compute_graph.flexml_layers[85].compute_node[1][3], compute_graph.flexml_layers[85].compute_node[2][0], compute_graph.flexml_layers[85].compute_node[2][1], compute_graph.flexml_layers[85].compute_node[2][2], compute_graph.flexml_layers[85].compute_node[2][3], compute_graph.flexml_layers[85].compute_node[3][0], compute_graph.flexml_layers[85].compute_node[3][1], compute_graph.flexml_layers[85].compute_node[3][2], compute_graph.flexml_layers[85].compute_node[3][3], {compute_graph.l2_98.out[0], compute_graph.l2_98.out[1], compute_graph.l2_98.out[2], compute_graph.l2_98.out[3], compute_graph.l2_99.in[0], compute_graph.l2_99.in[1], compute_graph.l2_99.in[2], compute_graph.l2_99.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1414), layer(110): compute_graph.flexml_layers[86].compute_node[0][0], compute_graph.flexml_layers[86].compute_node[0][1], compute_graph.flexml_layers[86].compute_node[0][2], compute_graph.flexml_layers[86].compute_node[0][3], compute_graph.flexml_layers[86].compute_node[1][0], compute_graph.flexml_layers[86].compute_node[1][1], compute_graph.flexml_layers[86].compute_node[1][2], compute_graph.flexml_layers[86].compute_node[1][3], compute_graph.flexml_layers[86].compute_node[2][0], compute_graph.flexml_layers[86].compute_node[2][1], compute_graph.flexml_layers[86].compute_node[2][2], compute_graph.flexml_layers[86].compute_node[2][3], compute_graph.flexml_layers[86].compute_node[3][0], compute_graph.flexml_layers[86].compute_node[3][1], compute_graph.flexml_layers[86].compute_node[3][2], compute_graph.flexml_layers[86].compute_node[3][3], {compute_graph.l2_99.out[0], compute_graph.l2_99.out[1], compute_graph.l2_99.out[2], compute_graph.l2_99.out[3], compute_graph.l2_100.in[0], compute_graph.l2_100.in[1], compute_graph.l2_100.in[2], compute_graph.l2_100.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1415), layer(111): compute_graph.flexml_layers[87].compute_node[0][0], compute_graph.flexml_layers[87].compute_node[0][1], compute_graph.flexml_layers[87].compute_node[0][2], compute_graph.flexml_layers[87].compute_node[0][3], compute_graph.flexml_layers[87].compute_node[1][0], compute_graph.flexml_layers[87].compute_node[1][1], compute_graph.flexml_layers[87].compute_node[1][2], compute_graph.flexml_layers[87].compute_node[1][3], compute_graph.flexml_layers[87].compute_node[2][0], compute_graph.flexml_layers[87].compute_node[2][1], compute_graph.flexml_layers[87].compute_node[2][2], compute_graph.flexml_layers[87].compute_node[2][3], compute_graph.flexml_layers[87].compute_node[3][0], compute_graph.flexml_layers[87].compute_node[3][1], compute_graph.flexml_layers[87].compute_node[3][2], compute_graph.flexml_layers[87].compute_node[3][3], {compute_graph.l2_100.out[0], compute_graph.l2_100.out[1], compute_graph.l2_100.out[2], compute_graph.l2_100.out[3], compute_graph.l2_101.in[0], compute_graph.l2_101.in[1], compute_graph.l2_101.in[2], compute_graph.l2_101.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1416), layer(112): compute_graph.flexml_layers[88].compute_node[0][0], compute_graph.flexml_layers[88].compute_node[0][1], compute_graph.flexml_layers[88].compute_node[0][2], compute_graph.flexml_layers[88].compute_node[0][3], compute_graph.flexml_layers[88].compute_node[1][0], compute_graph.flexml_layers[88].compute_node[1][1], compute_graph.flexml_layers[88].compute_node[1][2], compute_graph.flexml_layers[88].compute_node[1][3], compute_graph.flexml_layers[88].compute_node[2][0], compute_graph.flexml_layers[88].compute_node[2][1], compute_graph.flexml_layers[88].compute_node[2][2], compute_graph.flexml_layers[88].compute_node[2][3], compute_graph.flexml_layers[88].compute_node[3][0], compute_graph.flexml_layers[88].compute_node[3][1], compute_graph.flexml_layers[88].compute_node[3][2], compute_graph.flexml_layers[88].compute_node[3][3], {compute_graph.l2_98.out[4], compute_graph.l2_98.out[5], compute_graph.l2_98.out[6], compute_graph.l2_98.out[7], compute_graph.l2_101.out[0], compute_graph.l2_101.out[1], compute_graph.l2_101.out[2], compute_graph.l2_101.out[3], compute_graph.l2_102.in[0], compute_graph.l2_102.in[1], compute_graph.l2_102.in[2], compute_graph.l2_102.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(854): mode(3), layer(113): {compute_graph.Layer_103_wts_ddr.out[0], compute_graph.Layer_103_wts_ddr.out[1], compute_graph.Layer_103_l2_wts.in[0], compute_graph.Layer_103_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(854) is pipelined with KernelLayerNode(1416) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.l2_92 overlaps with memory space used by compute_graph.l2_103. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For KernelLayerNode(1417), layer(113): compute_graph.flexml_layers[89].compute_node[0][0], compute_graph.flexml_layers[89].compute_node[0][1], compute_graph.flexml_layers[89].compute_node[0][2], compute_graph.flexml_layers[89].compute_node[0][3], compute_graph.flexml_layers[89].compute_node[1][0], compute_graph.flexml_layers[89].compute_node[1][1], compute_graph.flexml_layers[89].compute_node[1][2], compute_graph.flexml_layers[89].compute_node[1][3], compute_graph.flexml_layers[89].compute_node[2][0], compute_graph.flexml_layers[89].compute_node[2][1], compute_graph.flexml_layers[89].compute_node[2][2], compute_graph.flexml_layers[89].compute_node[2][3], compute_graph.flexml_layers[89].compute_node[3][0], compute_graph.flexml_layers[89].compute_node[3][1], compute_graph.flexml_layers[89].compute_node[3][2], compute_graph.flexml_layers[89].compute_node[3][3], {compute_graph.l2_92.out[4], compute_graph.l2_92.out[5], compute_graph.l2_92.out[6], compute_graph.l2_92.out[7], compute_graph.l2_102.out[0], compute_graph.l2_102.out[1], compute_graph.l2_102.out[2], compute_graph.l2_102.out[3], compute_graph.Layer_103_l2_wts.out[0], compute_graph.Layer_103_l2_wts.out[1], compute_graph.Layer_103_l2_wts.out[2], compute_graph.Layer_103_l2_wts.out[3], compute_graph.l2_103.in[0], compute_graph.l2_103.in[1], compute_graph.l2_103.in[2], compute_graph.l2_103.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(855): mode(3), layer(114): {compute_graph.Layer_104_wts_ddr.out[0], compute_graph.Layer_104_wts_ddr.out[1], compute_graph.Layer_104_l2_wts.in[0], compute_graph.Layer_104_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(855) is pipelined with KernelLayerNode(1417) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1418), layer(114): compute_graph.flexml_layers[90].compute_node[0][0], compute_graph.flexml_layers[90].compute_node[0][1], compute_graph.flexml_layers[90].compute_node[0][2], compute_graph.flexml_layers[90].compute_node[0][3], compute_graph.flexml_layers[90].compute_node[1][0], compute_graph.flexml_layers[90].compute_node[1][1], compute_graph.flexml_layers[90].compute_node[1][2], compute_graph.flexml_layers[90].compute_node[1][3], compute_graph.flexml_layers[90].compute_node[2][0], compute_graph.flexml_layers[90].compute_node[2][1], compute_graph.flexml_layers[90].compute_node[2][2], compute_graph.flexml_layers[90].compute_node[2][3], compute_graph.flexml_layers[90].compute_node[3][0], compute_graph.flexml_layers[90].compute_node[3][1], compute_graph.flexml_layers[90].compute_node[3][2], compute_graph.flexml_layers[90].compute_node[3][3], {compute_graph.l2_103.out[0], compute_graph.l2_103.out[1], compute_graph.l2_103.out[2], compute_graph.l2_103.out[3], compute_graph.Layer_104_l2_wts.out[0], compute_graph.Layer_104_l2_wts.out[1], compute_graph.Layer_104_l2_wts.out[2], compute_graph.Layer_104_l2_wts.out[3], compute_graph.l2_104.in[0], compute_graph.l2_104.in[1], compute_graph.l2_104.in[2], compute_graph.l2_104.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1419), layer(115): compute_graph.flexml_layers[91].compute_node[0][0], compute_graph.flexml_layers[91].compute_node[0][1], compute_graph.flexml_layers[91].compute_node[0][2], compute_graph.flexml_layers[91].compute_node[0][3], compute_graph.flexml_layers[91].compute_node[1][0], compute_graph.flexml_layers[91].compute_node[1][1], compute_graph.flexml_layers[91].compute_node[1][2], compute_graph.flexml_layers[91].compute_node[1][3], compute_graph.flexml_layers[91].compute_node[2][0], compute_graph.flexml_layers[91].compute_node[2][1], compute_graph.flexml_layers[91].compute_node[2][2], compute_graph.flexml_layers[91].compute_node[2][3], compute_graph.flexml_layers[91].compute_node[3][0], compute_graph.flexml_layers[91].compute_node[3][1], compute_graph.flexml_layers[91].compute_node[3][2], compute_graph.flexml_layers[91].compute_node[3][3], {compute_graph.l2_104.out[0], compute_graph.l2_104.out[1], compute_graph.l2_104.out[2], compute_graph.l2_104.out[3], compute_graph.l2_105.in[0], compute_graph.l2_105.in[1], compute_graph.l2_105.in[2], compute_graph.l2_105.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1420), layer(116): compute_graph.flexml_layers[92].compute_node[0][0], compute_graph.flexml_layers[92].compute_node[0][1], compute_graph.flexml_layers[92].compute_node[0][2], compute_graph.flexml_layers[92].compute_node[0][3], compute_graph.flexml_layers[92].compute_node[1][0], compute_graph.flexml_layers[92].compute_node[1][1], compute_graph.flexml_layers[92].compute_node[1][2], compute_graph.flexml_layers[92].compute_node[1][3], compute_graph.flexml_layers[92].compute_node[2][0], compute_graph.flexml_layers[92].compute_node[2][1], compute_graph.flexml_layers[92].compute_node[2][2], compute_graph.flexml_layers[92].compute_node[2][3], compute_graph.flexml_layers[92].compute_node[3][0], compute_graph.flexml_layers[92].compute_node[3][1], compute_graph.flexml_layers[92].compute_node[3][2], compute_graph.flexml_layers[92].compute_node[3][3], {compute_graph.l2_105.out[0], compute_graph.l2_105.out[1], compute_graph.l2_105.out[2], compute_graph.l2_105.out[3], compute_graph.l2_106.in[0], compute_graph.l2_106.in[1], compute_graph.l2_106.in[2], compute_graph.l2_106.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1421), layer(117): compute_graph.flexml_layers[93].compute_node[0][0], compute_graph.flexml_layers[93].compute_node[0][1], compute_graph.flexml_layers[93].compute_node[0][2], compute_graph.flexml_layers[93].compute_node[0][3], compute_graph.flexml_layers[93].compute_node[1][0], compute_graph.flexml_layers[93].compute_node[1][1], compute_graph.flexml_layers[93].compute_node[1][2], compute_graph.flexml_layers[93].compute_node[1][3], compute_graph.flexml_layers[93].compute_node[2][0], compute_graph.flexml_layers[93].compute_node[2][1], compute_graph.flexml_layers[93].compute_node[2][2], compute_graph.flexml_layers[93].compute_node[2][3], compute_graph.flexml_layers[93].compute_node[3][0], compute_graph.flexml_layers[93].compute_node[3][1], compute_graph.flexml_layers[93].compute_node[3][2], compute_graph.flexml_layers[93].compute_node[3][3], {compute_graph.l2_106.out[0], compute_graph.l2_106.out[1], compute_graph.l2_106.out[2], compute_graph.l2_106.out[3], compute_graph.l2_107.in[0], compute_graph.l2_107.in[1], compute_graph.l2_107.in[2], compute_graph.l2_107.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1422), layer(118): compute_graph.flexml_layers[94].compute_node[0][0], compute_graph.flexml_layers[94].compute_node[0][1], compute_graph.flexml_layers[94].compute_node[0][2], compute_graph.flexml_layers[94].compute_node[0][3], compute_graph.flexml_layers[94].compute_node[1][0], compute_graph.flexml_layers[94].compute_node[1][1], compute_graph.flexml_layers[94].compute_node[1][2], compute_graph.flexml_layers[94].compute_node[1][3], compute_graph.flexml_layers[94].compute_node[2][0], compute_graph.flexml_layers[94].compute_node[2][1], compute_graph.flexml_layers[94].compute_node[2][2], compute_graph.flexml_layers[94].compute_node[2][3], compute_graph.flexml_layers[94].compute_node[3][0], compute_graph.flexml_layers[94].compute_node[3][1], compute_graph.flexml_layers[94].compute_node[3][2], compute_graph.flexml_layers[94].compute_node[3][3], {compute_graph.l2_104.out[4], compute_graph.l2_104.out[5], compute_graph.l2_104.out[6], compute_graph.l2_104.out[7], compute_graph.l2_107.out[0], compute_graph.l2_107.out[1], compute_graph.l2_107.out[2], compute_graph.l2_107.out[3], compute_graph.l2_108.in[0], compute_graph.l2_108.in[1], compute_graph.l2_108.in[2], compute_graph.l2_108.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(856): mode(3), layer(119): {compute_graph.Layer_109_wts_ddr.out[0], compute_graph.Layer_109_wts_ddr.out[1], compute_graph.Layer_109_l2_wts.in[0], compute_graph.Layer_109_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(856) is pipelined with KernelLayerNode(1422) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1423), layer(119): compute_graph.flexml_layers[95].compute_node[0][0], compute_graph.flexml_layers[95].compute_node[0][1], compute_graph.flexml_layers[95].compute_node[0][2], compute_graph.flexml_layers[95].compute_node[0][3], compute_graph.flexml_layers[95].compute_node[1][0], compute_graph.flexml_layers[95].compute_node[1][1], compute_graph.flexml_layers[95].compute_node[1][2], compute_graph.flexml_layers[95].compute_node[1][3], compute_graph.flexml_layers[95].compute_node[2][0], compute_graph.flexml_layers[95].compute_node[2][1], compute_graph.flexml_layers[95].compute_node[2][2], compute_graph.flexml_layers[95].compute_node[2][3], compute_graph.flexml_layers[95].compute_node[3][0], compute_graph.flexml_layers[95].compute_node[3][1], compute_graph.flexml_layers[95].compute_node[3][2], compute_graph.flexml_layers[95].compute_node[3][3], {compute_graph.l2_108.out[0], compute_graph.l2_108.out[1], compute_graph.l2_108.out[2], compute_graph.l2_108.out[3], compute_graph.Layer_109_l2_wts.out[0], compute_graph.Layer_109_l2_wts.out[1], compute_graph.Layer_109_l2_wts.out[2], compute_graph.Layer_109_l2_wts.out[3], compute_graph.l2_109.in[0], compute_graph.l2_109.in[1], compute_graph.l2_109.in[2], compute_graph.l2_109.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1424), layer(120): compute_graph.flexml_layers[96].compute_node[0][0], compute_graph.flexml_layers[96].compute_node[0][1], compute_graph.flexml_layers[96].compute_node[0][2], compute_graph.flexml_layers[96].compute_node[0][3], compute_graph.flexml_layers[96].compute_node[1][0], compute_graph.flexml_layers[96].compute_node[1][1], compute_graph.flexml_layers[96].compute_node[1][2], compute_graph.flexml_layers[96].compute_node[1][3], compute_graph.flexml_layers[96].compute_node[2][0], compute_graph.flexml_layers[96].compute_node[2][1], compute_graph.flexml_layers[96].compute_node[2][2], compute_graph.flexml_layers[96].compute_node[2][3], compute_graph.flexml_layers[96].compute_node[3][0], compute_graph.flexml_layers[96].compute_node[3][1], compute_graph.flexml_layers[96].compute_node[3][2], compute_graph.flexml_layers[96].compute_node[3][3], {compute_graph.l2_109.out[0], compute_graph.l2_109.out[1], compute_graph.l2_109.out[2], compute_graph.l2_109.out[3], compute_graph.l2_110.in[0], compute_graph.l2_110.in[1], compute_graph.l2_110.in[2], compute_graph.l2_110.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1425), layer(121): compute_graph.flexml_layers[97].compute_node[0][0], compute_graph.flexml_layers[97].compute_node[0][1], compute_graph.flexml_layers[97].compute_node[0][2], compute_graph.flexml_layers[97].compute_node[0][3], compute_graph.flexml_layers[97].compute_node[1][0], compute_graph.flexml_layers[97].compute_node[1][1], compute_graph.flexml_layers[97].compute_node[1][2], compute_graph.flexml_layers[97].compute_node[1][3], compute_graph.flexml_layers[97].compute_node[2][0], compute_graph.flexml_layers[97].compute_node[2][1], compute_graph.flexml_layers[97].compute_node[2][2], compute_graph.flexml_layers[97].compute_node[2][3], compute_graph.flexml_layers[97].compute_node[3][0], compute_graph.flexml_layers[97].compute_node[3][1], compute_graph.flexml_layers[97].compute_node[3][2], compute_graph.flexml_layers[97].compute_node[3][3], {compute_graph.l2_110.out[0], compute_graph.l2_110.out[1], compute_graph.l2_110.out[2], compute_graph.l2_110.out[3], compute_graph.l2_111.in[0], compute_graph.l2_111.in[1], compute_graph.l2_111.in[2], compute_graph.l2_111.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1426), layer(122): compute_graph.flexml_layers[98].compute_node[0][0], compute_graph.flexml_layers[98].compute_node[0][1], compute_graph.flexml_layers[98].compute_node[0][2], compute_graph.flexml_layers[98].compute_node[0][3], compute_graph.flexml_layers[98].compute_node[1][0], compute_graph.flexml_layers[98].compute_node[1][1], compute_graph.flexml_layers[98].compute_node[1][2], compute_graph.flexml_layers[98].compute_node[1][3], compute_graph.flexml_layers[98].compute_node[2][0], compute_graph.flexml_layers[98].compute_node[2][1], compute_graph.flexml_layers[98].compute_node[2][2], compute_graph.flexml_layers[98].compute_node[2][3], compute_graph.flexml_layers[98].compute_node[3][0], compute_graph.flexml_layers[98].compute_node[3][1], compute_graph.flexml_layers[98].compute_node[3][2], compute_graph.flexml_layers[98].compute_node[3][3], {compute_graph.l2_111.out[0], compute_graph.l2_111.out[1], compute_graph.l2_111.out[2], compute_graph.l2_111.out[3], compute_graph.l2_112.in[0], compute_graph.l2_112.in[1], compute_graph.l2_112.in[2], compute_graph.l2_112.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1427), layer(123): compute_graph.flexml_layers[99].compute_node[0][0], compute_graph.flexml_layers[99].compute_node[0][1], compute_graph.flexml_layers[99].compute_node[0][2], compute_graph.flexml_layers[99].compute_node[0][3], compute_graph.flexml_layers[99].compute_node[1][0], compute_graph.flexml_layers[99].compute_node[1][1], compute_graph.flexml_layers[99].compute_node[1][2], compute_graph.flexml_layers[99].compute_node[1][3], compute_graph.flexml_layers[99].compute_node[2][0], compute_graph.flexml_layers[99].compute_node[2][1], compute_graph.flexml_layers[99].compute_node[2][2], compute_graph.flexml_layers[99].compute_node[2][3], compute_graph.flexml_layers[99].compute_node[3][0], compute_graph.flexml_layers[99].compute_node[3][1], compute_graph.flexml_layers[99].compute_node[3][2], compute_graph.flexml_layers[99].compute_node[3][3], {compute_graph.l2_109.out[4], compute_graph.l2_109.out[5], compute_graph.l2_109.out[6], compute_graph.l2_109.out[7], compute_graph.l2_112.out[0], compute_graph.l2_112.out[1], compute_graph.l2_112.out[2], compute_graph.l2_112.out[3], compute_graph.l2_113.in[0], compute_graph.l2_113.in[1], compute_graph.l2_113.in[2], compute_graph.l2_113.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(930): mode(0), layer(123): {compute_graph.l2_113.out[2], compute_graph.l2_113.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(929): mode(0), layer(123): {compute_graph.l2_113.out[0], compute_graph.l2_113.out[1], compute_graph.l2l3_113_spill.in[0], compute_graph.l2l3_113_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1074): mode(0), layer(124): {compute_graph.l2l3_113_spill.out[0], compute_graph.templated_graph_114.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1258): mode(0), layer(124): {compute_graph.templated_graph_114.trans_mt_ifm.out[0], compute_graph.templated_graph_114.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1259): mode(0), layer(124): {compute_graph.templated_graph_114.trans_mt_ofm.out[0], compute_graph.l2l3_114_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1075): mode(0), layer(125): {compute_graph.l2l3_114_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1428), layer(125): compute_graph.flexml_layers[100].compute_node[0][0], compute_graph.flexml_layers[100].compute_node[0][1], compute_graph.flexml_layers[100].compute_node[0][2], compute_graph.flexml_layers[100].compute_node[0][3], compute_graph.flexml_layers[100].compute_node[1][0], compute_graph.flexml_layers[100].compute_node[1][1], compute_graph.flexml_layers[100].compute_node[1][2], compute_graph.flexml_layers[100].compute_node[1][3], compute_graph.flexml_layers[100].compute_node[2][0], compute_graph.flexml_layers[100].compute_node[2][1], compute_graph.flexml_layers[100].compute_node[2][2], compute_graph.flexml_layers[100].compute_node[2][3], compute_graph.flexml_layers[100].compute_node[3][0], compute_graph.flexml_layers[100].compute_node[3][1], compute_graph.flexml_layers[100].compute_node[3][2], compute_graph.flexml_layers[100].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[3], compute_graph.l2_115.in[0], compute_graph.l2_115.in[1], compute_graph.l2_115.in[2], compute_graph.l2_115.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(857): mode(3), layer(126): {compute_graph.Layer_116_wts_ddr.out[0], compute_graph.Layer_116_wts_ddr.out[1], compute_graph.Layer_116_l2_wts.in[0], compute_graph.Layer_116_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(857) is pipelined with KernelLayerNode(1428) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1429), layer(126): compute_graph.flexml_layers[101].compute_node[0][0], compute_graph.flexml_layers[101].compute_node[0][1], compute_graph.flexml_layers[101].compute_node[0][2], compute_graph.flexml_layers[101].compute_node[0][3], compute_graph.flexml_layers[101].compute_node[1][0], compute_graph.flexml_layers[101].compute_node[1][1], compute_graph.flexml_layers[101].compute_node[1][2], compute_graph.flexml_layers[101].compute_node[1][3], compute_graph.flexml_layers[101].compute_node[2][0], compute_graph.flexml_layers[101].compute_node[2][1], compute_graph.flexml_layers[101].compute_node[2][2], compute_graph.flexml_layers[101].compute_node[2][3], compute_graph.flexml_layers[101].compute_node[3][0], compute_graph.flexml_layers[101].compute_node[3][1], compute_graph.flexml_layers[101].compute_node[3][2], compute_graph.flexml_layers[101].compute_node[3][3], {compute_graph.l2_115.out[0], compute_graph.l2_115.out[1], compute_graph.l2_115.out[2], compute_graph.l2_115.out[3], compute_graph.Layer_116_l2_wts.out[0], compute_graph.Layer_116_l2_wts.out[1], compute_graph.Layer_116_l2_wts.out[2], compute_graph.Layer_116_l2_wts.out[3], compute_graph.l2_116.in[0], compute_graph.l2_116.in[1], compute_graph.l2_116.in[2], compute_graph.l2_116.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(858): mode(3), layer(127): {compute_graph.Layer_117_wts_ddr.out[0], compute_graph.Layer_117_wts_ddr.out[1], compute_graph.Layer_117_l2_wts.in[0], compute_graph.Layer_117_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(858) is pipelined with KernelLayerNode(1429) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1430), layer(127): compute_graph.flexml_layers[102].compute_node[0][0], compute_graph.flexml_layers[102].compute_node[0][1], compute_graph.flexml_layers[102].compute_node[0][2], compute_graph.flexml_layers[102].compute_node[0][3], compute_graph.flexml_layers[102].compute_node[1][0], compute_graph.flexml_layers[102].compute_node[1][1], compute_graph.flexml_layers[102].compute_node[1][2], compute_graph.flexml_layers[102].compute_node[1][3], compute_graph.flexml_layers[102].compute_node[2][0], compute_graph.flexml_layers[102].compute_node[2][1], compute_graph.flexml_layers[102].compute_node[2][2], compute_graph.flexml_layers[102].compute_node[2][3], compute_graph.flexml_layers[102].compute_node[3][0], compute_graph.flexml_layers[102].compute_node[3][1], compute_graph.flexml_layers[102].compute_node[3][2], compute_graph.flexml_layers[102].compute_node[3][3], {compute_graph.l2_116.out[0], compute_graph.l2_116.out[1], compute_graph.l2_116.out[2], compute_graph.l2_116.out[3], compute_graph.Layer_117_l2_wts.out[0], compute_graph.Layer_117_l2_wts.out[1], compute_graph.Layer_117_l2_wts.out[2], compute_graph.Layer_117_l2_wts.out[3], compute_graph.l2_117.in[0], compute_graph.l2_117.in[1], compute_graph.l2_117.in[2], compute_graph.l2_117.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1431), layer(128): compute_graph.flexml_layers[103].compute_node[0][0], compute_graph.flexml_layers[103].compute_node[0][1], compute_graph.flexml_layers[103].compute_node[0][2], compute_graph.flexml_layers[103].compute_node[0][3], compute_graph.flexml_layers[103].compute_node[1][0], compute_graph.flexml_layers[103].compute_node[1][1], compute_graph.flexml_layers[103].compute_node[1][2], compute_graph.flexml_layers[103].compute_node[1][3], compute_graph.flexml_layers[103].compute_node[2][0], compute_graph.flexml_layers[103].compute_node[2][1], compute_graph.flexml_layers[103].compute_node[2][2], compute_graph.flexml_layers[103].compute_node[2][3], compute_graph.flexml_layers[103].compute_node[3][0], compute_graph.flexml_layers[103].compute_node[3][1], compute_graph.flexml_layers[103].compute_node[3][2], compute_graph.flexml_layers[103].compute_node[3][3], {compute_graph.l2_117.out[0], compute_graph.l2_117.out[1], compute_graph.l2_117.out[2], compute_graph.l2_117.out[3], compute_graph.l2_118.in[0], compute_graph.l2_118.in[1], compute_graph.l2_118.in[2], compute_graph.l2_118.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1432), layer(129): compute_graph.flexml_layers[104].compute_node[0][0], compute_graph.flexml_layers[104].compute_node[0][1], compute_graph.flexml_layers[104].compute_node[0][2], compute_graph.flexml_layers[104].compute_node[0][3], compute_graph.flexml_layers[104].compute_node[1][0], compute_graph.flexml_layers[104].compute_node[1][1], compute_graph.flexml_layers[104].compute_node[1][2], compute_graph.flexml_layers[104].compute_node[1][3], compute_graph.flexml_layers[104].compute_node[2][0], compute_graph.flexml_layers[104].compute_node[2][1], compute_graph.flexml_layers[104].compute_node[2][2], compute_graph.flexml_layers[104].compute_node[2][3], compute_graph.flexml_layers[104].compute_node[3][0], compute_graph.flexml_layers[104].compute_node[3][1], compute_graph.flexml_layers[104].compute_node[3][2], compute_graph.flexml_layers[104].compute_node[3][3], {compute_graph.l2_118.out[0], compute_graph.l2_118.out[1], compute_graph.l2_118.out[2], compute_graph.l2_118.out[3], compute_graph.l2_119.in[0], compute_graph.l2_119.in[1], compute_graph.l2_119.in[2], compute_graph.l2_119.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1433), layer(130): compute_graph.flexml_layers[105].compute_node[0][0], compute_graph.flexml_layers[105].compute_node[0][1], compute_graph.flexml_layers[105].compute_node[0][2], compute_graph.flexml_layers[105].compute_node[0][3], compute_graph.flexml_layers[105].compute_node[1][0], compute_graph.flexml_layers[105].compute_node[1][1], compute_graph.flexml_layers[105].compute_node[1][2], compute_graph.flexml_layers[105].compute_node[1][3], compute_graph.flexml_layers[105].compute_node[2][0], compute_graph.flexml_layers[105].compute_node[2][1], compute_graph.flexml_layers[105].compute_node[2][2], compute_graph.flexml_layers[105].compute_node[2][3], compute_graph.flexml_layers[105].compute_node[3][0], compute_graph.flexml_layers[105].compute_node[3][1], compute_graph.flexml_layers[105].compute_node[3][2], compute_graph.flexml_layers[105].compute_node[3][3], {compute_graph.l2_119.out[0], compute_graph.l2_119.out[1], compute_graph.l2_119.out[2], compute_graph.l2_119.out[3], compute_graph.l2_120.in[0], compute_graph.l2_120.in[1], compute_graph.l2_120.in[2], compute_graph.l2_120.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(931): mode(0), layer(130): {compute_graph.l2_120.out[1], compute_graph.l2l3_120_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1076): mode(0), layer(131): {compute_graph.l2l3_120_spill.out[0], compute_graph.templated_graph_121.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1260): mode(0), layer(131): {compute_graph.templated_graph_121.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_121_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1077): mode(0), layer(132): {compute_graph.l2l3_scratch_0_121_spill.out[0], compute_graph.templated_graph_121.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1261): mode(0), layer(132): {compute_graph.templated_graph_121.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_121_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1078): mode(0), layer(133): {compute_graph.l2l3_scratch_1_121_spill.out[0], compute_graph.templated_graph_121.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1262): mode(0), layer(133): {compute_graph.templated_graph_121.g2.ifm_mem.out[0], compute_graph.l2l3_121_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1073): mode(0), layer(134): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1079): mode(0), layer(134): {compute_graph.l2l3_121_spill.out[0], compute_graph.l2l3_121_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1434), layer(134): compute_graph.flexml_layers[106].compute_node[0][0], compute_graph.flexml_layers[106].compute_node[0][1], compute_graph.flexml_layers[106].compute_node[0][2], compute_graph.flexml_layers[106].compute_node[0][3], compute_graph.flexml_layers[106].compute_node[1][0], compute_graph.flexml_layers[106].compute_node[1][1], compute_graph.flexml_layers[106].compute_node[1][2], compute_graph.flexml_layers[106].compute_node[1][3], compute_graph.flexml_layers[106].compute_node[2][0], compute_graph.flexml_layers[106].compute_node[2][1], compute_graph.flexml_layers[106].compute_node[2][2], compute_graph.flexml_layers[106].compute_node[2][3], compute_graph.flexml_layers[106].compute_node[3][0], compute_graph.flexml_layers[106].compute_node[3][1], compute_graph.flexml_layers[106].compute_node[3][2], compute_graph.flexml_layers[106].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[3], compute_graph.l2_122.in[0], compute_graph.l2_122.in[1], compute_graph.l2_122.in[2], compute_graph.l2_122.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(859): mode(3), layer(135): {compute_graph.Layer_123_wts_ddr.out[0], compute_graph.Layer_123_wts_ddr.out[1], compute_graph.Layer_123_l2_wts.in[0], compute_graph.Layer_123_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(859) is pipelined with KernelLayerNode(1434) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1435), layer(135): compute_graph.flexml_layers[107].compute_node[0][0], compute_graph.flexml_layers[107].compute_node[0][1], compute_graph.flexml_layers[107].compute_node[0][2], compute_graph.flexml_layers[107].compute_node[0][3], compute_graph.flexml_layers[107].compute_node[1][0], compute_graph.flexml_layers[107].compute_node[1][1], compute_graph.flexml_layers[107].compute_node[1][2], compute_graph.flexml_layers[107].compute_node[1][3], compute_graph.flexml_layers[107].compute_node[2][0], compute_graph.flexml_layers[107].compute_node[2][1], compute_graph.flexml_layers[107].compute_node[2][2], compute_graph.flexml_layers[107].compute_node[2][3], compute_graph.flexml_layers[107].compute_node[3][0], compute_graph.flexml_layers[107].compute_node[3][1], compute_graph.flexml_layers[107].compute_node[3][2], compute_graph.flexml_layers[107].compute_node[3][3], {compute_graph.l2_122.out[0], compute_graph.l2_122.out[1], compute_graph.l2_122.out[2], compute_graph.l2_122.out[3], compute_graph.Layer_123_l2_wts.out[0], compute_graph.Layer_123_l2_wts.out[1], compute_graph.Layer_123_l2_wts.out[2], compute_graph.Layer_123_l2_wts.out[3], compute_graph.l2_123.in[0], compute_graph.l2_123.in[1], compute_graph.l2_123.in[2], compute_graph.l2_123.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(860): mode(3), layer(136): {compute_graph.Layer_124_wts_ddr.out[0], compute_graph.Layer_124_wts_ddr.out[1], compute_graph.Layer_124_l2_wts.in[0], compute_graph.Layer_124_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(860) is pipelined with KernelLayerNode(1435) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1436), layer(136): compute_graph.flexml_layers[108].compute_node[0][0], compute_graph.flexml_layers[108].compute_node[0][1], compute_graph.flexml_layers[108].compute_node[0][2], compute_graph.flexml_layers[108].compute_node[0][3], compute_graph.flexml_layers[108].compute_node[1][0], compute_graph.flexml_layers[108].compute_node[1][1], compute_graph.flexml_layers[108].compute_node[1][2], compute_graph.flexml_layers[108].compute_node[1][3], compute_graph.flexml_layers[108].compute_node[2][0], compute_graph.flexml_layers[108].compute_node[2][1], compute_graph.flexml_layers[108].compute_node[2][2], compute_graph.flexml_layers[108].compute_node[2][3], compute_graph.flexml_layers[108].compute_node[3][0], compute_graph.flexml_layers[108].compute_node[3][1], compute_graph.flexml_layers[108].compute_node[3][2], compute_graph.flexml_layers[108].compute_node[3][3], {compute_graph.l2_123.out[0], compute_graph.l2_123.out[1], compute_graph.l2_123.out[2], compute_graph.l2_123.out[3], compute_graph.Layer_124_l2_wts.out[0], compute_graph.Layer_124_l2_wts.out[1], compute_graph.Layer_124_l2_wts.out[2], compute_graph.Layer_124_l2_wts.out[3], compute_graph.l2_124.in[0], compute_graph.l2_124.in[1], compute_graph.l2_124.in[2], compute_graph.l2_124.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(932): mode(3), layer(136): {compute_graph.l2_123.out[4], compute_graph.l2_123.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(932) is pipelined with KernelLayerNode(1436) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1437), layer(137): compute_graph.flexml_layers[109].compute_node[0][0], compute_graph.flexml_layers[109].compute_node[0][1], compute_graph.flexml_layers[109].compute_node[0][2], compute_graph.flexml_layers[109].compute_node[0][3], compute_graph.flexml_layers[109].compute_node[1][0], compute_graph.flexml_layers[109].compute_node[1][1], compute_graph.flexml_layers[109].compute_node[1][2], compute_graph.flexml_layers[109].compute_node[1][3], compute_graph.flexml_layers[109].compute_node[2][0], compute_graph.flexml_layers[109].compute_node[2][1], compute_graph.flexml_layers[109].compute_node[2][2], compute_graph.flexml_layers[109].compute_node[2][3], compute_graph.flexml_layers[109].compute_node[3][0], compute_graph.flexml_layers[109].compute_node[3][1], compute_graph.flexml_layers[109].compute_node[3][2], compute_graph.flexml_layers[109].compute_node[3][3], {compute_graph.l2_124.out[0], compute_graph.l2_124.out[1], compute_graph.l2_124.out[2], compute_graph.l2_124.out[3], compute_graph.l2_125.in[0], compute_graph.l2_125.in[1], compute_graph.l2_125.in[2], compute_graph.l2_125.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1438), layer(138): compute_graph.flexml_layers[110].compute_node[0][0], compute_graph.flexml_layers[110].compute_node[0][1], compute_graph.flexml_layers[110].compute_node[0][2], compute_graph.flexml_layers[110].compute_node[0][3], compute_graph.flexml_layers[110].compute_node[1][0], compute_graph.flexml_layers[110].compute_node[1][1], compute_graph.flexml_layers[110].compute_node[1][2], compute_graph.flexml_layers[110].compute_node[1][3], compute_graph.flexml_layers[110].compute_node[2][0], compute_graph.flexml_layers[110].compute_node[2][1], compute_graph.flexml_layers[110].compute_node[2][2], compute_graph.flexml_layers[110].compute_node[2][3], compute_graph.flexml_layers[110].compute_node[3][0], compute_graph.flexml_layers[110].compute_node[3][1], compute_graph.flexml_layers[110].compute_node[3][2], compute_graph.flexml_layers[110].compute_node[3][3], {compute_graph.l2_125.out[0], compute_graph.l2_125.out[1], compute_graph.l2_125.out[2], compute_graph.l2_125.out[3], compute_graph.l2_126.in[0], compute_graph.l2_126.in[1], compute_graph.l2_126.in[2], compute_graph.l2_126.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1439), layer(139): compute_graph.flexml_layers[111].compute_node[0][0], compute_graph.flexml_layers[111].compute_node[0][1], compute_graph.flexml_layers[111].compute_node[0][2], compute_graph.flexml_layers[111].compute_node[0][3], compute_graph.flexml_layers[111].compute_node[1][0], compute_graph.flexml_layers[111].compute_node[1][1], compute_graph.flexml_layers[111].compute_node[1][2], compute_graph.flexml_layers[111].compute_node[1][3], compute_graph.flexml_layers[111].compute_node[2][0], compute_graph.flexml_layers[111].compute_node[2][1], compute_graph.flexml_layers[111].compute_node[2][2], compute_graph.flexml_layers[111].compute_node[2][3], compute_graph.flexml_layers[111].compute_node[3][0], compute_graph.flexml_layers[111].compute_node[3][1], compute_graph.flexml_layers[111].compute_node[3][2], compute_graph.flexml_layers[111].compute_node[3][3], {compute_graph.l2_126.out[0], compute_graph.l2_126.out[1], compute_graph.l2_126.out[2], compute_graph.l2_126.out[3], compute_graph.l2_127.in[0], compute_graph.l2_127.in[1], compute_graph.l2_127.in[2], compute_graph.l2_127.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1440), layer(140): compute_graph.flexml_layers[112].compute_node[0][0], compute_graph.flexml_layers[112].compute_node[0][1], compute_graph.flexml_layers[112].compute_node[0][2], compute_graph.flexml_layers[112].compute_node[0][3], compute_graph.flexml_layers[112].compute_node[1][0], compute_graph.flexml_layers[112].compute_node[1][1], compute_graph.flexml_layers[112].compute_node[1][2], compute_graph.flexml_layers[112].compute_node[1][3], compute_graph.flexml_layers[112].compute_node[2][0], compute_graph.flexml_layers[112].compute_node[2][1], compute_graph.flexml_layers[112].compute_node[2][2], compute_graph.flexml_layers[112].compute_node[2][3], compute_graph.flexml_layers[112].compute_node[3][0], compute_graph.flexml_layers[112].compute_node[3][1], compute_graph.flexml_layers[112].compute_node[3][2], compute_graph.flexml_layers[112].compute_node[3][3], {compute_graph.l2_124.out[4], compute_graph.l2_124.out[5], compute_graph.l2_124.out[6], compute_graph.l2_124.out[7], compute_graph.l2_127.out[0], compute_graph.l2_127.out[1], compute_graph.l2_127.out[2], compute_graph.l2_127.out[3], compute_graph.l2_128.in[0], compute_graph.l2_128.in[1], compute_graph.l2_128.in[2], compute_graph.l2_128.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(861): mode(3), layer(141): {compute_graph.Layer_129_wts_ddr.out[0], compute_graph.Layer_129_wts_ddr.out[1], compute_graph.Layer_129_l2_wts.in[0], compute_graph.Layer_129_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(861) is pipelined with KernelLayerNode(1440) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1441), layer(141): compute_graph.flexml_layers[113].compute_node[0][0], compute_graph.flexml_layers[113].compute_node[0][1], compute_graph.flexml_layers[113].compute_node[0][2], compute_graph.flexml_layers[113].compute_node[0][3], compute_graph.flexml_layers[113].compute_node[1][0], compute_graph.flexml_layers[113].compute_node[1][1], compute_graph.flexml_layers[113].compute_node[1][2], compute_graph.flexml_layers[113].compute_node[1][3], compute_graph.flexml_layers[113].compute_node[2][0], compute_graph.flexml_layers[113].compute_node[2][1], compute_graph.flexml_layers[113].compute_node[2][2], compute_graph.flexml_layers[113].compute_node[2][3], compute_graph.flexml_layers[113].compute_node[3][0], compute_graph.flexml_layers[113].compute_node[3][1], compute_graph.flexml_layers[113].compute_node[3][2], compute_graph.flexml_layers[113].compute_node[3][3], {compute_graph.l2_128.out[0], compute_graph.l2_128.out[1], compute_graph.l2_128.out[2], compute_graph.l2_128.out[3], compute_graph.Layer_129_l2_wts.out[0], compute_graph.Layer_129_l2_wts.out[1], compute_graph.Layer_129_l2_wts.out[2], compute_graph.Layer_129_l2_wts.out[3], compute_graph.l2_129.in[0], compute_graph.l2_129.in[1], compute_graph.l2_129.in[2], compute_graph.l2_129.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1442), layer(142): compute_graph.flexml_layers[114].compute_node[0][0], compute_graph.flexml_layers[114].compute_node[0][1], compute_graph.flexml_layers[114].compute_node[0][2], compute_graph.flexml_layers[114].compute_node[0][3], compute_graph.flexml_layers[114].compute_node[1][0], compute_graph.flexml_layers[114].compute_node[1][1], compute_graph.flexml_layers[114].compute_node[1][2], compute_graph.flexml_layers[114].compute_node[1][3], compute_graph.flexml_layers[114].compute_node[2][0], compute_graph.flexml_layers[114].compute_node[2][1], compute_graph.flexml_layers[114].compute_node[2][2], compute_graph.flexml_layers[114].compute_node[2][3], compute_graph.flexml_layers[114].compute_node[3][0], compute_graph.flexml_layers[114].compute_node[3][1], compute_graph.flexml_layers[114].compute_node[3][2], compute_graph.flexml_layers[114].compute_node[3][3], {compute_graph.l2_129.out[0], compute_graph.l2_129.out[1], compute_graph.l2_129.out[2], compute_graph.l2_129.out[3], compute_graph.l2_130.in[0], compute_graph.l2_130.in[1], compute_graph.l2_130.in[2], compute_graph.l2_130.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1443), layer(143): compute_graph.flexml_layers[115].compute_node[0][0], compute_graph.flexml_layers[115].compute_node[0][1], compute_graph.flexml_layers[115].compute_node[0][2], compute_graph.flexml_layers[115].compute_node[0][3], compute_graph.flexml_layers[115].compute_node[1][0], compute_graph.flexml_layers[115].compute_node[1][1], compute_graph.flexml_layers[115].compute_node[1][2], compute_graph.flexml_layers[115].compute_node[1][3], compute_graph.flexml_layers[115].compute_node[2][0], compute_graph.flexml_layers[115].compute_node[2][1], compute_graph.flexml_layers[115].compute_node[2][2], compute_graph.flexml_layers[115].compute_node[2][3], compute_graph.flexml_layers[115].compute_node[3][0], compute_graph.flexml_layers[115].compute_node[3][1], compute_graph.flexml_layers[115].compute_node[3][2], compute_graph.flexml_layers[115].compute_node[3][3], {compute_graph.l2_130.out[0], compute_graph.l2_130.out[1], compute_graph.l2_130.out[2], compute_graph.l2_130.out[3], compute_graph.l2_131.in[0], compute_graph.l2_131.in[1], compute_graph.l2_131.in[2], compute_graph.l2_131.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1444), layer(144): compute_graph.flexml_layers[116].compute_node[0][0], compute_graph.flexml_layers[116].compute_node[0][1], compute_graph.flexml_layers[116].compute_node[0][2], compute_graph.flexml_layers[116].compute_node[0][3], compute_graph.flexml_layers[116].compute_node[1][0], compute_graph.flexml_layers[116].compute_node[1][1], compute_graph.flexml_layers[116].compute_node[1][2], compute_graph.flexml_layers[116].compute_node[1][3], compute_graph.flexml_layers[116].compute_node[2][0], compute_graph.flexml_layers[116].compute_node[2][1], compute_graph.flexml_layers[116].compute_node[2][2], compute_graph.flexml_layers[116].compute_node[2][3], compute_graph.flexml_layers[116].compute_node[3][0], compute_graph.flexml_layers[116].compute_node[3][1], compute_graph.flexml_layers[116].compute_node[3][2], compute_graph.flexml_layers[116].compute_node[3][3], {compute_graph.l2_131.out[0], compute_graph.l2_131.out[1], compute_graph.l2_131.out[2], compute_graph.l2_131.out[3], compute_graph.l2_132.in[0], compute_graph.l2_132.in[1], compute_graph.l2_132.in[2], compute_graph.l2_132.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1445), layer(145): compute_graph.flexml_layers[117].compute_node[0][0], compute_graph.flexml_layers[117].compute_node[0][1], compute_graph.flexml_layers[117].compute_node[0][2], compute_graph.flexml_layers[117].compute_node[0][3], compute_graph.flexml_layers[117].compute_node[1][0], compute_graph.flexml_layers[117].compute_node[1][1], compute_graph.flexml_layers[117].compute_node[1][2], compute_graph.flexml_layers[117].compute_node[1][3], compute_graph.flexml_layers[117].compute_node[2][0], compute_graph.flexml_layers[117].compute_node[2][1], compute_graph.flexml_layers[117].compute_node[2][2], compute_graph.flexml_layers[117].compute_node[2][3], compute_graph.flexml_layers[117].compute_node[3][0], compute_graph.flexml_layers[117].compute_node[3][1], compute_graph.flexml_layers[117].compute_node[3][2], compute_graph.flexml_layers[117].compute_node[3][3], {compute_graph.l2_129.out[4], compute_graph.l2_129.out[5], compute_graph.l2_129.out[6], compute_graph.l2_129.out[7], compute_graph.l2_132.out[0], compute_graph.l2_132.out[1], compute_graph.l2_132.out[2], compute_graph.l2_132.out[3], compute_graph.l2_133.in[0], compute_graph.l2_133.in[1], compute_graph.l2_133.in[2], compute_graph.l2_133.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(933): mode(0), layer(145): {compute_graph.l2_133.out[0], compute_graph.l2_133.out[1], compute_graph.l2l3_133_spill.in[0], compute_graph.l2l3_133_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(934): mode(0), layer(145): {compute_graph.l2_133.out[2], compute_graph.l2_133.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1082): mode(0), layer(146): {compute_graph.l2l3_133_spill.out[0], compute_graph.templated_graph_134.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1263): mode(0), layer(146): {compute_graph.templated_graph_134.trans_mt_ifm.out[0], compute_graph.templated_graph_134.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1264): mode(0), layer(146): {compute_graph.templated_graph_134.trans_mt_ofm.out[0], compute_graph.l2l3_134_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1083): mode(0), layer(147): {compute_graph.l2l3_134_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1446), layer(147): compute_graph.flexml_layers[118].compute_node[0][0], compute_graph.flexml_layers[118].compute_node[0][1], compute_graph.flexml_layers[118].compute_node[0][2], compute_graph.flexml_layers[118].compute_node[0][3], compute_graph.flexml_layers[118].compute_node[1][0], compute_graph.flexml_layers[118].compute_node[1][1], compute_graph.flexml_layers[118].compute_node[1][2], compute_graph.flexml_layers[118].compute_node[1][3], compute_graph.flexml_layers[118].compute_node[2][0], compute_graph.flexml_layers[118].compute_node[2][1], compute_graph.flexml_layers[118].compute_node[2][2], compute_graph.flexml_layers[118].compute_node[2][3], compute_graph.flexml_layers[118].compute_node[3][0], compute_graph.flexml_layers[118].compute_node[3][1], compute_graph.flexml_layers[118].compute_node[3][2], compute_graph.flexml_layers[118].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[3], compute_graph.l2_135.in[0], compute_graph.l2_135.in[1], compute_graph.l2_135.in[2], compute_graph.l2_135.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(862): mode(3), layer(148): {compute_graph.Layer_136_wts_ddr.out[0], compute_graph.Layer_136_wts_ddr.out[1], compute_graph.Layer_136_l2_wts.in[0], compute_graph.Layer_136_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(862) is pipelined with KernelLayerNode(1446) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1447), layer(148): compute_graph.flexml_layers[119].compute_node[0][0], compute_graph.flexml_layers[119].compute_node[0][1], compute_graph.flexml_layers[119].compute_node[0][2], compute_graph.flexml_layers[119].compute_node[0][3], compute_graph.flexml_layers[119].compute_node[1][0], compute_graph.flexml_layers[119].compute_node[1][1], compute_graph.flexml_layers[119].compute_node[1][2], compute_graph.flexml_layers[119].compute_node[1][3], compute_graph.flexml_layers[119].compute_node[2][0], compute_graph.flexml_layers[119].compute_node[2][1], compute_graph.flexml_layers[119].compute_node[2][2], compute_graph.flexml_layers[119].compute_node[2][3], compute_graph.flexml_layers[119].compute_node[3][0], compute_graph.flexml_layers[119].compute_node[3][1], compute_graph.flexml_layers[119].compute_node[3][2], compute_graph.flexml_layers[119].compute_node[3][3], {compute_graph.l2_135.out[0], compute_graph.l2_135.out[1], compute_graph.l2_135.out[2], compute_graph.l2_135.out[3], compute_graph.Layer_136_l2_wts.out[0], compute_graph.Layer_136_l2_wts.out[1], compute_graph.Layer_136_l2_wts.out[2], compute_graph.Layer_136_l2_wts.out[3], compute_graph.l2_136.in[0], compute_graph.l2_136.in[1], compute_graph.l2_136.in[2], compute_graph.l2_136.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(863): mode(3), layer(149): {compute_graph.Layer_137_wts_ddr.out[0], compute_graph.Layer_137_wts_ddr.out[1], compute_graph.Layer_137_l2_wts.in[0], compute_graph.Layer_137_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(863) is pipelined with KernelLayerNode(1447) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1448), layer(149): compute_graph.flexml_layers[120].compute_node[0][0], compute_graph.flexml_layers[120].compute_node[0][1], compute_graph.flexml_layers[120].compute_node[0][2], compute_graph.flexml_layers[120].compute_node[0][3], compute_graph.flexml_layers[120].compute_node[1][0], compute_graph.flexml_layers[120].compute_node[1][1], compute_graph.flexml_layers[120].compute_node[1][2], compute_graph.flexml_layers[120].compute_node[1][3], compute_graph.flexml_layers[120].compute_node[2][0], compute_graph.flexml_layers[120].compute_node[2][1], compute_graph.flexml_layers[120].compute_node[2][2], compute_graph.flexml_layers[120].compute_node[2][3], compute_graph.flexml_layers[120].compute_node[3][0], compute_graph.flexml_layers[120].compute_node[3][1], compute_graph.flexml_layers[120].compute_node[3][2], compute_graph.flexml_layers[120].compute_node[3][3], {compute_graph.l2_136.out[0], compute_graph.l2_136.out[1], compute_graph.l2_136.out[2], compute_graph.l2_136.out[3], compute_graph.Layer_137_l2_wts.out[0], compute_graph.Layer_137_l2_wts.out[1], compute_graph.Layer_137_l2_wts.out[2], compute_graph.Layer_137_l2_wts.out[3], compute_graph.l2_137.in[0], compute_graph.l2_137.in[1], compute_graph.l2_137.in[2], compute_graph.l2_137.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1449), layer(150): compute_graph.flexml_layers[121].compute_node[0][0], compute_graph.flexml_layers[121].compute_node[0][1], compute_graph.flexml_layers[121].compute_node[0][2], compute_graph.flexml_layers[121].compute_node[0][3], compute_graph.flexml_layers[121].compute_node[1][0], compute_graph.flexml_layers[121].compute_node[1][1], compute_graph.flexml_layers[121].compute_node[1][2], compute_graph.flexml_layers[121].compute_node[1][3], compute_graph.flexml_layers[121].compute_node[2][0], compute_graph.flexml_layers[121].compute_node[2][1], compute_graph.flexml_layers[121].compute_node[2][2], compute_graph.flexml_layers[121].compute_node[2][3], compute_graph.flexml_layers[121].compute_node[3][0], compute_graph.flexml_layers[121].compute_node[3][1], compute_graph.flexml_layers[121].compute_node[3][2], compute_graph.flexml_layers[121].compute_node[3][3], {compute_graph.l2_137.out[0], compute_graph.l2_137.out[1], compute_graph.l2_137.out[2], compute_graph.l2_137.out[3], compute_graph.l2_138.in[0], compute_graph.l2_138.in[1], compute_graph.l2_138.in[2], compute_graph.l2_138.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1450), layer(151): compute_graph.flexml_layers[122].compute_node[0][0], compute_graph.flexml_layers[122].compute_node[0][1], compute_graph.flexml_layers[122].compute_node[0][2], compute_graph.flexml_layers[122].compute_node[0][3], compute_graph.flexml_layers[122].compute_node[1][0], compute_graph.flexml_layers[122].compute_node[1][1], compute_graph.flexml_layers[122].compute_node[1][2], compute_graph.flexml_layers[122].compute_node[1][3], compute_graph.flexml_layers[122].compute_node[2][0], compute_graph.flexml_layers[122].compute_node[2][1], compute_graph.flexml_layers[122].compute_node[2][2], compute_graph.flexml_layers[122].compute_node[2][3], compute_graph.flexml_layers[122].compute_node[3][0], compute_graph.flexml_layers[122].compute_node[3][1], compute_graph.flexml_layers[122].compute_node[3][2], compute_graph.flexml_layers[122].compute_node[3][3], {compute_graph.l2_138.out[0], compute_graph.l2_138.out[1], compute_graph.l2_138.out[2], compute_graph.l2_138.out[3], compute_graph.l2_139.in[0], compute_graph.l2_139.in[1], compute_graph.l2_139.in[2], compute_graph.l2_139.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1451), layer(152): compute_graph.flexml_layers[123].compute_node[0][0], compute_graph.flexml_layers[123].compute_node[0][1], compute_graph.flexml_layers[123].compute_node[0][2], compute_graph.flexml_layers[123].compute_node[0][3], compute_graph.flexml_layers[123].compute_node[1][0], compute_graph.flexml_layers[123].compute_node[1][1], compute_graph.flexml_layers[123].compute_node[1][2], compute_graph.flexml_layers[123].compute_node[1][3], compute_graph.flexml_layers[123].compute_node[2][0], compute_graph.flexml_layers[123].compute_node[2][1], compute_graph.flexml_layers[123].compute_node[2][2], compute_graph.flexml_layers[123].compute_node[2][3], compute_graph.flexml_layers[123].compute_node[3][0], compute_graph.flexml_layers[123].compute_node[3][1], compute_graph.flexml_layers[123].compute_node[3][2], compute_graph.flexml_layers[123].compute_node[3][3], {compute_graph.l2_139.out[0], compute_graph.l2_139.out[1], compute_graph.l2_139.out[2], compute_graph.l2_139.out[3], compute_graph.l2_140.in[0], compute_graph.l2_140.in[1], compute_graph.l2_140.in[2], compute_graph.l2_140.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(935): mode(0), layer(152): {compute_graph.l2_140.out[1], compute_graph.l2l3_140_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1084): mode(0), layer(153): {compute_graph.l2l3_140_spill.out[0], compute_graph.templated_graph_141.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1265): mode(0), layer(153): {compute_graph.templated_graph_141.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_141_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1085): mode(0), layer(154): {compute_graph.l2l3_scratch_0_141_spill.out[0], compute_graph.templated_graph_141.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1266): mode(0), layer(154): {compute_graph.templated_graph_141.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_141_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1086): mode(0), layer(155): {compute_graph.l2l3_scratch_1_141_spill.out[0], compute_graph.templated_graph_141.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1267): mode(0), layer(155): {compute_graph.templated_graph_141.g2.ifm_mem.out[0], compute_graph.l2l3_141_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1081): mode(0), layer(156): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1087): mode(0), layer(156): {compute_graph.l2l3_141_spill.out[0], compute_graph.l2l3_141_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1452), layer(156): compute_graph.flexml_layers[124].compute_node[0][0], compute_graph.flexml_layers[124].compute_node[0][1], compute_graph.flexml_layers[124].compute_node[0][2], compute_graph.flexml_layers[124].compute_node[0][3], compute_graph.flexml_layers[124].compute_node[1][0], compute_graph.flexml_layers[124].compute_node[1][1], compute_graph.flexml_layers[124].compute_node[1][2], compute_graph.flexml_layers[124].compute_node[1][3], compute_graph.flexml_layers[124].compute_node[2][0], compute_graph.flexml_layers[124].compute_node[2][1], compute_graph.flexml_layers[124].compute_node[2][2], compute_graph.flexml_layers[124].compute_node[2][3], compute_graph.flexml_layers[124].compute_node[3][0], compute_graph.flexml_layers[124].compute_node[3][1], compute_graph.flexml_layers[124].compute_node[3][2], compute_graph.flexml_layers[124].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[3], compute_graph.l2_142.in[0], compute_graph.l2_142.in[1], compute_graph.l2_142.in[2], compute_graph.l2_142.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(864): mode(3), layer(157): {compute_graph.Layer_143_wts_ddr.out[0], compute_graph.Layer_143_wts_ddr.out[1], compute_graph.Layer_143_l2_wts.in[0], compute_graph.Layer_143_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(864) is pipelined with KernelLayerNode(1452) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123 overlaps with memory space used by compute_graph.l2_143. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1080): mode(0), layer(157): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1453), layer(157): compute_graph.flexml_layers[125].compute_node[0][0], compute_graph.flexml_layers[125].compute_node[0][1], compute_graph.flexml_layers[125].compute_node[0][2], compute_graph.flexml_layers[125].compute_node[0][3], compute_graph.flexml_layers[125].compute_node[1][0], compute_graph.flexml_layers[125].compute_node[1][1], compute_graph.flexml_layers[125].compute_node[1][2], compute_graph.flexml_layers[125].compute_node[1][3], compute_graph.flexml_layers[125].compute_node[2][0], compute_graph.flexml_layers[125].compute_node[2][1], compute_graph.flexml_layers[125].compute_node[2][2], compute_graph.flexml_layers[125].compute_node[2][3], compute_graph.flexml_layers[125].compute_node[3][0], compute_graph.flexml_layers[125].compute_node[3][1], compute_graph.flexml_layers[125].compute_node[3][2], compute_graph.flexml_layers[125].compute_node[3][3], {compute_graph.l2_142.out[0], compute_graph.l2_142.out[1], compute_graph.l2_142.out[2], compute_graph.l2_142.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[3], compute_graph.Layer_143_l2_wts.out[0], compute_graph.Layer_143_l2_wts.out[1], compute_graph.Layer_143_l2_wts.out[2], compute_graph.Layer_143_l2_wts.out[3], compute_graph.l2_143.in[0], compute_graph.l2_143.in[1], compute_graph.l2_143.in[2], compute_graph.l2_143.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(865): mode(3), layer(158): {compute_graph.Layer_144_wts_ddr.out[0], compute_graph.Layer_144_wts_ddr.out[1], compute_graph.Layer_144_l2_wts.in[0], compute_graph.Layer_144_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(865) is pipelined with KernelLayerNode(1453) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1454), layer(158): compute_graph.flexml_layers[126].compute_node[0][0], compute_graph.flexml_layers[126].compute_node[0][1], compute_graph.flexml_layers[126].compute_node[0][2], compute_graph.flexml_layers[126].compute_node[0][3], compute_graph.flexml_layers[126].compute_node[1][0], compute_graph.flexml_layers[126].compute_node[1][1], compute_graph.flexml_layers[126].compute_node[1][2], compute_graph.flexml_layers[126].compute_node[1][3], compute_graph.flexml_layers[126].compute_node[2][0], compute_graph.flexml_layers[126].compute_node[2][1], compute_graph.flexml_layers[126].compute_node[2][2], compute_graph.flexml_layers[126].compute_node[2][3], compute_graph.flexml_layers[126].compute_node[3][0], compute_graph.flexml_layers[126].compute_node[3][1], compute_graph.flexml_layers[126].compute_node[3][2], compute_graph.flexml_layers[126].compute_node[3][3], {compute_graph.l2_143.out[0], compute_graph.l2_143.out[1], compute_graph.l2_143.out[2], compute_graph.l2_143.out[3], compute_graph.Layer_144_l2_wts.out[0], compute_graph.Layer_144_l2_wts.out[1], compute_graph.Layer_144_l2_wts.out[2], compute_graph.Layer_144_l2_wts.out[3], compute_graph.l2_144.in[0], compute_graph.l2_144.in[1], compute_graph.l2_144.in[2], compute_graph.l2_144.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1455), layer(159): compute_graph.flexml_layers[127].compute_node[0][0], compute_graph.flexml_layers[127].compute_node[0][1], compute_graph.flexml_layers[127].compute_node[0][2], compute_graph.flexml_layers[127].compute_node[0][3], compute_graph.flexml_layers[127].compute_node[1][0], compute_graph.flexml_layers[127].compute_node[1][1], compute_graph.flexml_layers[127].compute_node[1][2], compute_graph.flexml_layers[127].compute_node[1][3], compute_graph.flexml_layers[127].compute_node[2][0], compute_graph.flexml_layers[127].compute_node[2][1], compute_graph.flexml_layers[127].compute_node[2][2], compute_graph.flexml_layers[127].compute_node[2][3], compute_graph.flexml_layers[127].compute_node[3][0], compute_graph.flexml_layers[127].compute_node[3][1], compute_graph.flexml_layers[127].compute_node[3][2], compute_graph.flexml_layers[127].compute_node[3][3], {compute_graph.l2_144.out[0], compute_graph.l2_144.out[1], compute_graph.l2_144.out[2], compute_graph.l2_144.out[3], compute_graph.l2_145.in[0], compute_graph.l2_145.in[1], compute_graph.l2_145.in[2], compute_graph.l2_145.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1456), layer(160): compute_graph.flexml_layers[128].compute_node[0][0], compute_graph.flexml_layers[128].compute_node[0][1], compute_graph.flexml_layers[128].compute_node[0][2], compute_graph.flexml_layers[128].compute_node[0][3], compute_graph.flexml_layers[128].compute_node[1][0], compute_graph.flexml_layers[128].compute_node[1][1], compute_graph.flexml_layers[128].compute_node[1][2], compute_graph.flexml_layers[128].compute_node[1][3], compute_graph.flexml_layers[128].compute_node[2][0], compute_graph.flexml_layers[128].compute_node[2][1], compute_graph.flexml_layers[128].compute_node[2][2], compute_graph.flexml_layers[128].compute_node[2][3], compute_graph.flexml_layers[128].compute_node[3][0], compute_graph.flexml_layers[128].compute_node[3][1], compute_graph.flexml_layers[128].compute_node[3][2], compute_graph.flexml_layers[128].compute_node[3][3], {compute_graph.l2_145.out[0], compute_graph.l2_145.out[1], compute_graph.l2_145.out[2], compute_graph.l2_145.out[3], compute_graph.l2_146.in[0], compute_graph.l2_146.in[1], compute_graph.l2_146.in[2], compute_graph.l2_146.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1457), layer(161): compute_graph.flexml_layers[129].compute_node[0][0], compute_graph.flexml_layers[129].compute_node[0][1], compute_graph.flexml_layers[129].compute_node[0][2], compute_graph.flexml_layers[129].compute_node[0][3], compute_graph.flexml_layers[129].compute_node[1][0], compute_graph.flexml_layers[129].compute_node[1][1], compute_graph.flexml_layers[129].compute_node[1][2], compute_graph.flexml_layers[129].compute_node[1][3], compute_graph.flexml_layers[129].compute_node[2][0], compute_graph.flexml_layers[129].compute_node[2][1], compute_graph.flexml_layers[129].compute_node[2][2], compute_graph.flexml_layers[129].compute_node[2][3], compute_graph.flexml_layers[129].compute_node[3][0], compute_graph.flexml_layers[129].compute_node[3][1], compute_graph.flexml_layers[129].compute_node[3][2], compute_graph.flexml_layers[129].compute_node[3][3], {compute_graph.l2_146.out[0], compute_graph.l2_146.out[1], compute_graph.l2_146.out[2], compute_graph.l2_146.out[3], compute_graph.l2_147.in[0], compute_graph.l2_147.in[1], compute_graph.l2_147.in[2], compute_graph.l2_147.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1458), layer(162): compute_graph.flexml_layers[130].compute_node[0][0], compute_graph.flexml_layers[130].compute_node[0][1], compute_graph.flexml_layers[130].compute_node[0][2], compute_graph.flexml_layers[130].compute_node[0][3], compute_graph.flexml_layers[130].compute_node[1][0], compute_graph.flexml_layers[130].compute_node[1][1], compute_graph.flexml_layers[130].compute_node[1][2], compute_graph.flexml_layers[130].compute_node[1][3], compute_graph.flexml_layers[130].compute_node[2][0], compute_graph.flexml_layers[130].compute_node[2][1], compute_graph.flexml_layers[130].compute_node[2][2], compute_graph.flexml_layers[130].compute_node[2][3], compute_graph.flexml_layers[130].compute_node[3][0], compute_graph.flexml_layers[130].compute_node[3][1], compute_graph.flexml_layers[130].compute_node[3][2], compute_graph.flexml_layers[130].compute_node[3][3], {compute_graph.l2_144.out[4], compute_graph.l2_144.out[5], compute_graph.l2_144.out[6], compute_graph.l2_144.out[7], compute_graph.l2_147.out[0], compute_graph.l2_147.out[1], compute_graph.l2_147.out[2], compute_graph.l2_147.out[3], compute_graph.l2_148.in[0], compute_graph.l2_148.in[1], compute_graph.l2_148.in[2], compute_graph.l2_148.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(866): mode(3), layer(163): {compute_graph.Layer_149_wts_ddr.out[0], compute_graph.Layer_149_wts_ddr.out[1], compute_graph.Layer_149_l2_wts.in[0], compute_graph.Layer_149_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(866) is pipelined with KernelLayerNode(1458) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1459), layer(163): compute_graph.flexml_layers[131].compute_node[0][0], compute_graph.flexml_layers[131].compute_node[0][1], compute_graph.flexml_layers[131].compute_node[0][2], compute_graph.flexml_layers[131].compute_node[0][3], compute_graph.flexml_layers[131].compute_node[1][0], compute_graph.flexml_layers[131].compute_node[1][1], compute_graph.flexml_layers[131].compute_node[1][2], compute_graph.flexml_layers[131].compute_node[1][3], compute_graph.flexml_layers[131].compute_node[2][0], compute_graph.flexml_layers[131].compute_node[2][1], compute_graph.flexml_layers[131].compute_node[2][2], compute_graph.flexml_layers[131].compute_node[2][3], compute_graph.flexml_layers[131].compute_node[3][0], compute_graph.flexml_layers[131].compute_node[3][1], compute_graph.flexml_layers[131].compute_node[3][2], compute_graph.flexml_layers[131].compute_node[3][3], {compute_graph.l2_148.out[0], compute_graph.l2_148.out[1], compute_graph.l2_148.out[2], compute_graph.l2_148.out[3], compute_graph.Layer_149_l2_wts.out[0], compute_graph.Layer_149_l2_wts.out[1], compute_graph.Layer_149_l2_wts.out[2], compute_graph.Layer_149_l2_wts.out[3], compute_graph.l2_149.in[0], compute_graph.l2_149.in[1], compute_graph.l2_149.in[2], compute_graph.l2_149.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1460), layer(164): compute_graph.flexml_layers[132].compute_node[0][0], compute_graph.flexml_layers[132].compute_node[0][1], compute_graph.flexml_layers[132].compute_node[0][2], compute_graph.flexml_layers[132].compute_node[0][3], compute_graph.flexml_layers[132].compute_node[1][0], compute_graph.flexml_layers[132].compute_node[1][1], compute_graph.flexml_layers[132].compute_node[1][2], compute_graph.flexml_layers[132].compute_node[1][3], compute_graph.flexml_layers[132].compute_node[2][0], compute_graph.flexml_layers[132].compute_node[2][1], compute_graph.flexml_layers[132].compute_node[2][2], compute_graph.flexml_layers[132].compute_node[2][3], compute_graph.flexml_layers[132].compute_node[3][0], compute_graph.flexml_layers[132].compute_node[3][1], compute_graph.flexml_layers[132].compute_node[3][2], compute_graph.flexml_layers[132].compute_node[3][3], {compute_graph.l2_149.out[0], compute_graph.l2_149.out[1], compute_graph.l2_149.out[2], compute_graph.l2_149.out[3], compute_graph.l2_150.in[0], compute_graph.l2_150.in[1], compute_graph.l2_150.in[2], compute_graph.l2_150.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1461), layer(165): compute_graph.flexml_layers[133].compute_node[0][0], compute_graph.flexml_layers[133].compute_node[0][1], compute_graph.flexml_layers[133].compute_node[0][2], compute_graph.flexml_layers[133].compute_node[0][3], compute_graph.flexml_layers[133].compute_node[1][0], compute_graph.flexml_layers[133].compute_node[1][1], compute_graph.flexml_layers[133].compute_node[1][2], compute_graph.flexml_layers[133].compute_node[1][3], compute_graph.flexml_layers[133].compute_node[2][0], compute_graph.flexml_layers[133].compute_node[2][1], compute_graph.flexml_layers[133].compute_node[2][2], compute_graph.flexml_layers[133].compute_node[2][3], compute_graph.flexml_layers[133].compute_node[3][0], compute_graph.flexml_layers[133].compute_node[3][1], compute_graph.flexml_layers[133].compute_node[3][2], compute_graph.flexml_layers[133].compute_node[3][3], {compute_graph.l2_150.out[0], compute_graph.l2_150.out[1], compute_graph.l2_150.out[2], compute_graph.l2_150.out[3], compute_graph.l2_151.in[0], compute_graph.l2_151.in[1], compute_graph.l2_151.in[2], compute_graph.l2_151.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1462), layer(166): compute_graph.flexml_layers[134].compute_node[0][0], compute_graph.flexml_layers[134].compute_node[0][1], compute_graph.flexml_layers[134].compute_node[0][2], compute_graph.flexml_layers[134].compute_node[0][3], compute_graph.flexml_layers[134].compute_node[1][0], compute_graph.flexml_layers[134].compute_node[1][1], compute_graph.flexml_layers[134].compute_node[1][2], compute_graph.flexml_layers[134].compute_node[1][3], compute_graph.flexml_layers[134].compute_node[2][0], compute_graph.flexml_layers[134].compute_node[2][1], compute_graph.flexml_layers[134].compute_node[2][2], compute_graph.flexml_layers[134].compute_node[2][3], compute_graph.flexml_layers[134].compute_node[3][0], compute_graph.flexml_layers[134].compute_node[3][1], compute_graph.flexml_layers[134].compute_node[3][2], compute_graph.flexml_layers[134].compute_node[3][3], {compute_graph.l2_151.out[0], compute_graph.l2_151.out[1], compute_graph.l2_151.out[2], compute_graph.l2_151.out[3], compute_graph.l2_152.in[0], compute_graph.l2_152.in[1], compute_graph.l2_152.in[2], compute_graph.l2_152.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1463), layer(167): compute_graph.flexml_layers[135].compute_node[0][0], compute_graph.flexml_layers[135].compute_node[0][1], compute_graph.flexml_layers[135].compute_node[0][2], compute_graph.flexml_layers[135].compute_node[0][3], compute_graph.flexml_layers[135].compute_node[1][0], compute_graph.flexml_layers[135].compute_node[1][1], compute_graph.flexml_layers[135].compute_node[1][2], compute_graph.flexml_layers[135].compute_node[1][3], compute_graph.flexml_layers[135].compute_node[2][0], compute_graph.flexml_layers[135].compute_node[2][1], compute_graph.flexml_layers[135].compute_node[2][2], compute_graph.flexml_layers[135].compute_node[2][3], compute_graph.flexml_layers[135].compute_node[3][0], compute_graph.flexml_layers[135].compute_node[3][1], compute_graph.flexml_layers[135].compute_node[3][2], compute_graph.flexml_layers[135].compute_node[3][3], {compute_graph.l2_149.out[4], compute_graph.l2_149.out[5], compute_graph.l2_149.out[6], compute_graph.l2_149.out[7], compute_graph.l2_152.out[0], compute_graph.l2_152.out[1], compute_graph.l2_152.out[2], compute_graph.l2_152.out[3], compute_graph.l2_153.in[0], compute_graph.l2_153.in[1], compute_graph.l2_153.in[2], compute_graph.l2_153.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(937): mode(0), layer(167): {compute_graph.l2_153.out[2], compute_graph.l2_153.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(936): mode(0), layer(167): {compute_graph.l2_153.out[0], compute_graph.l2_153.out[1], compute_graph.l2l3_153_spill.in[0], compute_graph.l2l3_153_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1089): mode(0), layer(168): {compute_graph.l2l3_153_spill.out[0], compute_graph.templated_graph_154.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1268): mode(0), layer(168): {compute_graph.templated_graph_154.trans_mt_ifm.out[0], compute_graph.templated_graph_154.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1269): mode(0), layer(168): {compute_graph.templated_graph_154.trans_mt_ofm.out[0], compute_graph.l2l3_154_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1090): mode(0), layer(169): {compute_graph.l2l3_154_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1464), layer(169): compute_graph.flexml_layers[136].compute_node[0][0], compute_graph.flexml_layers[136].compute_node[0][1], compute_graph.flexml_layers[136].compute_node[0][2], compute_graph.flexml_layers[136].compute_node[0][3], compute_graph.flexml_layers[136].compute_node[1][0], compute_graph.flexml_layers[136].compute_node[1][1], compute_graph.flexml_layers[136].compute_node[1][2], compute_graph.flexml_layers[136].compute_node[1][3], compute_graph.flexml_layers[136].compute_node[2][0], compute_graph.flexml_layers[136].compute_node[2][1], compute_graph.flexml_layers[136].compute_node[2][2], compute_graph.flexml_layers[136].compute_node[2][3], compute_graph.flexml_layers[136].compute_node[3][0], compute_graph.flexml_layers[136].compute_node[3][1], compute_graph.flexml_layers[136].compute_node[3][2], compute_graph.flexml_layers[136].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[3], compute_graph.l2_155.in[0], compute_graph.l2_155.in[1], compute_graph.l2_155.in[2], compute_graph.l2_155.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(867): mode(3), layer(170): {compute_graph.Layer_156_wts_ddr.out[0], compute_graph.Layer_156_wts_ddr.out[1], compute_graph.Layer_156_l2_wts.in[0], compute_graph.Layer_156_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(867) is pipelined with KernelLayerNode(1464) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1465), layer(170): compute_graph.flexml_layers[137].compute_node[0][0], compute_graph.flexml_layers[137].compute_node[0][1], compute_graph.flexml_layers[137].compute_node[0][2], compute_graph.flexml_layers[137].compute_node[0][3], compute_graph.flexml_layers[137].compute_node[1][0], compute_graph.flexml_layers[137].compute_node[1][1], compute_graph.flexml_layers[137].compute_node[1][2], compute_graph.flexml_layers[137].compute_node[1][3], compute_graph.flexml_layers[137].compute_node[2][0], compute_graph.flexml_layers[137].compute_node[2][1], compute_graph.flexml_layers[137].compute_node[2][2], compute_graph.flexml_layers[137].compute_node[2][3], compute_graph.flexml_layers[137].compute_node[3][0], compute_graph.flexml_layers[137].compute_node[3][1], compute_graph.flexml_layers[137].compute_node[3][2], compute_graph.flexml_layers[137].compute_node[3][3], {compute_graph.l2_155.out[0], compute_graph.l2_155.out[1], compute_graph.l2_155.out[2], compute_graph.l2_155.out[3], compute_graph.Layer_156_l2_wts.out[0], compute_graph.Layer_156_l2_wts.out[1], compute_graph.Layer_156_l2_wts.out[2], compute_graph.Layer_156_l2_wts.out[3], compute_graph.l2_156.in[0], compute_graph.l2_156.in[1], compute_graph.l2_156.in[2], compute_graph.l2_156.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(868): mode(3), layer(171): {compute_graph.Layer_157_wts_ddr.out[0], compute_graph.Layer_157_wts_ddr.out[1], compute_graph.Layer_157_l2_wts.in[0], compute_graph.Layer_157_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(868) is pipelined with KernelLayerNode(1465) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1466), layer(171): compute_graph.flexml_layers[138].compute_node[0][0], compute_graph.flexml_layers[138].compute_node[0][1], compute_graph.flexml_layers[138].compute_node[0][2], compute_graph.flexml_layers[138].compute_node[0][3], compute_graph.flexml_layers[138].compute_node[1][0], compute_graph.flexml_layers[138].compute_node[1][1], compute_graph.flexml_layers[138].compute_node[1][2], compute_graph.flexml_layers[138].compute_node[1][3], compute_graph.flexml_layers[138].compute_node[2][0], compute_graph.flexml_layers[138].compute_node[2][1], compute_graph.flexml_layers[138].compute_node[2][2], compute_graph.flexml_layers[138].compute_node[2][3], compute_graph.flexml_layers[138].compute_node[3][0], compute_graph.flexml_layers[138].compute_node[3][1], compute_graph.flexml_layers[138].compute_node[3][2], compute_graph.flexml_layers[138].compute_node[3][3], {compute_graph.l2_156.out[0], compute_graph.l2_156.out[1], compute_graph.l2_156.out[2], compute_graph.l2_156.out[3], compute_graph.Layer_157_l2_wts.out[0], compute_graph.Layer_157_l2_wts.out[1], compute_graph.Layer_157_l2_wts.out[2], compute_graph.Layer_157_l2_wts.out[3], compute_graph.l2_157.in[0], compute_graph.l2_157.in[1], compute_graph.l2_157.in[2], compute_graph.l2_157.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1467), layer(172): compute_graph.flexml_layers[139].compute_node[0][0], compute_graph.flexml_layers[139].compute_node[0][1], compute_graph.flexml_layers[139].compute_node[0][2], compute_graph.flexml_layers[139].compute_node[0][3], compute_graph.flexml_layers[139].compute_node[1][0], compute_graph.flexml_layers[139].compute_node[1][1], compute_graph.flexml_layers[139].compute_node[1][2], compute_graph.flexml_layers[139].compute_node[1][3], compute_graph.flexml_layers[139].compute_node[2][0], compute_graph.flexml_layers[139].compute_node[2][1], compute_graph.flexml_layers[139].compute_node[2][2], compute_graph.flexml_layers[139].compute_node[2][3], compute_graph.flexml_layers[139].compute_node[3][0], compute_graph.flexml_layers[139].compute_node[3][1], compute_graph.flexml_layers[139].compute_node[3][2], compute_graph.flexml_layers[139].compute_node[3][3], {compute_graph.l2_157.out[0], compute_graph.l2_157.out[1], compute_graph.l2_157.out[2], compute_graph.l2_157.out[3], compute_graph.l2_158.in[0], compute_graph.l2_158.in[1], compute_graph.l2_158.in[2], compute_graph.l2_158.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1468), layer(173): compute_graph.flexml_layers[140].compute_node[0][0], compute_graph.flexml_layers[140].compute_node[0][1], compute_graph.flexml_layers[140].compute_node[0][2], compute_graph.flexml_layers[140].compute_node[0][3], compute_graph.flexml_layers[140].compute_node[1][0], compute_graph.flexml_layers[140].compute_node[1][1], compute_graph.flexml_layers[140].compute_node[1][2], compute_graph.flexml_layers[140].compute_node[1][3], compute_graph.flexml_layers[140].compute_node[2][0], compute_graph.flexml_layers[140].compute_node[2][1], compute_graph.flexml_layers[140].compute_node[2][2], compute_graph.flexml_layers[140].compute_node[2][3], compute_graph.flexml_layers[140].compute_node[3][0], compute_graph.flexml_layers[140].compute_node[3][1], compute_graph.flexml_layers[140].compute_node[3][2], compute_graph.flexml_layers[140].compute_node[3][3], {compute_graph.l2_158.out[0], compute_graph.l2_158.out[1], compute_graph.l2_158.out[2], compute_graph.l2_158.out[3], compute_graph.l2_159.in[0], compute_graph.l2_159.in[1], compute_graph.l2_159.in[2], compute_graph.l2_159.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1469), layer(174): compute_graph.flexml_layers[141].compute_node[0][0], compute_graph.flexml_layers[141].compute_node[0][1], compute_graph.flexml_layers[141].compute_node[0][2], compute_graph.flexml_layers[141].compute_node[0][3], compute_graph.flexml_layers[141].compute_node[1][0], compute_graph.flexml_layers[141].compute_node[1][1], compute_graph.flexml_layers[141].compute_node[1][2], compute_graph.flexml_layers[141].compute_node[1][3], compute_graph.flexml_layers[141].compute_node[2][0], compute_graph.flexml_layers[141].compute_node[2][1], compute_graph.flexml_layers[141].compute_node[2][2], compute_graph.flexml_layers[141].compute_node[2][3], compute_graph.flexml_layers[141].compute_node[3][0], compute_graph.flexml_layers[141].compute_node[3][1], compute_graph.flexml_layers[141].compute_node[3][2], compute_graph.flexml_layers[141].compute_node[3][3], {compute_graph.l2_159.out[0], compute_graph.l2_159.out[1], compute_graph.l2_159.out[2], compute_graph.l2_159.out[3], compute_graph.l2_160.in[0], compute_graph.l2_160.in[1], compute_graph.l2_160.in[2], compute_graph.l2_160.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(938): mode(0), layer(174): {compute_graph.l2_160.out[1], compute_graph.l2l3_160_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1091): mode(0), layer(175): {compute_graph.l2l3_160_spill.out[0], compute_graph.templated_graph_161.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1270): mode(0), layer(175): {compute_graph.templated_graph_161.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_161_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1093): mode(0), layer(176): {compute_graph.l2l3_scratch_0_161_spill.out[0], compute_graph.templated_graph_161.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1271): mode(0), layer(176): {compute_graph.templated_graph_161.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_161_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1092): mode(0), layer(177): {compute_graph.l2l3_scratch_1_161_spill.out[0], compute_graph.templated_graph_161.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1272): mode(0), layer(177): {compute_graph.templated_graph_161.g2.ifm_mem.out[0], compute_graph.l2l3_161_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1088): mode(0), layer(178): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1094): mode(0), layer(178): {compute_graph.l2l3_161_spill.out[0], compute_graph.l2l3_161_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1470), layer(178): compute_graph.flexml_layers[142].compute_node[0][0], compute_graph.flexml_layers[142].compute_node[0][1], compute_graph.flexml_layers[142].compute_node[0][2], compute_graph.flexml_layers[142].compute_node[0][3], compute_graph.flexml_layers[142].compute_node[1][0], compute_graph.flexml_layers[142].compute_node[1][1], compute_graph.flexml_layers[142].compute_node[1][2], compute_graph.flexml_layers[142].compute_node[1][3], compute_graph.flexml_layers[142].compute_node[2][0], compute_graph.flexml_layers[142].compute_node[2][1], compute_graph.flexml_layers[142].compute_node[2][2], compute_graph.flexml_layers[142].compute_node[2][3], compute_graph.flexml_layers[142].compute_node[3][0], compute_graph.flexml_layers[142].compute_node[3][1], compute_graph.flexml_layers[142].compute_node[3][2], compute_graph.flexml_layers[142].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[3], compute_graph.l2_162.in[0], compute_graph.l2_162.in[1], compute_graph.l2_162.in[2], compute_graph.l2_162.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(869): mode(3), layer(179): {compute_graph.Layer_163_wts_ddr.out[0], compute_graph.Layer_163_wts_ddr.out[1], compute_graph.Layer_163_l2_wts.in[0], compute_graph.Layer_163_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(869) is pipelined with KernelLayerNode(1470) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1471), layer(179): compute_graph.flexml_layers[143].compute_node[0][0], compute_graph.flexml_layers[143].compute_node[0][1], compute_graph.flexml_layers[143].compute_node[0][2], compute_graph.flexml_layers[143].compute_node[0][3], compute_graph.flexml_layers[143].compute_node[1][0], compute_graph.flexml_layers[143].compute_node[1][1], compute_graph.flexml_layers[143].compute_node[1][2], compute_graph.flexml_layers[143].compute_node[1][3], compute_graph.flexml_layers[143].compute_node[2][0], compute_graph.flexml_layers[143].compute_node[2][1], compute_graph.flexml_layers[143].compute_node[2][2], compute_graph.flexml_layers[143].compute_node[2][3], compute_graph.flexml_layers[143].compute_node[3][0], compute_graph.flexml_layers[143].compute_node[3][1], compute_graph.flexml_layers[143].compute_node[3][2], compute_graph.flexml_layers[143].compute_node[3][3], {compute_graph.Layer_163_l2_wts.out[0], compute_graph.Layer_163_l2_wts.out[1], compute_graph.Layer_163_l2_wts.out[2], compute_graph.Layer_163_l2_wts.out[3], compute_graph.l2_162.out[0], compute_graph.l2_162.out[1], compute_graph.l2_162.out[2], compute_graph.l2_162.out[3], compute_graph.l2_163.in[0], compute_graph.l2_163.in[1], compute_graph.l2_163.in[2], compute_graph.l2_163.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(870): mode(0), layer(180): {compute_graph.Layer_164_wts_ddr.out[0], compute_graph.Layer_164_wts_ddr.out[1], compute_graph.Layer_164_l2_wts.in[0], compute_graph.Layer_164_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.Layer_164_l2_wts.out[0], compute_graph.Layer_164_l2_wts.out[1], compute_graph.Layer_164_l2_wts.out[2], compute_graph.Layer_164_l2_wts.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1472), layer(180): compute_graph.flexml_layers[144].compute_node[0][0], compute_graph.flexml_layers[144].compute_node[0][1], compute_graph.flexml_layers[144].compute_node[0][2], compute_graph.flexml_layers[144].compute_node[0][3], compute_graph.flexml_layers[144].compute_node[1][0], compute_graph.flexml_layers[144].compute_node[1][1], compute_graph.flexml_layers[144].compute_node[1][2], compute_graph.flexml_layers[144].compute_node[1][3], compute_graph.flexml_layers[144].compute_node[2][0], compute_graph.flexml_layers[144].compute_node[2][1], compute_graph.flexml_layers[144].compute_node[2][2], compute_graph.flexml_layers[144].compute_node[2][3], compute_graph.flexml_layers[144].compute_node[3][0], compute_graph.flexml_layers[144].compute_node[3][1], compute_graph.flexml_layers[144].compute_node[3][2], compute_graph.flexml_layers[144].compute_node[3][3], {compute_graph.Layer_164_l2_wts.out[0], compute_graph.Layer_164_l2_wts.out[1], compute_graph.Layer_164_l2_wts.out[2], compute_graph.Layer_164_l2_wts.out[3], compute_graph.l2_163.out[0], compute_graph.l2_163.out[1], compute_graph.l2_163.out[2], compute_graph.l2_163.out[3], compute_graph.l2_164.in[0], compute_graph.l2_164.in[1], compute_graph.l2_164.in[2], compute_graph.l2_164.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(939): mode(3), layer(180): {compute_graph.l2_163.out[4], compute_graph.l2_163.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(939) is pipelined with KernelLayerNode(1472) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1473), layer(181): compute_graph.flexml_layers[145].compute_node[0][0], compute_graph.flexml_layers[145].compute_node[0][1], compute_graph.flexml_layers[145].compute_node[0][2], compute_graph.flexml_layers[145].compute_node[0][3], compute_graph.flexml_layers[145].compute_node[1][0], compute_graph.flexml_layers[145].compute_node[1][1], compute_graph.flexml_layers[145].compute_node[1][2], compute_graph.flexml_layers[145].compute_node[1][3], compute_graph.flexml_layers[145].compute_node[2][0], compute_graph.flexml_layers[145].compute_node[2][1], compute_graph.flexml_layers[145].compute_node[2][2], compute_graph.flexml_layers[145].compute_node[2][3], compute_graph.flexml_layers[145].compute_node[3][0], compute_graph.flexml_layers[145].compute_node[3][1], compute_graph.flexml_layers[145].compute_node[3][2], compute_graph.flexml_layers[145].compute_node[3][3], {compute_graph.l2_164.out[0], compute_graph.l2_164.out[1], compute_graph.l2_164.out[2], compute_graph.l2_164.out[3], compute_graph.l2_165.in[0], compute_graph.l2_165.in[1], compute_graph.l2_165.in[2], compute_graph.l2_165.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(940): mode(3), layer(181): {compute_graph.l2_164.out[4], compute_graph.l2_164.out[5], compute_graph.l2l3_164_spill.in[0], compute_graph.l2l3_164_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(940) is pipelined with KernelLayerNode(1473) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1474), layer(182): compute_graph.flexml_layers[146].compute_node[0][0], compute_graph.flexml_layers[146].compute_node[0][1], compute_graph.flexml_layers[146].compute_node[0][2], compute_graph.flexml_layers[146].compute_node[0][3], compute_graph.flexml_layers[146].compute_node[1][0], compute_graph.flexml_layers[146].compute_node[1][1], compute_graph.flexml_layers[146].compute_node[1][2], compute_graph.flexml_layers[146].compute_node[1][3], compute_graph.flexml_layers[146].compute_node[2][0], compute_graph.flexml_layers[146].compute_node[2][1], compute_graph.flexml_layers[146].compute_node[2][2], compute_graph.flexml_layers[146].compute_node[2][3], compute_graph.flexml_layers[146].compute_node[3][0], compute_graph.flexml_layers[146].compute_node[3][1], compute_graph.flexml_layers[146].compute_node[3][2], compute_graph.flexml_layers[146].compute_node[3][3], {compute_graph.l2_165.out[0], compute_graph.l2_165.out[1], compute_graph.l2_165.out[2], compute_graph.l2_165.out[3], compute_graph.l2_166.in[0], compute_graph.l2_166.in[1], compute_graph.l2_166.in[2], compute_graph.l2_166.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1475), layer(183): compute_graph.flexml_layers[147].compute_node[0][0], compute_graph.flexml_layers[147].compute_node[0][1], compute_graph.flexml_layers[147].compute_node[0][2], compute_graph.flexml_layers[147].compute_node[0][3], compute_graph.flexml_layers[147].compute_node[1][0], compute_graph.flexml_layers[147].compute_node[1][1], compute_graph.flexml_layers[147].compute_node[1][2], compute_graph.flexml_layers[147].compute_node[1][3], compute_graph.flexml_layers[147].compute_node[2][0], compute_graph.flexml_layers[147].compute_node[2][1], compute_graph.flexml_layers[147].compute_node[2][2], compute_graph.flexml_layers[147].compute_node[2][3], compute_graph.flexml_layers[147].compute_node[3][0], compute_graph.flexml_layers[147].compute_node[3][1], compute_graph.flexml_layers[147].compute_node[3][2], compute_graph.flexml_layers[147].compute_node[3][3], {compute_graph.l2_166.out[0], compute_graph.l2_166.out[1], compute_graph.l2_166.out[2], compute_graph.l2_166.out[3], compute_graph.l2_167.in[0], compute_graph.l2_167.in[1], compute_graph.l2_167.in[2], compute_graph.l2_167.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(941): mode(0), layer(183): {compute_graph.l2_167.out[0], compute_graph.l2_167.out[1], compute_graph.l2l3_167_spill.in[0], compute_graph.l2l3_167_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 942 will not be pipelined because it's in the same layer 184 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1589), layer(184): {compute_graph.l2l3_164_spill.out[0], compute_graph.l2l3_167_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1590), layer(184): {compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1591), layer(184): {compute_graph.l2l3_164_spill.out[1], compute_graph.l2l3_167_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1592), layer(184): {compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1476), layer(184): compute_graph.flexml_layers[148].compute_node[0][0], compute_graph.flexml_layers[148].compute_node[0][1], compute_graph.flexml_layers[148].compute_node[0][2], compute_graph.flexml_layers[148].compute_node[0][3], compute_graph.flexml_layers[148].compute_node[1][0], compute_graph.flexml_layers[148].compute_node[1][1], compute_graph.flexml_layers[148].compute_node[1][2], compute_graph.flexml_layers[148].compute_node[1][3], compute_graph.flexml_layers[148].compute_node[2][0], compute_graph.flexml_layers[148].compute_node[2][1], compute_graph.flexml_layers[148].compute_node[2][2], compute_graph.flexml_layers[148].compute_node[2][3], compute_graph.flexml_layers[148].compute_node[3][0], compute_graph.flexml_layers[148].compute_node[3][1], compute_graph.flexml_layers[148].compute_node[3][2], compute_graph.flexml_layers[148].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[3], compute_graph.l2_168.in[0], compute_graph.l2_168.in[1], compute_graph.l2_168.in[2], compute_graph.l2_168.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(871): mode(3), layer(185): {compute_graph.Layer_169_wts_ddr.out[0], compute_graph.Layer_169_wts_ddr.out[1], compute_graph.Layer_169_l2_wts.in[0], compute_graph.Layer_169_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(871) is pipelined with KernelLayerNode(1476) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(942): mode(3), layer(184): {compute_graph.l2_168.out[0], compute_graph.l2_168.out[1], compute_graph.l2l3_168_spill.in[0], compute_graph.l2l3_168_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1098): mode(0), layer(185): {compute_graph.l2l3_168_spill.out[0], compute_graph.l2l3_168_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1477), layer(185): compute_graph.flexml_layers[149].compute_node[0][0], compute_graph.flexml_layers[149].compute_node[0][1], compute_graph.flexml_layers[149].compute_node[0][2], compute_graph.flexml_layers[149].compute_node[0][3], compute_graph.flexml_layers[149].compute_node[1][0], compute_graph.flexml_layers[149].compute_node[1][1], compute_graph.flexml_layers[149].compute_node[1][2], compute_graph.flexml_layers[149].compute_node[1][3], compute_graph.flexml_layers[149].compute_node[2][0], compute_graph.flexml_layers[149].compute_node[2][1], compute_graph.flexml_layers[149].compute_node[2][2], compute_graph.flexml_layers[149].compute_node[2][3], compute_graph.flexml_layers[149].compute_node[3][0], compute_graph.flexml_layers[149].compute_node[3][1], compute_graph.flexml_layers[149].compute_node[3][2], compute_graph.flexml_layers[149].compute_node[3][3], {compute_graph.Layer_169_l2_wts.out[0], compute_graph.Layer_169_l2_wts.out[1], compute_graph.Layer_169_l2_wts.out[2], compute_graph.Layer_169_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[3], compute_graph.l2_169.in[0], compute_graph.l2_169.in[1], compute_graph.l2_169.in[2], compute_graph.l2_169.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1478), layer(186): compute_graph.flexml_layers[150].compute_node[0][0], compute_graph.flexml_layers[150].compute_node[0][1], compute_graph.flexml_layers[150].compute_node[0][2], compute_graph.flexml_layers[150].compute_node[0][3], compute_graph.flexml_layers[150].compute_node[1][0], compute_graph.flexml_layers[150].compute_node[1][1], compute_graph.flexml_layers[150].compute_node[1][2], compute_graph.flexml_layers[150].compute_node[1][3], compute_graph.flexml_layers[150].compute_node[2][0], compute_graph.flexml_layers[150].compute_node[2][1], compute_graph.flexml_layers[150].compute_node[2][2], compute_graph.flexml_layers[150].compute_node[2][3], compute_graph.flexml_layers[150].compute_node[3][0], compute_graph.flexml_layers[150].compute_node[3][1], compute_graph.flexml_layers[150].compute_node[3][2], compute_graph.flexml_layers[150].compute_node[3][3], {compute_graph.l2_169.out[0], compute_graph.l2_169.out[1], compute_graph.l2_169.out[2], compute_graph.l2_169.out[3], compute_graph.l2_170.in[0], compute_graph.l2_170.in[1], compute_graph.l2_170.in[2], compute_graph.l2_170.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(943): mode(3), layer(186): {compute_graph.l2_169.out[4], compute_graph.l2_169.out[5], compute_graph.l2l3_169_spill.in[0], compute_graph.l2l3_169_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(943) is pipelined with KernelLayerNode(1478) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1479), layer(187): compute_graph.flexml_layers[151].compute_node[0][0], compute_graph.flexml_layers[151].compute_node[0][1], compute_graph.flexml_layers[151].compute_node[0][2], compute_graph.flexml_layers[151].compute_node[0][3], compute_graph.flexml_layers[151].compute_node[1][0], compute_graph.flexml_layers[151].compute_node[1][1], compute_graph.flexml_layers[151].compute_node[1][2], compute_graph.flexml_layers[151].compute_node[1][3], compute_graph.flexml_layers[151].compute_node[2][0], compute_graph.flexml_layers[151].compute_node[2][1], compute_graph.flexml_layers[151].compute_node[2][2], compute_graph.flexml_layers[151].compute_node[2][3], compute_graph.flexml_layers[151].compute_node[3][0], compute_graph.flexml_layers[151].compute_node[3][1], compute_graph.flexml_layers[151].compute_node[3][2], compute_graph.flexml_layers[151].compute_node[3][3], {compute_graph.l2_170.out[0], compute_graph.l2_170.out[1], compute_graph.l2_170.out[2], compute_graph.l2_170.out[3], compute_graph.l2_171.in[0], compute_graph.l2_171.in[1], compute_graph.l2_171.in[2], compute_graph.l2_171.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1480), layer(188): compute_graph.flexml_layers[152].compute_node[0][0], compute_graph.flexml_layers[152].compute_node[0][1], compute_graph.flexml_layers[152].compute_node[0][2], compute_graph.flexml_layers[152].compute_node[0][3], compute_graph.flexml_layers[152].compute_node[1][0], compute_graph.flexml_layers[152].compute_node[1][1], compute_graph.flexml_layers[152].compute_node[1][2], compute_graph.flexml_layers[152].compute_node[1][3], compute_graph.flexml_layers[152].compute_node[2][0], compute_graph.flexml_layers[152].compute_node[2][1], compute_graph.flexml_layers[152].compute_node[2][2], compute_graph.flexml_layers[152].compute_node[2][3], compute_graph.flexml_layers[152].compute_node[3][0], compute_graph.flexml_layers[152].compute_node[3][1], compute_graph.flexml_layers[152].compute_node[3][2], compute_graph.flexml_layers[152].compute_node[3][3], {compute_graph.l2_171.out[0], compute_graph.l2_171.out[1], compute_graph.l2_171.out[2], compute_graph.l2_171.out[3], compute_graph.l2_172.in[0], compute_graph.l2_172.in[1], compute_graph.l2_172.in[2], compute_graph.l2_172.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(944): mode(0), layer(188): {compute_graph.l2_172.out[0], compute_graph.l2_172.out[1], compute_graph.l2l3_172_spill.in[0], compute_graph.l2l3_172_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 945 will not be pipelined because it's in the same layer 189 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1593), layer(189): {compute_graph.l2l3_169_spill.out[0], compute_graph.l2l3_172_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1594), layer(189): {compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1595), layer(189): {compute_graph.l2l3_169_spill.out[1], compute_graph.l2l3_172_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1596), layer(189): {compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1481), layer(189): compute_graph.flexml_layers[153].compute_node[0][0], compute_graph.flexml_layers[153].compute_node[0][1], compute_graph.flexml_layers[153].compute_node[0][2], compute_graph.flexml_layers[153].compute_node[0][3], compute_graph.flexml_layers[153].compute_node[1][0], compute_graph.flexml_layers[153].compute_node[1][1], compute_graph.flexml_layers[153].compute_node[1][2], compute_graph.flexml_layers[153].compute_node[1][3], compute_graph.flexml_layers[153].compute_node[2][0], compute_graph.flexml_layers[153].compute_node[2][1], compute_graph.flexml_layers[153].compute_node[2][2], compute_graph.flexml_layers[153].compute_node[2][3], compute_graph.flexml_layers[153].compute_node[3][0], compute_graph.flexml_layers[153].compute_node[3][1], compute_graph.flexml_layers[153].compute_node[3][2], compute_graph.flexml_layers[153].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[3], compute_graph.l2_173.in[0], compute_graph.l2_173.in[1], compute_graph.l2_173.in[2], compute_graph.l2_173.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(945): mode(3), layer(189): {compute_graph.l2_173.out[0], compute_graph.l2_173.out[1], compute_graph.l2l3_173_spill.in[0], compute_graph.l2l3_173_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1101): mode(0), layer(190): {compute_graph.l2l3_173_spill.out[0], compute_graph.templated_graph_174.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1273): mode(0), layer(190): {compute_graph.templated_graph_174.trans_mt_ifm.out[0], compute_graph.templated_graph_174.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1274): mode(0), layer(190): {compute_graph.templated_graph_174.trans_mt_ofm.out[0], compute_graph.l2l3_174_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1103): mode(0), layer(191): {compute_graph.l2l3_174_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1482), layer(191): compute_graph.flexml_layers[154].compute_node[0][0], compute_graph.flexml_layers[154].compute_node[0][1], compute_graph.flexml_layers[154].compute_node[0][2], compute_graph.flexml_layers[154].compute_node[0][3], compute_graph.flexml_layers[154].compute_node[1][0], compute_graph.flexml_layers[154].compute_node[1][1], compute_graph.flexml_layers[154].compute_node[1][2], compute_graph.flexml_layers[154].compute_node[1][3], compute_graph.flexml_layers[154].compute_node[2][0], compute_graph.flexml_layers[154].compute_node[2][1], compute_graph.flexml_layers[154].compute_node[2][2], compute_graph.flexml_layers[154].compute_node[2][3], compute_graph.flexml_layers[154].compute_node[3][0], compute_graph.flexml_layers[154].compute_node[3][1], compute_graph.flexml_layers[154].compute_node[3][2], compute_graph.flexml_layers[154].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[3], compute_graph.l2_175.in[0], compute_graph.l2_175.in[1], compute_graph.l2_175.in[2], compute_graph.l2_175.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(872): mode(3), layer(192): {compute_graph.Layer_176_wts_ddr.out[0], compute_graph.Layer_176_wts_ddr.out[1], compute_graph.Layer_176_l2_wts.in[0], compute_graph.Layer_176_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(872) is pipelined with KernelLayerNode(1482) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1483), layer(192): compute_graph.flexml_layers[155].compute_node[0][0], compute_graph.flexml_layers[155].compute_node[0][1], compute_graph.flexml_layers[155].compute_node[0][2], compute_graph.flexml_layers[155].compute_node[0][3], compute_graph.flexml_layers[155].compute_node[1][0], compute_graph.flexml_layers[155].compute_node[1][1], compute_graph.flexml_layers[155].compute_node[1][2], compute_graph.flexml_layers[155].compute_node[1][3], compute_graph.flexml_layers[155].compute_node[2][0], compute_graph.flexml_layers[155].compute_node[2][1], compute_graph.flexml_layers[155].compute_node[2][2], compute_graph.flexml_layers[155].compute_node[2][3], compute_graph.flexml_layers[155].compute_node[3][0], compute_graph.flexml_layers[155].compute_node[3][1], compute_graph.flexml_layers[155].compute_node[3][2], compute_graph.flexml_layers[155].compute_node[3][3], {compute_graph.Layer_176_l2_wts.out[0], compute_graph.Layer_176_l2_wts.out[1], compute_graph.Layer_176_l2_wts.out[2], compute_graph.Layer_176_l2_wts.out[3], compute_graph.l2_175.out[0], compute_graph.l2_175.out[1], compute_graph.l2_175.out[2], compute_graph.l2_175.out[3], compute_graph.l2_176.in[0], compute_graph.l2_176.in[1], compute_graph.l2_176.in[2], compute_graph.l2_176.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(873): mode(0), layer(193): {compute_graph.Layer_177_wts_ddr.out[0], compute_graph.Layer_177_wts_ddr.out[1], compute_graph.Layer_177_l2_wts.in[0], compute_graph.Layer_177_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1484), layer(193): compute_graph.flexml_layers[156].compute_node[0][0], compute_graph.flexml_layers[156].compute_node[0][1], compute_graph.flexml_layers[156].compute_node[0][2], compute_graph.flexml_layers[156].compute_node[0][3], compute_graph.flexml_layers[156].compute_node[1][0], compute_graph.flexml_layers[156].compute_node[1][1], compute_graph.flexml_layers[156].compute_node[1][2], compute_graph.flexml_layers[156].compute_node[1][3], compute_graph.flexml_layers[156].compute_node[2][0], compute_graph.flexml_layers[156].compute_node[2][1], compute_graph.flexml_layers[156].compute_node[2][2], compute_graph.flexml_layers[156].compute_node[2][3], compute_graph.flexml_layers[156].compute_node[3][0], compute_graph.flexml_layers[156].compute_node[3][1], compute_graph.flexml_layers[156].compute_node[3][2], compute_graph.flexml_layers[156].compute_node[3][3], {compute_graph.Layer_177_l2_wts.out[0], compute_graph.Layer_177_l2_wts.out[1], compute_graph.Layer_177_l2_wts.out[2], compute_graph.Layer_177_l2_wts.out[3], compute_graph.l2_176.out[0], compute_graph.l2_176.out[1], compute_graph.l2_176.out[2], compute_graph.l2_176.out[3], compute_graph.l2_177.in[0], compute_graph.l2_177.in[1], compute_graph.l2_177.in[2], compute_graph.l2_177.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1485), layer(194): compute_graph.flexml_layers[157].compute_node[0][0], compute_graph.flexml_layers[157].compute_node[0][1], compute_graph.flexml_layers[157].compute_node[0][2], compute_graph.flexml_layers[157].compute_node[0][3], compute_graph.flexml_layers[157].compute_node[1][0], compute_graph.flexml_layers[157].compute_node[1][1], compute_graph.flexml_layers[157].compute_node[1][2], compute_graph.flexml_layers[157].compute_node[1][3], compute_graph.flexml_layers[157].compute_node[2][0], compute_graph.flexml_layers[157].compute_node[2][1], compute_graph.flexml_layers[157].compute_node[2][2], compute_graph.flexml_layers[157].compute_node[2][3], compute_graph.flexml_layers[157].compute_node[3][0], compute_graph.flexml_layers[157].compute_node[3][1], compute_graph.flexml_layers[157].compute_node[3][2], compute_graph.flexml_layers[157].compute_node[3][3], {compute_graph.l2_177.out[0], compute_graph.l2_177.out[1], compute_graph.l2_177.out[2], compute_graph.l2_177.out[3], compute_graph.l2_178.in[0], compute_graph.l2_178.in[1], compute_graph.l2_178.in[2], compute_graph.l2_178.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1486), layer(195): compute_graph.flexml_layers[158].compute_node[0][0], compute_graph.flexml_layers[158].compute_node[0][1], compute_graph.flexml_layers[158].compute_node[0][2], compute_graph.flexml_layers[158].compute_node[0][3], compute_graph.flexml_layers[158].compute_node[1][0], compute_graph.flexml_layers[158].compute_node[1][1], compute_graph.flexml_layers[158].compute_node[1][2], compute_graph.flexml_layers[158].compute_node[1][3], compute_graph.flexml_layers[158].compute_node[2][0], compute_graph.flexml_layers[158].compute_node[2][1], compute_graph.flexml_layers[158].compute_node[2][2], compute_graph.flexml_layers[158].compute_node[2][3], compute_graph.flexml_layers[158].compute_node[3][0], compute_graph.flexml_layers[158].compute_node[3][1], compute_graph.flexml_layers[158].compute_node[3][2], compute_graph.flexml_layers[158].compute_node[3][3], {compute_graph.l2_178.out[0], compute_graph.l2_178.out[1], compute_graph.l2_178.out[2], compute_graph.l2_178.out[3], compute_graph.l2_179.in[0], compute_graph.l2_179.in[1], compute_graph.l2_179.in[2], compute_graph.l2_179.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1487), layer(196): compute_graph.flexml_layers[159].compute_node[0][0], compute_graph.flexml_layers[159].compute_node[0][1], compute_graph.flexml_layers[159].compute_node[0][2], compute_graph.flexml_layers[159].compute_node[0][3], compute_graph.flexml_layers[159].compute_node[1][0], compute_graph.flexml_layers[159].compute_node[1][1], compute_graph.flexml_layers[159].compute_node[1][2], compute_graph.flexml_layers[159].compute_node[1][3], compute_graph.flexml_layers[159].compute_node[2][0], compute_graph.flexml_layers[159].compute_node[2][1], compute_graph.flexml_layers[159].compute_node[2][2], compute_graph.flexml_layers[159].compute_node[2][3], compute_graph.flexml_layers[159].compute_node[3][0], compute_graph.flexml_layers[159].compute_node[3][1], compute_graph.flexml_layers[159].compute_node[3][2], compute_graph.flexml_layers[159].compute_node[3][3], {compute_graph.l2_179.out[0], compute_graph.l2_179.out[1], compute_graph.l2_179.out[2], compute_graph.l2_179.out[3], compute_graph.l2_180.in[0], compute_graph.l2_180.in[1], compute_graph.l2_180.in[2], compute_graph.l2_180.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(946): mode(0), layer(196): {compute_graph.l2_180.out[1], compute_graph.l2l3_180_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1104): mode(0), layer(197): {compute_graph.l2l3_180_spill.out[0], compute_graph.templated_graph_181.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1275): mode(0), layer(197): {compute_graph.templated_graph_181.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_181_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1105): mode(0), layer(198): {compute_graph.l2l3_scratch_0_181_spill.out[0], compute_graph.templated_graph_181.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1276): mode(0), layer(198): {compute_graph.templated_graph_181.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_181_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1106): mode(0), layer(199): {compute_graph.l2l3_scratch_1_181_spill.out[0], compute_graph.templated_graph_181.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1277): mode(0), layer(199): {compute_graph.templated_graph_181.g2.ifm_mem.out[0], compute_graph.l2l3_181_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1102): mode(0), layer(200): {compute_graph.l2l3_173_spill.out[1], compute_graph.l2l3_173_spill.out[2], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1107): mode(0), layer(200): {compute_graph.l2l3_181_spill.out[0], compute_graph.l2l3_181_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1488), layer(200): compute_graph.flexml_layers[160].compute_node[0][0], compute_graph.flexml_layers[160].compute_node[0][1], compute_graph.flexml_layers[160].compute_node[0][2], compute_graph.flexml_layers[160].compute_node[0][3], compute_graph.flexml_layers[160].compute_node[1][0], compute_graph.flexml_layers[160].compute_node[1][1], compute_graph.flexml_layers[160].compute_node[1][2], compute_graph.flexml_layers[160].compute_node[1][3], compute_graph.flexml_layers[160].compute_node[2][0], compute_graph.flexml_layers[160].compute_node[2][1], compute_graph.flexml_layers[160].compute_node[2][2], compute_graph.flexml_layers[160].compute_node[2][3], compute_graph.flexml_layers[160].compute_node[3][0], compute_graph.flexml_layers[160].compute_node[3][1], compute_graph.flexml_layers[160].compute_node[3][2], compute_graph.flexml_layers[160].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[3], compute_graph.l2_182.in[0], compute_graph.l2_182.in[1], compute_graph.l2_182.in[2], compute_graph.l2_182.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(874): mode(3), layer(201): {compute_graph.Layer_183_wts_ddr.out[0], compute_graph.Layer_183_wts_ddr.out[1], compute_graph.Layer_183_l2_wts.in[0], compute_graph.Layer_183_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(874) is pipelined with KernelLayerNode(1488) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163 overlaps with memory space used by compute_graph.l2_183. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1095): mode(0), layer(201): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[3], compute_graph.l2_182.out[0], compute_graph.l2_182.out[1], compute_graph.l2_182.out[2], compute_graph.l2_182.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1489), layer(201): compute_graph.flexml_layers[161].compute_node[0][0], compute_graph.flexml_layers[161].compute_node[0][1], compute_graph.flexml_layers[161].compute_node[0][2], compute_graph.flexml_layers[161].compute_node[0][3], compute_graph.flexml_layers[161].compute_node[1][0], compute_graph.flexml_layers[161].compute_node[1][1], compute_graph.flexml_layers[161].compute_node[1][2], compute_graph.flexml_layers[161].compute_node[1][3], compute_graph.flexml_layers[161].compute_node[2][0], compute_graph.flexml_layers[161].compute_node[2][1], compute_graph.flexml_layers[161].compute_node[2][2], compute_graph.flexml_layers[161].compute_node[2][3], compute_graph.flexml_layers[161].compute_node[3][0], compute_graph.flexml_layers[161].compute_node[3][1], compute_graph.flexml_layers[161].compute_node[3][2], compute_graph.flexml_layers[161].compute_node[3][3], {compute_graph.Layer_183_l2_wts.out[0], compute_graph.Layer_183_l2_wts.out[1], compute_graph.Layer_183_l2_wts.out[2], compute_graph.Layer_183_l2_wts.out[3], compute_graph.l2_182.out[0], compute_graph.l2_182.out[1], compute_graph.l2_182.out[2], compute_graph.l2_182.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[3], compute_graph.l2_183.in[0], compute_graph.l2_183.in[1], compute_graph.l2_183.in[2], compute_graph.l2_183.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(875): mode(0), layer(202): {compute_graph.Layer_184_wts_ddr.out[0], compute_graph.Layer_184_wts_ddr.out[1], compute_graph.Layer_184_l2_wts.in[0], compute_graph.Layer_184_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.Layer_184_l2_wts.out[0], compute_graph.Layer_184_l2_wts.out[1], compute_graph.Layer_184_l2_wts.out[2], compute_graph.Layer_184_l2_wts.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1490), layer(202): compute_graph.flexml_layers[162].compute_node[0][0], compute_graph.flexml_layers[162].compute_node[0][1], compute_graph.flexml_layers[162].compute_node[0][2], compute_graph.flexml_layers[162].compute_node[0][3], compute_graph.flexml_layers[162].compute_node[1][0], compute_graph.flexml_layers[162].compute_node[1][1], compute_graph.flexml_layers[162].compute_node[1][2], compute_graph.flexml_layers[162].compute_node[1][3], compute_graph.flexml_layers[162].compute_node[2][0], compute_graph.flexml_layers[162].compute_node[2][1], compute_graph.flexml_layers[162].compute_node[2][2], compute_graph.flexml_layers[162].compute_node[2][3], compute_graph.flexml_layers[162].compute_node[3][0], compute_graph.flexml_layers[162].compute_node[3][1], compute_graph.flexml_layers[162].compute_node[3][2], compute_graph.flexml_layers[162].compute_node[3][3], {compute_graph.Layer_184_l2_wts.out[0], compute_graph.Layer_184_l2_wts.out[1], compute_graph.Layer_184_l2_wts.out[2], compute_graph.Layer_184_l2_wts.out[3], compute_graph.l2_183.out[0], compute_graph.l2_183.out[1], compute_graph.l2_183.out[2], compute_graph.l2_183.out[3], compute_graph.l2_184.in[0], compute_graph.l2_184.in[1], compute_graph.l2_184.in[2], compute_graph.l2_184.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(947): mode(3), layer(202): {compute_graph.l2_183.out[4], compute_graph.l2_183.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(947) is pipelined with KernelLayerNode(1490) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1491), layer(203): compute_graph.flexml_layers[163].compute_node[0][0], compute_graph.flexml_layers[163].compute_node[0][1], compute_graph.flexml_layers[163].compute_node[0][2], compute_graph.flexml_layers[163].compute_node[0][3], compute_graph.flexml_layers[163].compute_node[1][0], compute_graph.flexml_layers[163].compute_node[1][1], compute_graph.flexml_layers[163].compute_node[1][2], compute_graph.flexml_layers[163].compute_node[1][3], compute_graph.flexml_layers[163].compute_node[2][0], compute_graph.flexml_layers[163].compute_node[2][1], compute_graph.flexml_layers[163].compute_node[2][2], compute_graph.flexml_layers[163].compute_node[2][3], compute_graph.flexml_layers[163].compute_node[3][0], compute_graph.flexml_layers[163].compute_node[3][1], compute_graph.flexml_layers[163].compute_node[3][2], compute_graph.flexml_layers[163].compute_node[3][3], {compute_graph.l2_184.out[0], compute_graph.l2_184.out[1], compute_graph.l2_184.out[2], compute_graph.l2_184.out[3], compute_graph.l2_185.in[0], compute_graph.l2_185.in[1], compute_graph.l2_185.in[2], compute_graph.l2_185.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(948): mode(3), layer(203): {compute_graph.l2_184.out[4], compute_graph.l2_184.out[5], compute_graph.l2l3_184_spill.in[0], compute_graph.l2l3_184_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(948) is pipelined with KernelLayerNode(1491) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1492), layer(204): compute_graph.flexml_layers[164].compute_node[0][0], compute_graph.flexml_layers[164].compute_node[0][1], compute_graph.flexml_layers[164].compute_node[0][2], compute_graph.flexml_layers[164].compute_node[0][3], compute_graph.flexml_layers[164].compute_node[1][0], compute_graph.flexml_layers[164].compute_node[1][1], compute_graph.flexml_layers[164].compute_node[1][2], compute_graph.flexml_layers[164].compute_node[1][3], compute_graph.flexml_layers[164].compute_node[2][0], compute_graph.flexml_layers[164].compute_node[2][1], compute_graph.flexml_layers[164].compute_node[2][2], compute_graph.flexml_layers[164].compute_node[2][3], compute_graph.flexml_layers[164].compute_node[3][0], compute_graph.flexml_layers[164].compute_node[3][1], compute_graph.flexml_layers[164].compute_node[3][2], compute_graph.flexml_layers[164].compute_node[3][3], {compute_graph.l2_185.out[0], compute_graph.l2_185.out[1], compute_graph.l2_185.out[2], compute_graph.l2_185.out[3], compute_graph.l2_186.in[0], compute_graph.l2_186.in[1], compute_graph.l2_186.in[2], compute_graph.l2_186.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1493), layer(205): compute_graph.flexml_layers[165].compute_node[0][0], compute_graph.flexml_layers[165].compute_node[0][1], compute_graph.flexml_layers[165].compute_node[0][2], compute_graph.flexml_layers[165].compute_node[0][3], compute_graph.flexml_layers[165].compute_node[1][0], compute_graph.flexml_layers[165].compute_node[1][1], compute_graph.flexml_layers[165].compute_node[1][2], compute_graph.flexml_layers[165].compute_node[1][3], compute_graph.flexml_layers[165].compute_node[2][0], compute_graph.flexml_layers[165].compute_node[2][1], compute_graph.flexml_layers[165].compute_node[2][2], compute_graph.flexml_layers[165].compute_node[2][3], compute_graph.flexml_layers[165].compute_node[3][0], compute_graph.flexml_layers[165].compute_node[3][1], compute_graph.flexml_layers[165].compute_node[3][2], compute_graph.flexml_layers[165].compute_node[3][3], {compute_graph.l2_186.out[0], compute_graph.l2_186.out[1], compute_graph.l2_186.out[2], compute_graph.l2_186.out[3], compute_graph.l2_187.in[0], compute_graph.l2_187.in[1], compute_graph.l2_187.in[2], compute_graph.l2_187.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(949): mode(0), layer(205): {compute_graph.l2_187.out[0], compute_graph.l2_187.out[1], compute_graph.l2l3_187_spill.in[0], compute_graph.l2l3_187_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 950 will not be pipelined because it's in the same layer 206 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1597), layer(206): {compute_graph.l2l3_184_spill.out[0], compute_graph.l2l3_187_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1598), layer(206): {compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1599), layer(206): {compute_graph.l2l3_184_spill.out[1], compute_graph.l2l3_187_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1600), layer(206): {compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1494), layer(206): compute_graph.flexml_layers[166].compute_node[0][0], compute_graph.flexml_layers[166].compute_node[0][1], compute_graph.flexml_layers[166].compute_node[0][2], compute_graph.flexml_layers[166].compute_node[0][3], compute_graph.flexml_layers[166].compute_node[1][0], compute_graph.flexml_layers[166].compute_node[1][1], compute_graph.flexml_layers[166].compute_node[1][2], compute_graph.flexml_layers[166].compute_node[1][3], compute_graph.flexml_layers[166].compute_node[2][0], compute_graph.flexml_layers[166].compute_node[2][1], compute_graph.flexml_layers[166].compute_node[2][2], compute_graph.flexml_layers[166].compute_node[2][3], compute_graph.flexml_layers[166].compute_node[3][0], compute_graph.flexml_layers[166].compute_node[3][1], compute_graph.flexml_layers[166].compute_node[3][2], compute_graph.flexml_layers[166].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[3], compute_graph.l2_188.in[0], compute_graph.l2_188.in[1], compute_graph.l2_188.in[2], compute_graph.l2_188.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(876): mode(3), layer(207): {compute_graph.Layer_189_wts_ddr.out[0], compute_graph.Layer_189_wts_ddr.out[1], compute_graph.Layer_189_l2_wts.in[0], compute_graph.Layer_189_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(876) is pipelined with KernelLayerNode(1494) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(950): mode(3), layer(206): {compute_graph.l2_188.out[0], compute_graph.l2_188.out[1], compute_graph.l2l3_188_spill.in[0], compute_graph.l2l3_188_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1111): mode(0), layer(207): {compute_graph.l2l3_188_spill.out[0], compute_graph.l2l3_188_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1495), layer(207): compute_graph.flexml_layers[167].compute_node[0][0], compute_graph.flexml_layers[167].compute_node[0][1], compute_graph.flexml_layers[167].compute_node[0][2], compute_graph.flexml_layers[167].compute_node[0][3], compute_graph.flexml_layers[167].compute_node[1][0], compute_graph.flexml_layers[167].compute_node[1][1], compute_graph.flexml_layers[167].compute_node[1][2], compute_graph.flexml_layers[167].compute_node[1][3], compute_graph.flexml_layers[167].compute_node[2][0], compute_graph.flexml_layers[167].compute_node[2][1], compute_graph.flexml_layers[167].compute_node[2][2], compute_graph.flexml_layers[167].compute_node[2][3], compute_graph.flexml_layers[167].compute_node[3][0], compute_graph.flexml_layers[167].compute_node[3][1], compute_graph.flexml_layers[167].compute_node[3][2], compute_graph.flexml_layers[167].compute_node[3][3], {compute_graph.Layer_189_l2_wts.out[0], compute_graph.Layer_189_l2_wts.out[1], compute_graph.Layer_189_l2_wts.out[2], compute_graph.Layer_189_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[3], compute_graph.l2_189.in[0], compute_graph.l2_189.in[1], compute_graph.l2_189.in[2], compute_graph.l2_189.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1496), layer(208): compute_graph.flexml_layers[168].compute_node[0][0], compute_graph.flexml_layers[168].compute_node[0][1], compute_graph.flexml_layers[168].compute_node[0][2], compute_graph.flexml_layers[168].compute_node[0][3], compute_graph.flexml_layers[168].compute_node[1][0], compute_graph.flexml_layers[168].compute_node[1][1], compute_graph.flexml_layers[168].compute_node[1][2], compute_graph.flexml_layers[168].compute_node[1][3], compute_graph.flexml_layers[168].compute_node[2][0], compute_graph.flexml_layers[168].compute_node[2][1], compute_graph.flexml_layers[168].compute_node[2][2], compute_graph.flexml_layers[168].compute_node[2][3], compute_graph.flexml_layers[168].compute_node[3][0], compute_graph.flexml_layers[168].compute_node[3][1], compute_graph.flexml_layers[168].compute_node[3][2], compute_graph.flexml_layers[168].compute_node[3][3], {compute_graph.l2_189.out[0], compute_graph.l2_189.out[1], compute_graph.l2_189.out[2], compute_graph.l2_189.out[3], compute_graph.l2_190.in[0], compute_graph.l2_190.in[1], compute_graph.l2_190.in[2], compute_graph.l2_190.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(951): mode(3), layer(208): {compute_graph.l2_189.out[4], compute_graph.l2_189.out[5], compute_graph.l2l3_189_spill.in[0], compute_graph.l2l3_189_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(951) is pipelined with KernelLayerNode(1496) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1497), layer(209): compute_graph.flexml_layers[169].compute_node[0][0], compute_graph.flexml_layers[169].compute_node[0][1], compute_graph.flexml_layers[169].compute_node[0][2], compute_graph.flexml_layers[169].compute_node[0][3], compute_graph.flexml_layers[169].compute_node[1][0], compute_graph.flexml_layers[169].compute_node[1][1], compute_graph.flexml_layers[169].compute_node[1][2], compute_graph.flexml_layers[169].compute_node[1][3], compute_graph.flexml_layers[169].compute_node[2][0], compute_graph.flexml_layers[169].compute_node[2][1], compute_graph.flexml_layers[169].compute_node[2][2], compute_graph.flexml_layers[169].compute_node[2][3], compute_graph.flexml_layers[169].compute_node[3][0], compute_graph.flexml_layers[169].compute_node[3][1], compute_graph.flexml_layers[169].compute_node[3][2], compute_graph.flexml_layers[169].compute_node[3][3], {compute_graph.l2_190.out[0], compute_graph.l2_190.out[1], compute_graph.l2_190.out[2], compute_graph.l2_190.out[3], compute_graph.l2_191.in[0], compute_graph.l2_191.in[1], compute_graph.l2_191.in[2], compute_graph.l2_191.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1498), layer(210): compute_graph.flexml_layers[170].compute_node[0][0], compute_graph.flexml_layers[170].compute_node[0][1], compute_graph.flexml_layers[170].compute_node[0][2], compute_graph.flexml_layers[170].compute_node[0][3], compute_graph.flexml_layers[170].compute_node[1][0], compute_graph.flexml_layers[170].compute_node[1][1], compute_graph.flexml_layers[170].compute_node[1][2], compute_graph.flexml_layers[170].compute_node[1][3], compute_graph.flexml_layers[170].compute_node[2][0], compute_graph.flexml_layers[170].compute_node[2][1], compute_graph.flexml_layers[170].compute_node[2][2], compute_graph.flexml_layers[170].compute_node[2][3], compute_graph.flexml_layers[170].compute_node[3][0], compute_graph.flexml_layers[170].compute_node[3][1], compute_graph.flexml_layers[170].compute_node[3][2], compute_graph.flexml_layers[170].compute_node[3][3], {compute_graph.l2_191.out[0], compute_graph.l2_191.out[1], compute_graph.l2_191.out[2], compute_graph.l2_191.out[3], compute_graph.l2_192.in[0], compute_graph.l2_192.in[1], compute_graph.l2_192.in[2], compute_graph.l2_192.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(952): mode(0), layer(210): {compute_graph.l2_192.out[0], compute_graph.l2_192.out[1], compute_graph.l2l3_192_spill.in[0], compute_graph.l2l3_192_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 953 will not be pipelined because it's in the same layer 211 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1601), layer(211): {compute_graph.l2l3_189_spill.out[0], compute_graph.l2l3_192_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1602), layer(211): {compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1603), layer(211): {compute_graph.l2l3_189_spill.out[1], compute_graph.l2l3_192_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1604), layer(211): {compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1499), layer(211): compute_graph.flexml_layers[171].compute_node[0][0], compute_graph.flexml_layers[171].compute_node[0][1], compute_graph.flexml_layers[171].compute_node[0][2], compute_graph.flexml_layers[171].compute_node[0][3], compute_graph.flexml_layers[171].compute_node[1][0], compute_graph.flexml_layers[171].compute_node[1][1], compute_graph.flexml_layers[171].compute_node[1][2], compute_graph.flexml_layers[171].compute_node[1][3], compute_graph.flexml_layers[171].compute_node[2][0], compute_graph.flexml_layers[171].compute_node[2][1], compute_graph.flexml_layers[171].compute_node[2][2], compute_graph.flexml_layers[171].compute_node[2][3], compute_graph.flexml_layers[171].compute_node[3][0], compute_graph.flexml_layers[171].compute_node[3][1], compute_graph.flexml_layers[171].compute_node[3][2], compute_graph.flexml_layers[171].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[3], compute_graph.l2_193.in[0], compute_graph.l2_193.in[1], compute_graph.l2_193.in[2], compute_graph.l2_193.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(953): mode(3), layer(211): {compute_graph.l2_193.out[0], compute_graph.l2_193.out[1], compute_graph.l2l3_193_spill.in[0], compute_graph.l2l3_193_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1114): mode(0), layer(212): {compute_graph.l2l3_193_spill.out[0], compute_graph.templated_graph_194.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1278): mode(0), layer(212): {compute_graph.templated_graph_194.trans_mt_ifm.out[0], compute_graph.templated_graph_194.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1279): mode(0), layer(212): {compute_graph.templated_graph_194.trans_mt_ofm.out[0], compute_graph.l2l3_194_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1116): mode(0), layer(213): {compute_graph.l2l3_194_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1500), layer(213): compute_graph.flexml_layers[172].compute_node[0][0], compute_graph.flexml_layers[172].compute_node[0][1], compute_graph.flexml_layers[172].compute_node[0][2], compute_graph.flexml_layers[172].compute_node[0][3], compute_graph.flexml_layers[172].compute_node[1][0], compute_graph.flexml_layers[172].compute_node[1][1], compute_graph.flexml_layers[172].compute_node[1][2], compute_graph.flexml_layers[172].compute_node[1][3], compute_graph.flexml_layers[172].compute_node[2][0], compute_graph.flexml_layers[172].compute_node[2][1], compute_graph.flexml_layers[172].compute_node[2][2], compute_graph.flexml_layers[172].compute_node[2][3], compute_graph.flexml_layers[172].compute_node[3][0], compute_graph.flexml_layers[172].compute_node[3][1], compute_graph.flexml_layers[172].compute_node[3][2], compute_graph.flexml_layers[172].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[3], compute_graph.l2_195.in[0], compute_graph.l2_195.in[1], compute_graph.l2_195.in[2], compute_graph.l2_195.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(877): mode(3), layer(214): {compute_graph.Layer_196_wts_ddr.out[0], compute_graph.Layer_196_wts_ddr.out[1], compute_graph.Layer_196_l2_wts.in[0], compute_graph.Layer_196_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(877) is pipelined with KernelLayerNode(1500) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1501), layer(214): compute_graph.flexml_layers[173].compute_node[0][0], compute_graph.flexml_layers[173].compute_node[0][1], compute_graph.flexml_layers[173].compute_node[0][2], compute_graph.flexml_layers[173].compute_node[0][3], compute_graph.flexml_layers[173].compute_node[1][0], compute_graph.flexml_layers[173].compute_node[1][1], compute_graph.flexml_layers[173].compute_node[1][2], compute_graph.flexml_layers[173].compute_node[1][3], compute_graph.flexml_layers[173].compute_node[2][0], compute_graph.flexml_layers[173].compute_node[2][1], compute_graph.flexml_layers[173].compute_node[2][2], compute_graph.flexml_layers[173].compute_node[2][3], compute_graph.flexml_layers[173].compute_node[3][0], compute_graph.flexml_layers[173].compute_node[3][1], compute_graph.flexml_layers[173].compute_node[3][2], compute_graph.flexml_layers[173].compute_node[3][3], {compute_graph.Layer_196_l2_wts.out[0], compute_graph.Layer_196_l2_wts.out[1], compute_graph.Layer_196_l2_wts.out[2], compute_graph.Layer_196_l2_wts.out[3], compute_graph.l2_195.out[0], compute_graph.l2_195.out[1], compute_graph.l2_195.out[2], compute_graph.l2_195.out[3], compute_graph.l2_196.in[0], compute_graph.l2_196.in[1], compute_graph.l2_196.in[2], compute_graph.l2_196.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(878): mode(0), layer(215): {compute_graph.Layer_197_wts_ddr.out[0], compute_graph.Layer_197_wts_ddr.out[1], compute_graph.Layer_197_l2_wts.in[0], compute_graph.Layer_197_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1502), layer(215): compute_graph.flexml_layers[174].compute_node[0][0], compute_graph.flexml_layers[174].compute_node[0][1], compute_graph.flexml_layers[174].compute_node[0][2], compute_graph.flexml_layers[174].compute_node[0][3], compute_graph.flexml_layers[174].compute_node[1][0], compute_graph.flexml_layers[174].compute_node[1][1], compute_graph.flexml_layers[174].compute_node[1][2], compute_graph.flexml_layers[174].compute_node[1][3], compute_graph.flexml_layers[174].compute_node[2][0], compute_graph.flexml_layers[174].compute_node[2][1], compute_graph.flexml_layers[174].compute_node[2][2], compute_graph.flexml_layers[174].compute_node[2][3], compute_graph.flexml_layers[174].compute_node[3][0], compute_graph.flexml_layers[174].compute_node[3][1], compute_graph.flexml_layers[174].compute_node[3][2], compute_graph.flexml_layers[174].compute_node[3][3], {compute_graph.Layer_197_l2_wts.out[0], compute_graph.Layer_197_l2_wts.out[1], compute_graph.Layer_197_l2_wts.out[2], compute_graph.Layer_197_l2_wts.out[3], compute_graph.l2_196.out[0], compute_graph.l2_196.out[1], compute_graph.l2_196.out[2], compute_graph.l2_196.out[3], compute_graph.l2_197.in[0], compute_graph.l2_197.in[1], compute_graph.l2_197.in[2], compute_graph.l2_197.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1503), layer(216): compute_graph.flexml_layers[175].compute_node[0][0], compute_graph.flexml_layers[175].compute_node[0][1], compute_graph.flexml_layers[175].compute_node[0][2], compute_graph.flexml_layers[175].compute_node[0][3], compute_graph.flexml_layers[175].compute_node[1][0], compute_graph.flexml_layers[175].compute_node[1][1], compute_graph.flexml_layers[175].compute_node[1][2], compute_graph.flexml_layers[175].compute_node[1][3], compute_graph.flexml_layers[175].compute_node[2][0], compute_graph.flexml_layers[175].compute_node[2][1], compute_graph.flexml_layers[175].compute_node[2][2], compute_graph.flexml_layers[175].compute_node[2][3], compute_graph.flexml_layers[175].compute_node[3][0], compute_graph.flexml_layers[175].compute_node[3][1], compute_graph.flexml_layers[175].compute_node[3][2], compute_graph.flexml_layers[175].compute_node[3][3], {compute_graph.l2_197.out[0], compute_graph.l2_197.out[1], compute_graph.l2_197.out[2], compute_graph.l2_197.out[3], compute_graph.l2_198.in[0], compute_graph.l2_198.in[1], compute_graph.l2_198.in[2], compute_graph.l2_198.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1504), layer(217): compute_graph.flexml_layers[176].compute_node[0][0], compute_graph.flexml_layers[176].compute_node[0][1], compute_graph.flexml_layers[176].compute_node[0][2], compute_graph.flexml_layers[176].compute_node[0][3], compute_graph.flexml_layers[176].compute_node[1][0], compute_graph.flexml_layers[176].compute_node[1][1], compute_graph.flexml_layers[176].compute_node[1][2], compute_graph.flexml_layers[176].compute_node[1][3], compute_graph.flexml_layers[176].compute_node[2][0], compute_graph.flexml_layers[176].compute_node[2][1], compute_graph.flexml_layers[176].compute_node[2][2], compute_graph.flexml_layers[176].compute_node[2][3], compute_graph.flexml_layers[176].compute_node[3][0], compute_graph.flexml_layers[176].compute_node[3][1], compute_graph.flexml_layers[176].compute_node[3][2], compute_graph.flexml_layers[176].compute_node[3][3], {compute_graph.l2_198.out[0], compute_graph.l2_198.out[1], compute_graph.l2_198.out[2], compute_graph.l2_198.out[3], compute_graph.l2_199.in[0], compute_graph.l2_199.in[1], compute_graph.l2_199.in[2], compute_graph.l2_199.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1505), layer(218): compute_graph.flexml_layers[177].compute_node[0][0], compute_graph.flexml_layers[177].compute_node[0][1], compute_graph.flexml_layers[177].compute_node[0][2], compute_graph.flexml_layers[177].compute_node[0][3], compute_graph.flexml_layers[177].compute_node[1][0], compute_graph.flexml_layers[177].compute_node[1][1], compute_graph.flexml_layers[177].compute_node[1][2], compute_graph.flexml_layers[177].compute_node[1][3], compute_graph.flexml_layers[177].compute_node[2][0], compute_graph.flexml_layers[177].compute_node[2][1], compute_graph.flexml_layers[177].compute_node[2][2], compute_graph.flexml_layers[177].compute_node[2][3], compute_graph.flexml_layers[177].compute_node[3][0], compute_graph.flexml_layers[177].compute_node[3][1], compute_graph.flexml_layers[177].compute_node[3][2], compute_graph.flexml_layers[177].compute_node[3][3], {compute_graph.l2_199.out[0], compute_graph.l2_199.out[1], compute_graph.l2_199.out[2], compute_graph.l2_199.out[3], compute_graph.l2_200.in[0], compute_graph.l2_200.in[1], compute_graph.l2_200.in[2], compute_graph.l2_200.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(954): mode(0), layer(218): {compute_graph.l2_200.out[1], compute_graph.l2l3_200_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1117): mode(0), layer(219): {compute_graph.l2l3_200_spill.out[0], compute_graph.templated_graph_201.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1280): mode(0), layer(219): {compute_graph.templated_graph_201.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_201_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1118): mode(0), layer(220): {compute_graph.l2l3_scratch_0_201_spill.out[0], compute_graph.templated_graph_201.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1281): mode(0), layer(220): {compute_graph.templated_graph_201.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_201_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1119): mode(0), layer(221): {compute_graph.l2l3_scratch_1_201_spill.out[0], compute_graph.templated_graph_201.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1282): mode(0), layer(221): {compute_graph.templated_graph_201.g2.ifm_mem.out[0], compute_graph.l2l3_201_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1115): mode(0), layer(222): {compute_graph.l2l3_193_spill.out[1], compute_graph.l2l3_193_spill.out[2], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1120): mode(0), layer(222): {compute_graph.l2l3_201_spill.out[0], compute_graph.l2l3_201_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1506), layer(222): compute_graph.flexml_layers[178].compute_node[0][0], compute_graph.flexml_layers[178].compute_node[0][1], compute_graph.flexml_layers[178].compute_node[0][2], compute_graph.flexml_layers[178].compute_node[0][3], compute_graph.flexml_layers[178].compute_node[1][0], compute_graph.flexml_layers[178].compute_node[1][1], compute_graph.flexml_layers[178].compute_node[1][2], compute_graph.flexml_layers[178].compute_node[1][3], compute_graph.flexml_layers[178].compute_node[2][0], compute_graph.flexml_layers[178].compute_node[2][1], compute_graph.flexml_layers[178].compute_node[2][2], compute_graph.flexml_layers[178].compute_node[2][3], compute_graph.flexml_layers[178].compute_node[3][0], compute_graph.flexml_layers[178].compute_node[3][1], compute_graph.flexml_layers[178].compute_node[3][2], compute_graph.flexml_layers[178].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[3], compute_graph.l2_202.in[0], compute_graph.l2_202.in[1], compute_graph.l2_202.in[2], compute_graph.l2_202.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(879): mode(3), layer(223): {compute_graph.Layer_203_wts_ddr.out[0], compute_graph.Layer_203_wts_ddr.out[1], compute_graph.Layer_203_l2_wts.in[0], compute_graph.Layer_203_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(879) is pipelined with KernelLayerNode(1506) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183 overlaps with memory space used by compute_graph.l2_203. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1108): mode(0), layer(223): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[3], compute_graph.l2_202.out[0], compute_graph.l2_202.out[1], compute_graph.l2_202.out[2], compute_graph.l2_202.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1507), layer(223): compute_graph.flexml_layers[179].compute_node[0][0], compute_graph.flexml_layers[179].compute_node[0][1], compute_graph.flexml_layers[179].compute_node[0][2], compute_graph.flexml_layers[179].compute_node[0][3], compute_graph.flexml_layers[179].compute_node[1][0], compute_graph.flexml_layers[179].compute_node[1][1], compute_graph.flexml_layers[179].compute_node[1][2], compute_graph.flexml_layers[179].compute_node[1][3], compute_graph.flexml_layers[179].compute_node[2][0], compute_graph.flexml_layers[179].compute_node[2][1], compute_graph.flexml_layers[179].compute_node[2][2], compute_graph.flexml_layers[179].compute_node[2][3], compute_graph.flexml_layers[179].compute_node[3][0], compute_graph.flexml_layers[179].compute_node[3][1], compute_graph.flexml_layers[179].compute_node[3][2], compute_graph.flexml_layers[179].compute_node[3][3], {compute_graph.Layer_203_l2_wts.out[0], compute_graph.Layer_203_l2_wts.out[1], compute_graph.Layer_203_l2_wts.out[2], compute_graph.Layer_203_l2_wts.out[3], compute_graph.l2_202.out[0], compute_graph.l2_202.out[1], compute_graph.l2_202.out[2], compute_graph.l2_202.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[3], compute_graph.l2_203.in[0], compute_graph.l2_203.in[1], compute_graph.l2_203.in[2], compute_graph.l2_203.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(880): mode(0), layer(224): {compute_graph.Layer_204_wts_ddr.out[0], compute_graph.Layer_204_wts_ddr.out[1], compute_graph.Layer_204_l2_wts.in[0], compute_graph.Layer_204_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.Layer_204_l2_wts.out[0], compute_graph.Layer_204_l2_wts.out[1], compute_graph.Layer_204_l2_wts.out[2], compute_graph.Layer_204_l2_wts.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1508), layer(224): compute_graph.flexml_layers[180].compute_node[0][0], compute_graph.flexml_layers[180].compute_node[0][1], compute_graph.flexml_layers[180].compute_node[0][2], compute_graph.flexml_layers[180].compute_node[0][3], compute_graph.flexml_layers[180].compute_node[1][0], compute_graph.flexml_layers[180].compute_node[1][1], compute_graph.flexml_layers[180].compute_node[1][2], compute_graph.flexml_layers[180].compute_node[1][3], compute_graph.flexml_layers[180].compute_node[2][0], compute_graph.flexml_layers[180].compute_node[2][1], compute_graph.flexml_layers[180].compute_node[2][2], compute_graph.flexml_layers[180].compute_node[2][3], compute_graph.flexml_layers[180].compute_node[3][0], compute_graph.flexml_layers[180].compute_node[3][1], compute_graph.flexml_layers[180].compute_node[3][2], compute_graph.flexml_layers[180].compute_node[3][3], {compute_graph.Layer_204_l2_wts.out[0], compute_graph.Layer_204_l2_wts.out[1], compute_graph.Layer_204_l2_wts.out[2], compute_graph.Layer_204_l2_wts.out[3], compute_graph.l2_203.out[0], compute_graph.l2_203.out[1], compute_graph.l2_203.out[2], compute_graph.l2_203.out[3], compute_graph.l2_204.in[0], compute_graph.l2_204.in[1], compute_graph.l2_204.in[2], compute_graph.l2_204.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1509), layer(225): compute_graph.flexml_layers[181].compute_node[0][0], compute_graph.flexml_layers[181].compute_node[0][1], compute_graph.flexml_layers[181].compute_node[0][2], compute_graph.flexml_layers[181].compute_node[0][3], compute_graph.flexml_layers[181].compute_node[1][0], compute_graph.flexml_layers[181].compute_node[1][1], compute_graph.flexml_layers[181].compute_node[1][2], compute_graph.flexml_layers[181].compute_node[1][3], compute_graph.flexml_layers[181].compute_node[2][0], compute_graph.flexml_layers[181].compute_node[2][1], compute_graph.flexml_layers[181].compute_node[2][2], compute_graph.flexml_layers[181].compute_node[2][3], compute_graph.flexml_layers[181].compute_node[3][0], compute_graph.flexml_layers[181].compute_node[3][1], compute_graph.flexml_layers[181].compute_node[3][2], compute_graph.flexml_layers[181].compute_node[3][3], {compute_graph.l2_204.out[0], compute_graph.l2_204.out[1], compute_graph.l2_204.out[2], compute_graph.l2_204.out[3], compute_graph.l2_205.in[0], compute_graph.l2_205.in[1], compute_graph.l2_205.in[2], compute_graph.l2_205.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(955): mode(3), layer(225): {compute_graph.l2_204.out[4], compute_graph.l2_204.out[5], compute_graph.l2l3_204_spill.in[0], compute_graph.l2l3_204_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(955) is pipelined with KernelLayerNode(1509) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1510), layer(226): compute_graph.flexml_layers[182].compute_node[0][0], compute_graph.flexml_layers[182].compute_node[0][1], compute_graph.flexml_layers[182].compute_node[0][2], compute_graph.flexml_layers[182].compute_node[0][3], compute_graph.flexml_layers[182].compute_node[1][0], compute_graph.flexml_layers[182].compute_node[1][1], compute_graph.flexml_layers[182].compute_node[1][2], compute_graph.flexml_layers[182].compute_node[1][3], compute_graph.flexml_layers[182].compute_node[2][0], compute_graph.flexml_layers[182].compute_node[2][1], compute_graph.flexml_layers[182].compute_node[2][2], compute_graph.flexml_layers[182].compute_node[2][3], compute_graph.flexml_layers[182].compute_node[3][0], compute_graph.flexml_layers[182].compute_node[3][1], compute_graph.flexml_layers[182].compute_node[3][2], compute_graph.flexml_layers[182].compute_node[3][3], {compute_graph.l2_205.out[0], compute_graph.l2_205.out[1], compute_graph.l2_205.out[2], compute_graph.l2_205.out[3], compute_graph.l2_206.in[0], compute_graph.l2_206.in[1], compute_graph.l2_206.in[2], compute_graph.l2_206.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1511), layer(227): compute_graph.flexml_layers[183].compute_node[0][0], compute_graph.flexml_layers[183].compute_node[0][1], compute_graph.flexml_layers[183].compute_node[0][2], compute_graph.flexml_layers[183].compute_node[0][3], compute_graph.flexml_layers[183].compute_node[1][0], compute_graph.flexml_layers[183].compute_node[1][1], compute_graph.flexml_layers[183].compute_node[1][2], compute_graph.flexml_layers[183].compute_node[1][3], compute_graph.flexml_layers[183].compute_node[2][0], compute_graph.flexml_layers[183].compute_node[2][1], compute_graph.flexml_layers[183].compute_node[2][2], compute_graph.flexml_layers[183].compute_node[2][3], compute_graph.flexml_layers[183].compute_node[3][0], compute_graph.flexml_layers[183].compute_node[3][1], compute_graph.flexml_layers[183].compute_node[3][2], compute_graph.flexml_layers[183].compute_node[3][3], {compute_graph.l2_206.out[0], compute_graph.l2_206.out[1], compute_graph.l2_206.out[2], compute_graph.l2_206.out[3], compute_graph.l2_207.in[0], compute_graph.l2_207.in[1], compute_graph.l2_207.in[2], compute_graph.l2_207.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(956): mode(0), layer(227): {compute_graph.l2_207.out[0], compute_graph.l2_207.out[1], compute_graph.l2l3_207_spill.in[0], compute_graph.l2l3_207_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 957 will not be pipelined because it's in the same layer 228 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1605), layer(228): {compute_graph.l2l3_204_spill.out[0], compute_graph.l2l3_207_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1606), layer(228): {compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1607), layer(228): {compute_graph.l2l3_204_spill.out[1], compute_graph.l2l3_207_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1608), layer(228): {compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1512), layer(228): compute_graph.flexml_layers[184].compute_node[0][0], compute_graph.flexml_layers[184].compute_node[0][1], compute_graph.flexml_layers[184].compute_node[0][2], compute_graph.flexml_layers[184].compute_node[0][3], compute_graph.flexml_layers[184].compute_node[1][0], compute_graph.flexml_layers[184].compute_node[1][1], compute_graph.flexml_layers[184].compute_node[1][2], compute_graph.flexml_layers[184].compute_node[1][3], compute_graph.flexml_layers[184].compute_node[2][0], compute_graph.flexml_layers[184].compute_node[2][1], compute_graph.flexml_layers[184].compute_node[2][2], compute_graph.flexml_layers[184].compute_node[2][3], compute_graph.flexml_layers[184].compute_node[3][0], compute_graph.flexml_layers[184].compute_node[3][1], compute_graph.flexml_layers[184].compute_node[3][2], compute_graph.flexml_layers[184].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[3], compute_graph.l2_208.in[0], compute_graph.l2_208.in[1], compute_graph.l2_208.in[2], compute_graph.l2_208.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(957): mode(3), layer(228): {compute_graph.l2_208.out[0], compute_graph.l2_208.out[1], compute_graph.l2l3_208_spill.in[0], compute_graph.l2l3_208_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1123): mode(0), layer(229): {compute_graph.l2l3_208_spill.out[0], compute_graph.templated_graph_209.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1283): mode(0), layer(229): {compute_graph.templated_graph_209.trans_mt_ifm.out[0], compute_graph.templated_graph_209.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1284): mode(0), layer(229): {compute_graph.templated_graph_209.trans_mt_ofm.out[0], compute_graph.l2l3_209_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1125): mode(0), layer(230): {compute_graph.l2l3_209_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1513), layer(230): compute_graph.flexml_layers[185].compute_node[0][0], compute_graph.flexml_layers[185].compute_node[0][1], compute_graph.flexml_layers[185].compute_node[0][2], compute_graph.flexml_layers[185].compute_node[0][3], compute_graph.flexml_layers[185].compute_node[1][0], compute_graph.flexml_layers[185].compute_node[1][1], compute_graph.flexml_layers[185].compute_node[1][2], compute_graph.flexml_layers[185].compute_node[1][3], compute_graph.flexml_layers[185].compute_node[2][0], compute_graph.flexml_layers[185].compute_node[2][1], compute_graph.flexml_layers[185].compute_node[2][2], compute_graph.flexml_layers[185].compute_node[2][3], compute_graph.flexml_layers[185].compute_node[3][0], compute_graph.flexml_layers[185].compute_node[3][1], compute_graph.flexml_layers[185].compute_node[3][2], compute_graph.flexml_layers[185].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[3], compute_graph.l2_210.in[0], compute_graph.l2_210.in[1], compute_graph.l2_210.in[2], compute_graph.l2_210.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(881): mode(3), layer(231): {compute_graph.Layer_211_wts_ddr.out[0], compute_graph.Layer_211_wts_ddr.out[1], compute_graph.Layer_211_l2_wts.in[0], compute_graph.Layer_211_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(881) is pipelined with KernelLayerNode(1513) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1514), layer(231): compute_graph.flexml_layers[186].compute_node[0][0], compute_graph.flexml_layers[186].compute_node[0][1], compute_graph.flexml_layers[186].compute_node[0][2], compute_graph.flexml_layers[186].compute_node[0][3], compute_graph.flexml_layers[186].compute_node[1][0], compute_graph.flexml_layers[186].compute_node[1][1], compute_graph.flexml_layers[186].compute_node[1][2], compute_graph.flexml_layers[186].compute_node[1][3], compute_graph.flexml_layers[186].compute_node[2][0], compute_graph.flexml_layers[186].compute_node[2][1], compute_graph.flexml_layers[186].compute_node[2][2], compute_graph.flexml_layers[186].compute_node[2][3], compute_graph.flexml_layers[186].compute_node[3][0], compute_graph.flexml_layers[186].compute_node[3][1], compute_graph.flexml_layers[186].compute_node[3][2], compute_graph.flexml_layers[186].compute_node[3][3], {compute_graph.Layer_211_l2_wts.out[0], compute_graph.Layer_211_l2_wts.out[1], compute_graph.Layer_211_l2_wts.out[2], compute_graph.Layer_211_l2_wts.out[3], compute_graph.l2_210.out[0], compute_graph.l2_210.out[1], compute_graph.l2_210.out[2], compute_graph.l2_210.out[3], compute_graph.l2_211.in[0], compute_graph.l2_211.in[1], compute_graph.l2_211.in[2], compute_graph.l2_211.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1515), layer(232): compute_graph.flexml_layers[187].compute_node[0][0], compute_graph.flexml_layers[187].compute_node[0][1], compute_graph.flexml_layers[187].compute_node[0][2], compute_graph.flexml_layers[187].compute_node[0][3], compute_graph.flexml_layers[187].compute_node[1][0], compute_graph.flexml_layers[187].compute_node[1][1], compute_graph.flexml_layers[187].compute_node[1][2], compute_graph.flexml_layers[187].compute_node[1][3], compute_graph.flexml_layers[187].compute_node[2][0], compute_graph.flexml_layers[187].compute_node[2][1], compute_graph.flexml_layers[187].compute_node[2][2], compute_graph.flexml_layers[187].compute_node[2][3], compute_graph.flexml_layers[187].compute_node[3][0], compute_graph.flexml_layers[187].compute_node[3][1], compute_graph.flexml_layers[187].compute_node[3][2], compute_graph.flexml_layers[187].compute_node[3][3], {compute_graph.l2_211.out[0], compute_graph.l2_211.out[1], compute_graph.l2_211.out[2], compute_graph.l2_211.out[3], compute_graph.l2_212.in[0], compute_graph.l2_212.in[1], compute_graph.l2_212.in[2], compute_graph.l2_212.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(958): mode(0), layer(232): {compute_graph.l2_212.out[1], compute_graph.l2l3_212_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1126): mode(0), layer(233): {compute_graph.l2l3_212_spill.out[0], compute_graph.templated_graph_213.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1285): mode(0), layer(233): {compute_graph.templated_graph_213.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_213_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1127): mode(0), layer(234): {compute_graph.l2l3_scratch_0_213_spill.out[0], compute_graph.templated_graph_213.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1286): mode(0), layer(234): {compute_graph.templated_graph_213.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_213_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1128): mode(0), layer(235): {compute_graph.l2l3_scratch_1_213_spill.out[0], compute_graph.templated_graph_213.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1287): mode(0), layer(235): {compute_graph.templated_graph_213.g2.ifm_mem.out[0], compute_graph.l2l3_213_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x70800): Memory space used by compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1 overlaps with memory space used by compute_graph.l2_214. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 1, row 0 (Address: 0x0): Memory space used by compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1 overlaps with memory space used by compute_graph.l2_214. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 1, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1129): mode(0), layer(236): {compute_graph.l2l3_213_spill.out[0], compute_graph.l2l3_213_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(882): mode(4), layer(236): {compute_graph.Layer_214_wts_ddr.out[0], compute_graph.Layer_214_wts_ddr.out[1], compute_graph.Layer_214_l2_wts.in[0], compute_graph.Layer_214_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(882) is pipelined with BufferToBufferNode(1129) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1124): mode(0), layer(236): {compute_graph.l2l3_208_spill.out[1], compute_graph.l2l3_208_spill.out[2], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1516), layer(236): compute_graph.flexml_layers[188].compute_node[0][0], compute_graph.flexml_layers[188].compute_node[0][1], compute_graph.flexml_layers[188].compute_node[0][2], compute_graph.flexml_layers[188].compute_node[0][3], compute_graph.flexml_layers[188].compute_node[1][0], compute_graph.flexml_layers[188].compute_node[1][1], compute_graph.flexml_layers[188].compute_node[1][2], compute_graph.flexml_layers[188].compute_node[1][3], compute_graph.flexml_layers[188].compute_node[2][0], compute_graph.flexml_layers[188].compute_node[2][1], compute_graph.flexml_layers[188].compute_node[2][2], compute_graph.flexml_layers[188].compute_node[2][3], compute_graph.flexml_layers[188].compute_node[3][0], compute_graph.flexml_layers[188].compute_node[3][1], compute_graph.flexml_layers[188].compute_node[3][2], compute_graph.flexml_layers[188].compute_node[3][3], {compute_graph.Layer_214_l2_wts.out[0], compute_graph.Layer_214_l2_wts.out[1], compute_graph.Layer_214_l2_wts.out[2], compute_graph.Layer_214_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[3], compute_graph.l2_214.in[0], compute_graph.l2_214.in[1], compute_graph.l2_214.in[2], compute_graph.l2_214.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(959): mode(0), layer(236): {compute_graph.l2_214.out[0], compute_graph.l2_214.out[1], compute_graph.l2l3_214_spill.in[0], compute_graph.l2l3_214_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1130): mode(0), layer(237): {compute_graph.l2l3_214_spill.out[0], compute_graph.l2l3_214_spill.out[1], compute_graph.templated_graph_215.ifm.in[0], compute_graph.templated_graph_215.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1288): mode(0), layer(237): {compute_graph.templated_graph_215.ifm.out[0], compute_graph.templated_graph_215.ifm.out[1], compute_graph.l2l3_215_spill.in[0], compute_graph.l2l3_215_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1131): mode(0), layer(238): {compute_graph.l2l3_214_spill.out[2], compute_graph.l2l3_214_spill.out[3], compute_graph.templated_graph_216.ifm.in[0], compute_graph.templated_graph_216.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1289): mode(0), layer(238): {compute_graph.templated_graph_216.ifm.out[0], compute_graph.templated_graph_216.ifm.out[1], compute_graph.l2l3_216_spill.in[0], compute_graph.l2l3_216_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1021): mode(0), layer(239): {compute_graph.l2l3_5_spill.out[4], compute_graph.l2l3_5_spill.out[5], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1005): mode(0), layer(239): {compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.out[1], compute_graph.spill_L3_Concat_Buffer_layer_294.in[2], compute_graph.spill_L3_Concat_Buffer_layer_294.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1199): mode(0), layer(240): {compute_graph.ifm_ddr_1.out[0], compute_graph.ifm_ddr_1.out[1], compute_graph.L2_IFM_Buffer_for_input1_0_port1.in[0], compute_graph.L2_IFM_Buffer_for_input1_0_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(992): mode(0), layer(240): {compute_graph.L2_IFM_Buffer_for_input1_0_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_0_port1.out[1], compute_graph.spill_L3_Concat_Buffer_layer_279.in[2], compute_graph.spill_L3_Concat_Buffer_layer_279.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1157): mode(0), layer(241): {compute_graph.ifm_ddr_3.out[0], compute_graph.ifm_ddr_3.out[1], compute_graph.L2_IFM_Buffer_for_input3_0_port1.in[0], compute_graph.L2_IFM_Buffer_for_input3_0_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(976): mode(0), layer(241): {compute_graph.L2_IFM_Buffer_for_input3_0_port1.out[0], compute_graph.L2_IFM_Buffer_for_input3_0_port1.out[1], compute_graph.spill_L3_Concat_Buffer_layer_240.in[2], compute_graph.spill_L3_Concat_Buffer_layer_240.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1132): mode(0), layer(242): {compute_graph.l2l3_215_spill.out[0], compute_graph.l2l3_215_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(968): mode(0), layer(242): {compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_230.in[0], compute_graph.spill_L3_Concat_Buffer_layer_230.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1134): mode(0), layer(243): {compute_graph.l2l3_216_spill.out[2], compute_graph.l2l3_216_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(965): mode(0), layer(243): {compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_225.in[0], compute_graph.spill_L3_Concat_Buffer_layer_225.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1135): mode(0), layer(244): {compute_graph.ifm_ddr_4.out[0], compute_graph.ifm_ddr_4.out[1], compute_graph.L2_IFM_Buffer_for_input4_0_port1.in[0], compute_graph.L2_IFM_Buffer_for_input4_0_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(961): mode(0), layer(244): {compute_graph.L2_IFM_Buffer_for_input4_0_port1.out[0], compute_graph.L2_IFM_Buffer_for_input4_0_port1.out[1], compute_graph.spill_L3_Concat_Buffer_layer_217.in[2], compute_graph.spill_L3_Concat_Buffer_layer_217.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1133): mode(0), layer(245): {compute_graph.l2l3_216_spill.out[0], compute_graph.l2l3_216_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(960): mode(0), layer(245): {compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_217.in[0], compute_graph.spill_L3_Concat_Buffer_layer_217.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1138): mode(0), layer(246): {compute_graph.spill_L3_Concat_Buffer_layer_217.out[0], compute_graph.spill_L3_Concat_Buffer_layer_217.out[1], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(883): mode(4), layer(246): {compute_graph.Layer_218_wts_ddr.out[0], compute_graph.Layer_218_wts_ddr.out[1], compute_graph.Layer_218_l2_wts.in[0], compute_graph.Layer_218_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(883) is pipelined with BufferToBufferNode(1138) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1517), layer(246): compute_graph.flexml_layers[189].compute_node[0][0], compute_graph.flexml_layers[189].compute_node[0][1], compute_graph.flexml_layers[189].compute_node[0][2], compute_graph.flexml_layers[189].compute_node[0][3], compute_graph.flexml_layers[189].compute_node[1][0], compute_graph.flexml_layers[189].compute_node[1][1], compute_graph.flexml_layers[189].compute_node[1][2], compute_graph.flexml_layers[189].compute_node[1][3], compute_graph.flexml_layers[189].compute_node[2][0], compute_graph.flexml_layers[189].compute_node[2][1], compute_graph.flexml_layers[189].compute_node[2][2], compute_graph.flexml_layers[189].compute_node[2][3], compute_graph.flexml_layers[189].compute_node[3][0], compute_graph.flexml_layers[189].compute_node[3][1], compute_graph.flexml_layers[189].compute_node[3][2], compute_graph.flexml_layers[189].compute_node[3][3], {compute_graph.Layer_218_l2_wts.out[0], compute_graph.Layer_218_l2_wts.out[1], compute_graph.Layer_218_l2_wts.out[2], compute_graph.Layer_218_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[3], compute_graph.l2_218.in[0], compute_graph.l2_218.in[1], compute_graph.l2_218.in[2], compute_graph.l2_218.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1518), layer(247): compute_graph.flexml_layers[190].compute_node[0][0], compute_graph.flexml_layers[190].compute_node[0][1], compute_graph.flexml_layers[190].compute_node[0][2], compute_graph.flexml_layers[190].compute_node[0][3], compute_graph.flexml_layers[190].compute_node[1][0], compute_graph.flexml_layers[190].compute_node[1][1], compute_graph.flexml_layers[190].compute_node[1][2], compute_graph.flexml_layers[190].compute_node[1][3], compute_graph.flexml_layers[190].compute_node[2][0], compute_graph.flexml_layers[190].compute_node[2][1], compute_graph.flexml_layers[190].compute_node[2][2], compute_graph.flexml_layers[190].compute_node[2][3], compute_graph.flexml_layers[190].compute_node[3][0], compute_graph.flexml_layers[190].compute_node[3][1], compute_graph.flexml_layers[190].compute_node[3][2], compute_graph.flexml_layers[190].compute_node[3][3], {compute_graph.l2_218.out[0], compute_graph.l2_218.out[1], compute_graph.l2_218.out[2], compute_graph.l2_218.out[3], compute_graph.l2_219.in[0], compute_graph.l2_219.in[1], compute_graph.l2_219.in[2], compute_graph.l2_219.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(962): mode(0), layer(247): {compute_graph.l2_219.out[0], compute_graph.l2_219.out[1], compute_graph.l2l3_219_spill.in[0], compute_graph.l2l3_219_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1139): mode(0), layer(248): {compute_graph.l2l3_219_spill.out[0], compute_graph.l2l3_219_spill.out[1], compute_graph.templated_graph_220.ifm.in[0], compute_graph.templated_graph_220.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1290): mode(0), layer(248): {compute_graph.templated_graph_220.ifm.out[0], compute_graph.templated_graph_220.ifm.out[1], compute_graph.l2l3_220_spill.in[0], compute_graph.l2l3_220_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1141): mode(0), layer(249): {compute_graph.l2l3_220_spill.out[0], compute_graph.l2l3_220_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1143): mode(0), layer(249): {compute_graph.const_ifm_ddr_3.out[0], compute_graph.const_ifm_ddr_3.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1519), layer(249): compute_graph.flexml_layers[191].compute_node[0][0], compute_graph.flexml_layers[191].compute_node[0][1], compute_graph.flexml_layers[191].compute_node[0][2], compute_graph.flexml_layers[191].compute_node[0][3], compute_graph.flexml_layers[191].compute_node[1][0], compute_graph.flexml_layers[191].compute_node[1][1], compute_graph.flexml_layers[191].compute_node[1][2], compute_graph.flexml_layers[191].compute_node[1][3], compute_graph.flexml_layers[191].compute_node[2][0], compute_graph.flexml_layers[191].compute_node[2][1], compute_graph.flexml_layers[191].compute_node[2][2], compute_graph.flexml_layers[191].compute_node[2][3], compute_graph.flexml_layers[191].compute_node[3][0], compute_graph.flexml_layers[191].compute_node[3][1], compute_graph.flexml_layers[191].compute_node[3][2], compute_graph.flexml_layers[191].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[3], compute_graph.l2_221.in[0], compute_graph.l2_221.in[1], compute_graph.l2_221.in[2], compute_graph.l2_221.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1136): mode(0), layer(250): {compute_graph.ifm_ddr_4.out[2], compute_graph.ifm_ddr_4.out[3], compute_graph.L2_IFM_Buffer_for_input4_1_port1.in[0], compute_graph.L2_IFM_Buffer_for_input4_1_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1520), layer(250): compute_graph.flexml_layers[192].compute_node[0][0], compute_graph.flexml_layers[192].compute_node[0][1], compute_graph.flexml_layers[192].compute_node[0][2], compute_graph.flexml_layers[192].compute_node[0][3], compute_graph.flexml_layers[192].compute_node[1][0], compute_graph.flexml_layers[192].compute_node[1][1], compute_graph.flexml_layers[192].compute_node[1][2], compute_graph.flexml_layers[192].compute_node[1][3], compute_graph.flexml_layers[192].compute_node[2][0], compute_graph.flexml_layers[192].compute_node[2][1], compute_graph.flexml_layers[192].compute_node[2][2], compute_graph.flexml_layers[192].compute_node[2][3], compute_graph.flexml_layers[192].compute_node[3][0], compute_graph.flexml_layers[192].compute_node[3][1], compute_graph.flexml_layers[192].compute_node[3][2], compute_graph.flexml_layers[192].compute_node[3][3], {compute_graph.l2_221.out[0], compute_graph.l2_221.out[1], compute_graph.l2_221.out[2], compute_graph.l2_221.out[3], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[3], compute_graph.l2_222.in[0], compute_graph.l2_222.in[1], compute_graph.l2_222.in[2], compute_graph.l2_222.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(963): mode(0), layer(250): {compute_graph.l2_222.out[0], compute_graph.l2_222.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1140): mode(0), layer(251): {compute_graph.l2l3_219_spill.out[2], compute_graph.l2l3_219_spill.out[3], compute_graph.templated_graph_223.ifm.in[0], compute_graph.templated_graph_223.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1291): mode(0), layer(251): {compute_graph.templated_graph_223.ifm.out[0], compute_graph.templated_graph_223.ifm.out[1], compute_graph.l2l3_223_spill.in[0], compute_graph.l2l3_223_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1145): mode(0), layer(252): {compute_graph.l2l3_223_spill.out[0], compute_graph.l2l3_223_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1137): mode(0), layer(252): {compute_graph.ifm_ddr_4.out[4], compute_graph.ifm_ddr_4.out[5], compute_graph.L2_IFM_Buffer_for_input4_2_port1.in[0], compute_graph.L2_IFM_Buffer_for_input4_2_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1521), layer(252): compute_graph.flexml_layers[193].compute_node[0][0], compute_graph.flexml_layers[193].compute_node[0][1], compute_graph.flexml_layers[193].compute_node[0][2], compute_graph.flexml_layers[193].compute_node[0][3], compute_graph.flexml_layers[193].compute_node[1][0], compute_graph.flexml_layers[193].compute_node[1][1], compute_graph.flexml_layers[193].compute_node[1][2], compute_graph.flexml_layers[193].compute_node[1][3], compute_graph.flexml_layers[193].compute_node[2][0], compute_graph.flexml_layers[193].compute_node[2][1], compute_graph.flexml_layers[193].compute_node[2][2], compute_graph.flexml_layers[193].compute_node[2][3], compute_graph.flexml_layers[193].compute_node[3][0], compute_graph.flexml_layers[193].compute_node[3][1], compute_graph.flexml_layers[193].compute_node[3][2], compute_graph.flexml_layers[193].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[0], compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[1], compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[2], compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[3], compute_graph.l2_224.in[0], compute_graph.l2_224.in[1], compute_graph.l2_224.in[2], compute_graph.l2_224.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(884): mode(3), layer(253): {compute_graph.Layer_226_wts_ddr.out[0], compute_graph.Layer_226_wts_ddr.out[1], compute_graph.Layer_226_l2_wts.in[0], compute_graph.Layer_226_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(884) is pipelined with KernelLayerNode(1521) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(964): mode(0), layer(252): {compute_graph.l2_224.out[0], compute_graph.l2_224.out[1], compute_graph.spill_L3_Concat_Buffer_layer_225.in[2], compute_graph.spill_L3_Concat_Buffer_layer_225.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1146): mode(0), layer(253): {compute_graph.spill_L3_Concat_Buffer_layer_225.out[0], compute_graph.spill_L3_Concat_Buffer_layer_225.out[1], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1522), layer(253): compute_graph.flexml_layers[194].compute_node[0][0], compute_graph.flexml_layers[194].compute_node[0][1], compute_graph.flexml_layers[194].compute_node[0][2], compute_graph.flexml_layers[194].compute_node[0][3], compute_graph.flexml_layers[194].compute_node[1][0], compute_graph.flexml_layers[194].compute_node[1][1], compute_graph.flexml_layers[194].compute_node[1][2], compute_graph.flexml_layers[194].compute_node[1][3], compute_graph.flexml_layers[194].compute_node[2][0], compute_graph.flexml_layers[194].compute_node[2][1], compute_graph.flexml_layers[194].compute_node[2][2], compute_graph.flexml_layers[194].compute_node[2][3], compute_graph.flexml_layers[194].compute_node[3][0], compute_graph.flexml_layers[194].compute_node[3][1], compute_graph.flexml_layers[194].compute_node[3][2], compute_graph.flexml_layers[194].compute_node[3][3], {compute_graph.Layer_226_l2_wts.out[0], compute_graph.Layer_226_l2_wts.out[1], compute_graph.Layer_226_l2_wts.out[2], compute_graph.Layer_226_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[3], compute_graph.l2_226.in[0], compute_graph.l2_226.in[1], compute_graph.l2_226.in[2], compute_graph.l2_226.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1523), layer(254): compute_graph.flexml_layers[195].compute_node[0][0], compute_graph.flexml_layers[195].compute_node[0][1], compute_graph.flexml_layers[195].compute_node[0][2], compute_graph.flexml_layers[195].compute_node[0][3], compute_graph.flexml_layers[195].compute_node[1][0], compute_graph.flexml_layers[195].compute_node[1][1], compute_graph.flexml_layers[195].compute_node[1][2], compute_graph.flexml_layers[195].compute_node[1][3], compute_graph.flexml_layers[195].compute_node[2][0], compute_graph.flexml_layers[195].compute_node[2][1], compute_graph.flexml_layers[195].compute_node[2][2], compute_graph.flexml_layers[195].compute_node[2][3], compute_graph.flexml_layers[195].compute_node[3][0], compute_graph.flexml_layers[195].compute_node[3][1], compute_graph.flexml_layers[195].compute_node[3][2], compute_graph.flexml_layers[195].compute_node[3][3], {compute_graph.l2_226.out[0], compute_graph.l2_226.out[1], compute_graph.l2_226.out[2], compute_graph.l2_226.out[3], compute_graph.l2_227.in[0], compute_graph.l2_227.in[1], compute_graph.l2_227.in[2], compute_graph.l2_227.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1142): mode(0), layer(255): {compute_graph.l2l3_220_spill.out[2], compute_graph.l2l3_220_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1524), layer(255): compute_graph.flexml_layers[196].compute_node[0][0], compute_graph.flexml_layers[196].compute_node[0][1], compute_graph.flexml_layers[196].compute_node[0][2], compute_graph.flexml_layers[196].compute_node[0][3], compute_graph.flexml_layers[196].compute_node[1][0], compute_graph.flexml_layers[196].compute_node[1][1], compute_graph.flexml_layers[196].compute_node[1][2], compute_graph.flexml_layers[196].compute_node[1][3], compute_graph.flexml_layers[196].compute_node[2][0], compute_graph.flexml_layers[196].compute_node[2][1], compute_graph.flexml_layers[196].compute_node[2][2], compute_graph.flexml_layers[196].compute_node[2][3], compute_graph.flexml_layers[196].compute_node[3][0], compute_graph.flexml_layers[196].compute_node[3][1], compute_graph.flexml_layers[196].compute_node[3][2], compute_graph.flexml_layers[196].compute_node[3][3], {compute_graph.l2_227.out[0], compute_graph.l2_227.out[1], compute_graph.l2_227.out[2], compute_graph.l2_227.out[3], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[3], compute_graph.l2_228.in[0], compute_graph.l2_228.in[1], compute_graph.l2_228.in[2], compute_graph.l2_228.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1144): mode(0), layer(256): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1525), layer(256): compute_graph.flexml_layers[197].compute_node[0][0], compute_graph.flexml_layers[197].compute_node[0][1], compute_graph.flexml_layers[197].compute_node[0][2], compute_graph.flexml_layers[197].compute_node[0][3], compute_graph.flexml_layers[197].compute_node[1][0], compute_graph.flexml_layers[197].compute_node[1][1], compute_graph.flexml_layers[197].compute_node[1][2], compute_graph.flexml_layers[197].compute_node[1][3], compute_graph.flexml_layers[197].compute_node[2][0], compute_graph.flexml_layers[197].compute_node[2][1], compute_graph.flexml_layers[197].compute_node[2][2], compute_graph.flexml_layers[197].compute_node[2][3], compute_graph.flexml_layers[197].compute_node[3][0], compute_graph.flexml_layers[197].compute_node[3][1], compute_graph.flexml_layers[197].compute_node[3][2], compute_graph.flexml_layers[197].compute_node[3][3], {compute_graph.l2_228.out[0], compute_graph.l2_228.out[1], compute_graph.l2_228.out[2], compute_graph.l2_228.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[3], compute_graph.l2_229.in[0], compute_graph.l2_229.in[1], compute_graph.l2_229.in[2], compute_graph.l2_229.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(967): mode(0), layer(256): {compute_graph.l2_229.out[2], compute_graph.l2_229.out[3], compute_graph.spill_L3_Concat_Buffer_layer_230.in[2], compute_graph.spill_L3_Concat_Buffer_layer_230.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(966): mode(0), layer(256): {compute_graph.l2_229.out[0], compute_graph.l2_229.out[1], compute_graph.ofm_ddr_0_l2l3_229_spill.in[0], compute_graph.ofm_ddr_0_l2l3_229_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1147): mode(0), layer(257): {compute_graph.spill_L3_Concat_Buffer_layer_230.out[0], compute_graph.spill_L3_Concat_Buffer_layer_230.out[1], compute_graph.templated_graph_231.ifm_mem.in[0], compute_graph.templated_graph_231.ifm_mem.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1293), layer(257): compute_graph.templated_graph_231.mk[0][0], compute_graph.templated_graph_231.mk[0][1], compute_graph.templated_graph_231.mk[0][2], compute_graph.templated_graph_231.mk[0][3], compute_graph.templated_graph_231.mk[1][0], compute_graph.templated_graph_231.mk[1][1], compute_graph.templated_graph_231.mk[1][2], compute_graph.templated_graph_231.mk[1][3], compute_graph.templated_graph_231.mk[2][0], compute_graph.templated_graph_231.mk[2][1], compute_graph.templated_graph_231.mk[2][2], compute_graph.templated_graph_231.mk[2][3], compute_graph.templated_graph_231.mk[3][0], compute_graph.templated_graph_231.mk[3][1], compute_graph.templated_graph_231.mk[3][2], compute_graph.templated_graph_231.mk[3][3], {compute_graph.templated_graph_231.ifm_mem.out[0], compute_graph.templated_graph_231.ifm_mem.out[1], compute_graph.templated_graph_231.ifm_mem.out[2], compute_graph.templated_graph_231.ifm_mem.out[3], compute_graph.templated_graph_231.ofm_mem.in[0], compute_graph.templated_graph_231.ofm_mem.in[1], compute_graph.templated_graph_231.ofm_mem.in[2], compute_graph.templated_graph_231.ofm_mem.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1292): mode(0), layer(257): {compute_graph.templated_graph_231.ofm_mem.out[0], compute_graph.templated_graph_231.ofm_mem.out[1], compute_graph.l2l3_231_spill.in[0], compute_graph.l2l3_231_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1148): mode(0), layer(258): {compute_graph.l2l3_231_spill.out[0], compute_graph.l2l3_231_spill.out[1], compute_graph.templated_graph_232.ifm.in[0], compute_graph.templated_graph_232.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1294): mode(0), layer(258): {compute_graph.templated_graph_232.ifm.out[0], compute_graph.templated_graph_232.ifm.out[1], compute_graph.l2l3_232_spill.in[0], compute_graph.l2l3_232_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1020): mode(0), layer(259): {compute_graph.l2l3_5_spill.out[2], compute_graph.l2l3_5_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1526), layer(259): compute_graph.flexml_layers[198].compute_node[0][0], compute_graph.flexml_layers[198].compute_node[0][1], compute_graph.flexml_layers[198].compute_node[0][2], compute_graph.flexml_layers[198].compute_node[0][3], compute_graph.flexml_layers[198].compute_node[1][0], compute_graph.flexml_layers[198].compute_node[1][1], compute_graph.flexml_layers[198].compute_node[1][2], compute_graph.flexml_layers[198].compute_node[1][3], compute_graph.flexml_layers[198].compute_node[2][0], compute_graph.flexml_layers[198].compute_node[2][1], compute_graph.flexml_layers[198].compute_node[2][2], compute_graph.flexml_layers[198].compute_node[2][3], compute_graph.flexml_layers[198].compute_node[3][0], compute_graph.flexml_layers[198].compute_node[3][1], compute_graph.flexml_layers[198].compute_node[3][2], compute_graph.flexml_layers[198].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[3], compute_graph.l2_233.in[0], compute_graph.l2_233.in[1], compute_graph.l2_233.in[2], compute_graph.l2_233.in[3]}, Scheduler computes number of sub-layer phases to be 24. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1609), layer(259): {compute_graph.l2_233.out[0], compute_graph.l2_233.out[2]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1610), layer(259): {compute_graph.l2l3_233_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[4]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1611), layer(259): {compute_graph.l2_233.out[1], compute_graph.l2_233.out[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1612), layer(259): {compute_graph.l2l3_233_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[5]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1150): mode(0), layer(260): {compute_graph.l2l3_233_spill.out[0], compute_graph.l2l3_233_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1527), layer(260): compute_graph.flexml_layers[199].compute_node[0][0], compute_graph.flexml_layers[199].compute_node[0][1], compute_graph.flexml_layers[199].compute_node[0][2], compute_graph.flexml_layers[199].compute_node[0][3], compute_graph.flexml_layers[199].compute_node[1][0], compute_graph.flexml_layers[199].compute_node[1][1], compute_graph.flexml_layers[199].compute_node[1][2], compute_graph.flexml_layers[199].compute_node[1][3], compute_graph.flexml_layers[199].compute_node[2][0], compute_graph.flexml_layers[199].compute_node[2][1], compute_graph.flexml_layers[199].compute_node[2][2], compute_graph.flexml_layers[199].compute_node[2][3], compute_graph.flexml_layers[199].compute_node[3][0], compute_graph.flexml_layers[199].compute_node[3][1], compute_graph.flexml_layers[199].compute_node[3][2], compute_graph.flexml_layers[199].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[3], compute_graph.l2_234.in[0], compute_graph.l2_234.in[1], compute_graph.l2_234.in[2], compute_graph.l2_234.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1528), layer(261): compute_graph.flexml_layers[200].compute_node[0][0], compute_graph.flexml_layers[200].compute_node[0][1], compute_graph.flexml_layers[200].compute_node[0][2], compute_graph.flexml_layers[200].compute_node[0][3], compute_graph.flexml_layers[200].compute_node[1][0], compute_graph.flexml_layers[200].compute_node[1][1], compute_graph.flexml_layers[200].compute_node[1][2], compute_graph.flexml_layers[200].compute_node[1][3], compute_graph.flexml_layers[200].compute_node[2][0], compute_graph.flexml_layers[200].compute_node[2][1], compute_graph.flexml_layers[200].compute_node[2][2], compute_graph.flexml_layers[200].compute_node[2][3], compute_graph.flexml_layers[200].compute_node[3][0], compute_graph.flexml_layers[200].compute_node[3][1], compute_graph.flexml_layers[200].compute_node[3][2], compute_graph.flexml_layers[200].compute_node[3][3], {compute_graph.l2_234.out[0], compute_graph.l2_234.out[1], compute_graph.l2_234.out[2], compute_graph.l2_234.out[3], compute_graph.l2_235.in[0], compute_graph.l2_235.in[1], compute_graph.l2_235.in[2], compute_graph.l2_235.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(972): mode(3), layer(261): {compute_graph.l2_234.out[4], compute_graph.l2_234.out[5], compute_graph.spill_L3_Concat_Buffer_layer_256.in[4], compute_graph.spill_L3_Concat_Buffer_layer_256.in[5]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(972) is pipelined with KernelLayerNode(1528) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(973): mode(0), layer(261): {compute_graph.l2_235.out[0], compute_graph.l2_235.out[1], compute_graph.spill_L3_Concat_Buffer_layer_236.in[4], compute_graph.spill_L3_Concat_Buffer_layer_236.in[5]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1149): mode(0), layer(262): {compute_graph.l2l3_232_spill.out[0], compute_graph.l2l3_232_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(969): mode(0), layer(262): {compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_236.in[0], compute_graph.spill_L3_Concat_Buffer_layer_236.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1151): mode(0), layer(263): {compute_graph.spill_L3_Concat_Buffer_layer_236.out[0], compute_graph.spill_L3_Concat_Buffer_layer_236.out[1], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(885): mode(4), layer(263): {compute_graph.Layer_237_wts_ddr.out[0], compute_graph.Layer_237_wts_ddr.out[1], compute_graph.Layer_237_l2_wts.in[0], compute_graph.Layer_237_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(885) is pipelined with BufferToBufferNode(1151) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1529), layer(263): compute_graph.flexml_layers[201].compute_node[0][0], compute_graph.flexml_layers[201].compute_node[0][1], compute_graph.flexml_layers[201].compute_node[0][2], compute_graph.flexml_layers[201].compute_node[0][3], compute_graph.flexml_layers[201].compute_node[1][0], compute_graph.flexml_layers[201].compute_node[1][1], compute_graph.flexml_layers[201].compute_node[1][2], compute_graph.flexml_layers[201].compute_node[1][3], compute_graph.flexml_layers[201].compute_node[2][0], compute_graph.flexml_layers[201].compute_node[2][1], compute_graph.flexml_layers[201].compute_node[2][2], compute_graph.flexml_layers[201].compute_node[2][3], compute_graph.flexml_layers[201].compute_node[3][0], compute_graph.flexml_layers[201].compute_node[3][1], compute_graph.flexml_layers[201].compute_node[3][2], compute_graph.flexml_layers[201].compute_node[3][3], {compute_graph.Layer_237_l2_wts.out[0], compute_graph.Layer_237_l2_wts.out[1], compute_graph.Layer_237_l2_wts.out[2], compute_graph.Layer_237_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[3], compute_graph.l2_237.in[0], compute_graph.l2_237.in[1], compute_graph.l2_237.in[2], compute_graph.l2_237.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(974): mode(0), layer(263): {compute_graph.l2_237.out[0], compute_graph.l2_237.out[1], compute_graph.l2l3_237_spill.in[0], compute_graph.l2l3_237_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1152): mode(0), layer(264): {compute_graph.l2l3_237_spill.out[0], compute_graph.l2l3_237_spill.out[1], compute_graph.templated_graph_238.ifm.in[0], compute_graph.templated_graph_238.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1295): mode(0), layer(264): {compute_graph.templated_graph_238.ifm.out[0], compute_graph.templated_graph_238.ifm.out[1], compute_graph.l2l3_238_spill.in[0], compute_graph.l2l3_238_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1153): mode(0), layer(265): {compute_graph.l2l3_237_spill.out[2], compute_graph.l2l3_237_spill.out[3], compute_graph.templated_graph_239.ifm.in[0], compute_graph.templated_graph_239.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1296): mode(0), layer(265): {compute_graph.templated_graph_239.ifm.out[0], compute_graph.templated_graph_239.ifm.out[1], compute_graph.l2l3_239_spill.in[0], compute_graph.l2l3_239_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1156): mode(0), layer(266): {compute_graph.l2l3_239_spill.out[2], compute_graph.l2l3_239_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(980): mode(0), layer(266): {compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_248.in[0], compute_graph.spill_L3_Concat_Buffer_layer_248.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1155): mode(0), layer(267): {compute_graph.l2l3_239_spill.out[0], compute_graph.l2l3_239_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(975): mode(0), layer(267): {compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_240.in[0], compute_graph.spill_L3_Concat_Buffer_layer_240.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1160): mode(0), layer(268): {compute_graph.spill_L3_Concat_Buffer_layer_240.out[0], compute_graph.spill_L3_Concat_Buffer_layer_240.out[1], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(886): mode(4), layer(268): {compute_graph.Layer_241_wts_ddr.out[0], compute_graph.Layer_241_wts_ddr.out[1], compute_graph.Layer_241_l2_wts.in[0], compute_graph.Layer_241_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(886) is pipelined with BufferToBufferNode(1160) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1530), layer(268): compute_graph.flexml_layers[202].compute_node[0][0], compute_graph.flexml_layers[202].compute_node[0][1], compute_graph.flexml_layers[202].compute_node[0][2], compute_graph.flexml_layers[202].compute_node[0][3], compute_graph.flexml_layers[202].compute_node[1][0], compute_graph.flexml_layers[202].compute_node[1][1], compute_graph.flexml_layers[202].compute_node[1][2], compute_graph.flexml_layers[202].compute_node[1][3], compute_graph.flexml_layers[202].compute_node[2][0], compute_graph.flexml_layers[202].compute_node[2][1], compute_graph.flexml_layers[202].compute_node[2][2], compute_graph.flexml_layers[202].compute_node[2][3], compute_graph.flexml_layers[202].compute_node[3][0], compute_graph.flexml_layers[202].compute_node[3][1], compute_graph.flexml_layers[202].compute_node[3][2], compute_graph.flexml_layers[202].compute_node[3][3], {compute_graph.Layer_241_l2_wts.out[0], compute_graph.Layer_241_l2_wts.out[1], compute_graph.Layer_241_l2_wts.out[2], compute_graph.Layer_241_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[3], compute_graph.l2_241.in[0], compute_graph.l2_241.in[1], compute_graph.l2_241.in[2], compute_graph.l2_241.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1531), layer(269): compute_graph.flexml_layers[203].compute_node[0][0], compute_graph.flexml_layers[203].compute_node[0][1], compute_graph.flexml_layers[203].compute_node[0][2], compute_graph.flexml_layers[203].compute_node[0][3], compute_graph.flexml_layers[203].compute_node[1][0], compute_graph.flexml_layers[203].compute_node[1][1], compute_graph.flexml_layers[203].compute_node[1][2], compute_graph.flexml_layers[203].compute_node[1][3], compute_graph.flexml_layers[203].compute_node[2][0], compute_graph.flexml_layers[203].compute_node[2][1], compute_graph.flexml_layers[203].compute_node[2][2], compute_graph.flexml_layers[203].compute_node[2][3], compute_graph.flexml_layers[203].compute_node[3][0], compute_graph.flexml_layers[203].compute_node[3][1], compute_graph.flexml_layers[203].compute_node[3][2], compute_graph.flexml_layers[203].compute_node[3][3], {compute_graph.l2_241.out[0], compute_graph.l2_241.out[1], compute_graph.l2_241.out[2], compute_graph.l2_241.out[3], compute_graph.l2_242.in[0], compute_graph.l2_242.in[1], compute_graph.l2_242.in[2], compute_graph.l2_242.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(977): mode(0), layer(269): {compute_graph.l2_242.out[0], compute_graph.l2_242.out[1], compute_graph.l2l3_242_spill.in[0], compute_graph.l2l3_242_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1161): mode(0), layer(270): {compute_graph.l2l3_242_spill.out[0], compute_graph.l2l3_242_spill.out[1], compute_graph.templated_graph_243.ifm.in[0], compute_graph.templated_graph_243.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1297): mode(0), layer(270): {compute_graph.templated_graph_243.ifm.out[0], compute_graph.templated_graph_243.ifm.out[1], compute_graph.l2l3_243_spill.in[0], compute_graph.l2l3_243_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1163): mode(0), layer(271): {compute_graph.l2l3_243_spill.out[0], compute_graph.l2l3_243_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1165): mode(0), layer(271): {compute_graph.const_ifm_ddr_2.out[0], compute_graph.const_ifm_ddr_2.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1532), layer(271): compute_graph.flexml_layers[204].compute_node[0][0], compute_graph.flexml_layers[204].compute_node[0][1], compute_graph.flexml_layers[204].compute_node[0][2], compute_graph.flexml_layers[204].compute_node[0][3], compute_graph.flexml_layers[204].compute_node[1][0], compute_graph.flexml_layers[204].compute_node[1][1], compute_graph.flexml_layers[204].compute_node[1][2], compute_graph.flexml_layers[204].compute_node[1][3], compute_graph.flexml_layers[204].compute_node[2][0], compute_graph.flexml_layers[204].compute_node[2][1], compute_graph.flexml_layers[204].compute_node[2][2], compute_graph.flexml_layers[204].compute_node[2][3], compute_graph.flexml_layers[204].compute_node[3][0], compute_graph.flexml_layers[204].compute_node[3][1], compute_graph.flexml_layers[204].compute_node[3][2], compute_graph.flexml_layers[204].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[3], compute_graph.l2_244.in[0], compute_graph.l2_244.in[1], compute_graph.l2_244.in[2], compute_graph.l2_244.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1158): mode(0), layer(272): {compute_graph.ifm_ddr_3.out[2], compute_graph.ifm_ddr_3.out[3], compute_graph.L2_IFM_Buffer_for_input3_1_port1.in[0], compute_graph.L2_IFM_Buffer_for_input3_1_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1533), layer(272): compute_graph.flexml_layers[205].compute_node[0][0], compute_graph.flexml_layers[205].compute_node[0][1], compute_graph.flexml_layers[205].compute_node[0][2], compute_graph.flexml_layers[205].compute_node[0][3], compute_graph.flexml_layers[205].compute_node[1][0], compute_graph.flexml_layers[205].compute_node[1][1], compute_graph.flexml_layers[205].compute_node[1][2], compute_graph.flexml_layers[205].compute_node[1][3], compute_graph.flexml_layers[205].compute_node[2][0], compute_graph.flexml_layers[205].compute_node[2][1], compute_graph.flexml_layers[205].compute_node[2][2], compute_graph.flexml_layers[205].compute_node[2][3], compute_graph.flexml_layers[205].compute_node[3][0], compute_graph.flexml_layers[205].compute_node[3][1], compute_graph.flexml_layers[205].compute_node[3][2], compute_graph.flexml_layers[205].compute_node[3][3], {compute_graph.l2_244.out[0], compute_graph.l2_244.out[1], compute_graph.l2_244.out[2], compute_graph.l2_244.out[3], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[3], compute_graph.l2_245.in[0], compute_graph.l2_245.in[1], compute_graph.l2_245.in[2], compute_graph.l2_245.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(978): mode(0), layer(272): {compute_graph.l2_245.out[0], compute_graph.l2_245.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1162): mode(0), layer(273): {compute_graph.l2l3_242_spill.out[2], compute_graph.l2l3_242_spill.out[3], compute_graph.templated_graph_246.ifm.in[0], compute_graph.templated_graph_246.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1298): mode(0), layer(273): {compute_graph.templated_graph_246.ifm.out[0], compute_graph.templated_graph_246.ifm.out[1], compute_graph.l2l3_246_spill.in[0], compute_graph.l2l3_246_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1167): mode(0), layer(274): {compute_graph.l2l3_246_spill.out[0], compute_graph.l2l3_246_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1159): mode(0), layer(274): {compute_graph.ifm_ddr_3.out[4], compute_graph.ifm_ddr_3.out[5], compute_graph.L2_IFM_Buffer_for_input3_2_port1.in[0], compute_graph.L2_IFM_Buffer_for_input3_2_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1534), layer(274): compute_graph.flexml_layers[206].compute_node[0][0], compute_graph.flexml_layers[206].compute_node[0][1], compute_graph.flexml_layers[206].compute_node[0][2], compute_graph.flexml_layers[206].compute_node[0][3], compute_graph.flexml_layers[206].compute_node[1][0], compute_graph.flexml_layers[206].compute_node[1][1], compute_graph.flexml_layers[206].compute_node[1][2], compute_graph.flexml_layers[206].compute_node[1][3], compute_graph.flexml_layers[206].compute_node[2][0], compute_graph.flexml_layers[206].compute_node[2][1], compute_graph.flexml_layers[206].compute_node[2][2], compute_graph.flexml_layers[206].compute_node[2][3], compute_graph.flexml_layers[206].compute_node[3][0], compute_graph.flexml_layers[206].compute_node[3][1], compute_graph.flexml_layers[206].compute_node[3][2], compute_graph.flexml_layers[206].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[0], compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[1], compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[2], compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[3], compute_graph.l2_247.in[0], compute_graph.l2_247.in[1], compute_graph.l2_247.in[2], compute_graph.l2_247.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(887): mode(3), layer(275): {compute_graph.Layer_249_wts_ddr.out[0], compute_graph.Layer_249_wts_ddr.out[1], compute_graph.Layer_249_l2_wts.in[0], compute_graph.Layer_249_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(887) is pipelined with KernelLayerNode(1534) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(979): mode(0), layer(274): {compute_graph.l2_247.out[0], compute_graph.l2_247.out[1], compute_graph.spill_L3_Concat_Buffer_layer_248.in[2], compute_graph.spill_L3_Concat_Buffer_layer_248.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1168): mode(0), layer(275): {compute_graph.spill_L3_Concat_Buffer_layer_248.out[0], compute_graph.spill_L3_Concat_Buffer_layer_248.out[1], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1535), layer(275): compute_graph.flexml_layers[207].compute_node[0][0], compute_graph.flexml_layers[207].compute_node[0][1], compute_graph.flexml_layers[207].compute_node[0][2], compute_graph.flexml_layers[207].compute_node[0][3], compute_graph.flexml_layers[207].compute_node[1][0], compute_graph.flexml_layers[207].compute_node[1][1], compute_graph.flexml_layers[207].compute_node[1][2], compute_graph.flexml_layers[207].compute_node[1][3], compute_graph.flexml_layers[207].compute_node[2][0], compute_graph.flexml_layers[207].compute_node[2][1], compute_graph.flexml_layers[207].compute_node[2][2], compute_graph.flexml_layers[207].compute_node[2][3], compute_graph.flexml_layers[207].compute_node[3][0], compute_graph.flexml_layers[207].compute_node[3][1], compute_graph.flexml_layers[207].compute_node[3][2], compute_graph.flexml_layers[207].compute_node[3][3], {compute_graph.Layer_249_l2_wts.out[0], compute_graph.Layer_249_l2_wts.out[1], compute_graph.Layer_249_l2_wts.out[2], compute_graph.Layer_249_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[3], compute_graph.l2_249.in[0], compute_graph.l2_249.in[1], compute_graph.l2_249.in[2], compute_graph.l2_249.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1536), layer(276): compute_graph.flexml_layers[208].compute_node[0][0], compute_graph.flexml_layers[208].compute_node[0][1], compute_graph.flexml_layers[208].compute_node[0][2], compute_graph.flexml_layers[208].compute_node[0][3], compute_graph.flexml_layers[208].compute_node[1][0], compute_graph.flexml_layers[208].compute_node[1][1], compute_graph.flexml_layers[208].compute_node[1][2], compute_graph.flexml_layers[208].compute_node[1][3], compute_graph.flexml_layers[208].compute_node[2][0], compute_graph.flexml_layers[208].compute_node[2][1], compute_graph.flexml_layers[208].compute_node[2][2], compute_graph.flexml_layers[208].compute_node[2][3], compute_graph.flexml_layers[208].compute_node[3][0], compute_graph.flexml_layers[208].compute_node[3][1], compute_graph.flexml_layers[208].compute_node[3][2], compute_graph.flexml_layers[208].compute_node[3][3], {compute_graph.l2_249.out[0], compute_graph.l2_249.out[1], compute_graph.l2_249.out[2], compute_graph.l2_249.out[3], compute_graph.l2_250.in[0], compute_graph.l2_250.in[1], compute_graph.l2_250.in[2], compute_graph.l2_250.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1164): mode(0), layer(277): {compute_graph.l2l3_243_spill.out[2], compute_graph.l2l3_243_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1537), layer(277): compute_graph.flexml_layers[209].compute_node[0][0], compute_graph.flexml_layers[209].compute_node[0][1], compute_graph.flexml_layers[209].compute_node[0][2], compute_graph.flexml_layers[209].compute_node[0][3], compute_graph.flexml_layers[209].compute_node[1][0], compute_graph.flexml_layers[209].compute_node[1][1], compute_graph.flexml_layers[209].compute_node[1][2], compute_graph.flexml_layers[209].compute_node[1][3], compute_graph.flexml_layers[209].compute_node[2][0], compute_graph.flexml_layers[209].compute_node[2][1], compute_graph.flexml_layers[209].compute_node[2][2], compute_graph.flexml_layers[209].compute_node[2][3], compute_graph.flexml_layers[209].compute_node[3][0], compute_graph.flexml_layers[209].compute_node[3][1], compute_graph.flexml_layers[209].compute_node[3][2], compute_graph.flexml_layers[209].compute_node[3][3], {compute_graph.l2_250.out[0], compute_graph.l2_250.out[1], compute_graph.l2_250.out[2], compute_graph.l2_250.out[3], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[3], compute_graph.l2_251.in[0], compute_graph.l2_251.in[1], compute_graph.l2_251.in[2], compute_graph.l2_251.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1166): mode(0), layer(278): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1538), layer(278): compute_graph.flexml_layers[210].compute_node[0][0], compute_graph.flexml_layers[210].compute_node[0][1], compute_graph.flexml_layers[210].compute_node[0][2], compute_graph.flexml_layers[210].compute_node[0][3], compute_graph.flexml_layers[210].compute_node[1][0], compute_graph.flexml_layers[210].compute_node[1][1], compute_graph.flexml_layers[210].compute_node[1][2], compute_graph.flexml_layers[210].compute_node[1][3], compute_graph.flexml_layers[210].compute_node[2][0], compute_graph.flexml_layers[210].compute_node[2][1], compute_graph.flexml_layers[210].compute_node[2][2], compute_graph.flexml_layers[210].compute_node[2][3], compute_graph.flexml_layers[210].compute_node[3][0], compute_graph.flexml_layers[210].compute_node[3][1], compute_graph.flexml_layers[210].compute_node[3][2], compute_graph.flexml_layers[210].compute_node[3][3], {compute_graph.l2_251.out[0], compute_graph.l2_251.out[1], compute_graph.l2_251.out[2], compute_graph.l2_251.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[3], compute_graph.l2_252.in[0], compute_graph.l2_252.in[1], compute_graph.l2_252.in[2], compute_graph.l2_252.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(981): mode(0), layer(278): {compute_graph.l2_252.out[0], compute_graph.l2_252.out[1], compute_graph.ofm_ddr_1_l2l3_252_spill.in[0], compute_graph.ofm_ddr_1_l2l3_252_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(982): mode(0), layer(278): {compute_graph.l2_252.out[2], compute_graph.l2_252.out[3], compute_graph.spill_L3_Concat_Buffer_layer_253.in[2], compute_graph.spill_L3_Concat_Buffer_layer_253.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1154): mode(0), layer(279): {compute_graph.l2l3_238_spill.out[0], compute_graph.l2l3_238_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(983): mode(0), layer(279): {compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_253.in[0], compute_graph.spill_L3_Concat_Buffer_layer_253.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1169): mode(0), layer(280): {compute_graph.spill_L3_Concat_Buffer_layer_253.out[0], compute_graph.spill_L3_Concat_Buffer_layer_253.out[1], compute_graph.templated_graph_254.ifm_mem.in[0], compute_graph.templated_graph_254.ifm_mem.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.templated_graph_254.ofm_mem.in[0], compute_graph.templated_graph_254.ofm_mem.in[1], compute_graph.templated_graph_254.ofm_mem.in[2], compute_graph.templated_graph_254.ofm_mem.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1300), layer(280): compute_graph.templated_graph_254.mk[0][0], compute_graph.templated_graph_254.mk[0][1], compute_graph.templated_graph_254.mk[0][2], compute_graph.templated_graph_254.mk[0][3], compute_graph.templated_graph_254.mk[1][0], compute_graph.templated_graph_254.mk[1][1], compute_graph.templated_graph_254.mk[1][2], compute_graph.templated_graph_254.mk[1][3], compute_graph.templated_graph_254.mk[2][0], compute_graph.templated_graph_254.mk[2][1], compute_graph.templated_graph_254.mk[2][2], compute_graph.templated_graph_254.mk[2][3], compute_graph.templated_graph_254.mk[3][0], compute_graph.templated_graph_254.mk[3][1], compute_graph.templated_graph_254.mk[3][2], compute_graph.templated_graph_254.mk[3][3], {compute_graph.templated_graph_254.ifm_mem.out[0], compute_graph.templated_graph_254.ifm_mem.out[1], compute_graph.templated_graph_254.ifm_mem.out[2], compute_graph.templated_graph_254.ifm_mem.out[3], compute_graph.templated_graph_254.ofm_mem.in[0], compute_graph.templated_graph_254.ofm_mem.in[1], compute_graph.templated_graph_254.ofm_mem.in[2], compute_graph.templated_graph_254.ofm_mem.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1299): mode(3), layer(280): {compute_graph.templated_graph_254.ofm_mem.out[0], compute_graph.templated_graph_254.ofm_mem.out[1], compute_graph.l2l3_254_spill.in[0], compute_graph.l2l3_254_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1299) is pipelined with KernelLayerNode(1300) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1170): mode(0), layer(281): {compute_graph.l2l3_254_spill.out[0], compute_graph.l2l3_254_spill.out[1], compute_graph.templated_graph_255.ifm.in[0], compute_graph.templated_graph_255.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1301): mode(0), layer(281): {compute_graph.templated_graph_255.ifm.out[0], compute_graph.templated_graph_255.ifm.out[1], compute_graph.l2l3_255_spill.in[0], compute_graph.l2l3_255_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1171): mode(0), layer(282): {compute_graph.l2l3_255_spill.out[0], compute_graph.l2l3_255_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(984): mode(0), layer(282): {compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_256.in[0], compute_graph.spill_L3_Concat_Buffer_layer_256.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1172): mode(0), layer(283): {compute_graph.spill_L3_Concat_Buffer_layer_256.out[0], compute_graph.spill_L3_Concat_Buffer_layer_256.out[1], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(888): mode(4), layer(283): {compute_graph.Layer_257_wts_ddr.out[0], compute_graph.Layer_257_wts_ddr.out[1], compute_graph.Layer_257_l2_wts.in[0], compute_graph.Layer_257_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(888) is pipelined with BufferToBufferNode(1172) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1539), layer(283): compute_graph.flexml_layers[211].compute_node[0][0], compute_graph.flexml_layers[211].compute_node[0][1], compute_graph.flexml_layers[211].compute_node[0][2], compute_graph.flexml_layers[211].compute_node[0][3], compute_graph.flexml_layers[211].compute_node[1][0], compute_graph.flexml_layers[211].compute_node[1][1], compute_graph.flexml_layers[211].compute_node[1][2], compute_graph.flexml_layers[211].compute_node[1][3], compute_graph.flexml_layers[211].compute_node[2][0], compute_graph.flexml_layers[211].compute_node[2][1], compute_graph.flexml_layers[211].compute_node[2][2], compute_graph.flexml_layers[211].compute_node[2][3], compute_graph.flexml_layers[211].compute_node[3][0], compute_graph.flexml_layers[211].compute_node[3][1], compute_graph.flexml_layers[211].compute_node[3][2], compute_graph.flexml_layers[211].compute_node[3][3], {compute_graph.Layer_257_l2_wts.out[0], compute_graph.Layer_257_l2_wts.out[1], compute_graph.Layer_257_l2_wts.out[2], compute_graph.Layer_257_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[3], compute_graph.l2_257.in[0], compute_graph.l2_257.in[1], compute_graph.l2_257.in[2], compute_graph.l2_257.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(985): mode(0), layer(283): {compute_graph.l2_257.out[0], compute_graph.l2_257.out[1], compute_graph.l2l3_257_spill.in[0], compute_graph.l2l3_257_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1173): mode(0), layer(284): {compute_graph.l2l3_257_spill.out[0], compute_graph.l2l3_257_spill.out[1], compute_graph.templated_graph_258.ifm.in[0], compute_graph.templated_graph_258.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1303), layer(284): compute_graph.templated_graph_258.compute_node[0][0], compute_graph.templated_graph_258.compute_node[0][1], compute_graph.templated_graph_258.compute_node[0][2], compute_graph.templated_graph_258.compute_node[0][3], compute_graph.templated_graph_258.compute_node[1][0], compute_graph.templated_graph_258.compute_node[1][1], compute_graph.templated_graph_258.compute_node[1][2], compute_graph.templated_graph_258.compute_node[1][3], compute_graph.templated_graph_258.compute_node[2][0], compute_graph.templated_graph_258.compute_node[2][1], compute_graph.templated_graph_258.compute_node[2][2], compute_graph.templated_graph_258.compute_node[2][3], compute_graph.templated_graph_258.compute_node[3][0], compute_graph.templated_graph_258.compute_node[3][1], compute_graph.templated_graph_258.compute_node[3][2], compute_graph.templated_graph_258.compute_node[3][3], {compute_graph.templated_graph_258.ifm.out[0], compute_graph.templated_graph_258.ifm.out[1], compute_graph.templated_graph_258.ifm.out[2], compute_graph.templated_graph_258.ifm.out[3], compute_graph.templated_graph_258.ofm.in[0], compute_graph.templated_graph_258.ofm.in[1], compute_graph.templated_graph_258.ofm.in[2], compute_graph.templated_graph_258.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1302): mode(3), layer(284): {compute_graph.templated_graph_258.ofm.out[0], compute_graph.templated_graph_258.ofm.out[1], compute_graph.l2l3_258_spill.in[0], compute_graph.l2l3_258_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1302) is pipelined with KernelLayerNode(1303) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1174): mode(0), layer(285): {compute_graph.l2l3_257_spill.out[2], compute_graph.l2l3_257_spill.out[3], compute_graph.templated_graph_259.ifm.in[0], compute_graph.templated_graph_259.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1305), layer(285): compute_graph.templated_graph_259.compute_node[0][0], compute_graph.templated_graph_259.compute_node[0][1], compute_graph.templated_graph_259.compute_node[0][2], compute_graph.templated_graph_259.compute_node[0][3], compute_graph.templated_graph_259.compute_node[1][0], compute_graph.templated_graph_259.compute_node[1][1], compute_graph.templated_graph_259.compute_node[1][2], compute_graph.templated_graph_259.compute_node[1][3], compute_graph.templated_graph_259.compute_node[2][0], compute_graph.templated_graph_259.compute_node[2][1], compute_graph.templated_graph_259.compute_node[2][2], compute_graph.templated_graph_259.compute_node[2][3], compute_graph.templated_graph_259.compute_node[3][0], compute_graph.templated_graph_259.compute_node[3][1], compute_graph.templated_graph_259.compute_node[3][2], compute_graph.templated_graph_259.compute_node[3][3], {compute_graph.templated_graph_259.ifm.out[0], compute_graph.templated_graph_259.ifm.out[1], compute_graph.templated_graph_259.ifm.out[2], compute_graph.templated_graph_259.ifm.out[3], compute_graph.templated_graph_259.ofm.in[0], compute_graph.templated_graph_259.ofm.in[1], compute_graph.templated_graph_259.ofm.in[2], compute_graph.templated_graph_259.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1304): mode(3), layer(285): {compute_graph.templated_graph_259.ofm.out[0], compute_graph.templated_graph_259.ofm.out[1], compute_graph.l2l3_259_spill.in[0], compute_graph.l2l3_259_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1304) is pipelined with KernelLayerNode(1305) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1178): mode(3), layer(286): {compute_graph.ifm_ddr_2.out[0], compute_graph.templated_graph_260.wts.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1178) is pipelined with KernelLayerNode(1305) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1176): mode(0), layer(286): {compute_graph.l2l3_259_spill.out[0], compute_graph.templated_graph_260.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1307), layer(286): compute_graph.templated_graph_260.mk[0][0], compute_graph.templated_graph_260.mk[0][1], compute_graph.templated_graph_260.mk[0][2], compute_graph.templated_graph_260.mk[0][3], compute_graph.templated_graph_260.mk[1][0], compute_graph.templated_graph_260.mk[1][1], compute_graph.templated_graph_260.mk[1][2], compute_graph.templated_graph_260.mk[1][3], compute_graph.templated_graph_260.mk[2][0], compute_graph.templated_graph_260.mk[2][1], compute_graph.templated_graph_260.mk[2][2], compute_graph.templated_graph_260.mk[2][3], compute_graph.templated_graph_260.mk[3][0], compute_graph.templated_graph_260.mk[3][1], compute_graph.templated_graph_260.mk[3][2], compute_graph.templated_graph_260.mk[3][3], {compute_graph.templated_graph_260.ifm.out[0], compute_graph.templated_graph_260.ifm.out[1], compute_graph.templated_graph_260.ifm.out[2], compute_graph.templated_graph_260.ifm.out[3], compute_graph.templated_graph_260.wts.out[0], compute_graph.templated_graph_260.wts.out[1], compute_graph.templated_graph_260.wts.out[2], compute_graph.templated_graph_260.wts.out[3], compute_graph.templated_graph_260.ofm.in[0], compute_graph.templated_graph_260.ofm.in[1], compute_graph.templated_graph_260.ofm.in[2], compute_graph.templated_graph_260.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1306): mode(3), layer(286): {compute_graph.templated_graph_260.ofm.out[0], compute_graph.l2l3_260_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1306) is pipelined with KernelLayerNode(1307) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(889): mode(3), layer(287): {compute_graph.Layer_261_wts_ddr.out[0], compute_graph.Layer_261_wts_ddr.out[1], compute_graph.Layer_261_l2_wts.in[0], compute_graph.Layer_261_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(889) is pipelined with KernelLayerNode(1307) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1181): mode(0), layer(287): {compute_graph.l2l3_260_spill.out[0], compute_graph.l2l3_260_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1540), layer(287): compute_graph.flexml_layers[212].compute_node[0][0], compute_graph.flexml_layers[212].compute_node[0][1], compute_graph.flexml_layers[212].compute_node[0][2], compute_graph.flexml_layers[212].compute_node[0][3], compute_graph.flexml_layers[212].compute_node[1][0], compute_graph.flexml_layers[212].compute_node[1][1], compute_graph.flexml_layers[212].compute_node[1][2], compute_graph.flexml_layers[212].compute_node[1][3], compute_graph.flexml_layers[212].compute_node[2][0], compute_graph.flexml_layers[212].compute_node[2][1], compute_graph.flexml_layers[212].compute_node[2][2], compute_graph.flexml_layers[212].compute_node[2][3], compute_graph.flexml_layers[212].compute_node[3][0], compute_graph.flexml_layers[212].compute_node[3][1], compute_graph.flexml_layers[212].compute_node[3][2], compute_graph.flexml_layers[212].compute_node[3][3], {compute_graph.Layer_261_l2_wts.out[0], compute_graph.Layer_261_l2_wts.out[1], compute_graph.Layer_261_l2_wts.out[2], compute_graph.Layer_261_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[3], compute_graph.l2_261.in[0], compute_graph.l2_261.in[1], compute_graph.l2_261.in[2], compute_graph.l2_261.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2_261.out[0], compute_graph.l2_261.out[1], compute_graph.l2_261.out[2], compute_graph.l2_261.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1541), layer(288): compute_graph.flexml_layers[213].compute_node[0][0], compute_graph.flexml_layers[213].compute_node[0][1], compute_graph.flexml_layers[213].compute_node[0][2], compute_graph.flexml_layers[213].compute_node[0][3], compute_graph.flexml_layers[213].compute_node[1][0], compute_graph.flexml_layers[213].compute_node[1][1], compute_graph.flexml_layers[213].compute_node[1][2], compute_graph.flexml_layers[213].compute_node[1][3], compute_graph.flexml_layers[213].compute_node[2][0], compute_graph.flexml_layers[213].compute_node[2][1], compute_graph.flexml_layers[213].compute_node[2][2], compute_graph.flexml_layers[213].compute_node[2][3], compute_graph.flexml_layers[213].compute_node[3][0], compute_graph.flexml_layers[213].compute_node[3][1], compute_graph.flexml_layers[213].compute_node[3][2], compute_graph.flexml_layers[213].compute_node[3][3], {compute_graph.l2_261.out[0], compute_graph.l2_261.out[1], compute_graph.l2_261.out[2], compute_graph.l2_261.out[3], compute_graph.l2_262.in[0], compute_graph.l2_262.in[1], compute_graph.l2_262.in[2], compute_graph.l2_262.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(986): mode(0), layer(288): {compute_graph.l2_262.out[0], compute_graph.l2_262.out[1], compute_graph.l2l3_262_spill.in[0], compute_graph.l2l3_262_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1182): mode(0), layer(289): {compute_graph.l2l3_262_spill.out[0], compute_graph.l2l3_262_spill.out[1], compute_graph.templated_graph_263.ifm.in[0], compute_graph.templated_graph_263.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1309), layer(289): compute_graph.templated_graph_263.compute_node[0][0], compute_graph.templated_graph_263.compute_node[0][1], compute_graph.templated_graph_263.compute_node[0][2], compute_graph.templated_graph_263.compute_node[0][3], compute_graph.templated_graph_263.compute_node[1][0], compute_graph.templated_graph_263.compute_node[1][1], compute_graph.templated_graph_263.compute_node[1][2], compute_graph.templated_graph_263.compute_node[1][3], compute_graph.templated_graph_263.compute_node[2][0], compute_graph.templated_graph_263.compute_node[2][1], compute_graph.templated_graph_263.compute_node[2][2], compute_graph.templated_graph_263.compute_node[2][3], compute_graph.templated_graph_263.compute_node[3][0], compute_graph.templated_graph_263.compute_node[3][1], compute_graph.templated_graph_263.compute_node[3][2], compute_graph.templated_graph_263.compute_node[3][3], {compute_graph.templated_graph_263.ifm.out[0], compute_graph.templated_graph_263.ifm.out[1], compute_graph.templated_graph_263.ifm.out[2], compute_graph.templated_graph_263.ifm.out[3], compute_graph.templated_graph_263.ofm.in[0], compute_graph.templated_graph_263.ofm.in[1], compute_graph.templated_graph_263.ofm.in[2], compute_graph.templated_graph_263.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1308): mode(3), layer(289): {compute_graph.templated_graph_263.ofm.out[0], compute_graph.templated_graph_263.ofm.out[1], compute_graph.l2l3_263_spill.in[0], compute_graph.l2l3_263_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1308) is pipelined with KernelLayerNode(1309) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1184): mode(0), layer(290): {compute_graph.l2l3_263_spill.out[0], compute_graph.l2l3_263_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1186): mode(0), layer(290): {compute_graph.const_ifm_ddr_1.out[0], compute_graph.const_ifm_ddr_1.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1542), layer(290): compute_graph.flexml_layers[214].compute_node[0][0], compute_graph.flexml_layers[214].compute_node[0][1], compute_graph.flexml_layers[214].compute_node[0][2], compute_graph.flexml_layers[214].compute_node[0][3], compute_graph.flexml_layers[214].compute_node[1][0], compute_graph.flexml_layers[214].compute_node[1][1], compute_graph.flexml_layers[214].compute_node[1][2], compute_graph.flexml_layers[214].compute_node[1][3], compute_graph.flexml_layers[214].compute_node[2][0], compute_graph.flexml_layers[214].compute_node[2][1], compute_graph.flexml_layers[214].compute_node[2][2], compute_graph.flexml_layers[214].compute_node[2][3], compute_graph.flexml_layers[214].compute_node[3][0], compute_graph.flexml_layers[214].compute_node[3][1], compute_graph.flexml_layers[214].compute_node[3][2], compute_graph.flexml_layers[214].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[3], compute_graph.l2_264.in[0], compute_graph.l2_264.in[1], compute_graph.l2_264.in[2], compute_graph.l2_264.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1179): mode(0), layer(291): {compute_graph.ifm_ddr_2.out[1], compute_graph.ifm_ddr_2.out[2], compute_graph.L2_IFM_Buffer_for_input2_1_port1.in[0], compute_graph.L2_IFM_Buffer_for_input2_1_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1543), layer(291): compute_graph.flexml_layers[215].compute_node[0][0], compute_graph.flexml_layers[215].compute_node[0][1], compute_graph.flexml_layers[215].compute_node[0][2], compute_graph.flexml_layers[215].compute_node[0][3], compute_graph.flexml_layers[215].compute_node[1][0], compute_graph.flexml_layers[215].compute_node[1][1], compute_graph.flexml_layers[215].compute_node[1][2], compute_graph.flexml_layers[215].compute_node[1][3], compute_graph.flexml_layers[215].compute_node[2][0], compute_graph.flexml_layers[215].compute_node[2][1], compute_graph.flexml_layers[215].compute_node[2][2], compute_graph.flexml_layers[215].compute_node[2][3], compute_graph.flexml_layers[215].compute_node[3][0], compute_graph.flexml_layers[215].compute_node[3][1], compute_graph.flexml_layers[215].compute_node[3][2], compute_graph.flexml_layers[215].compute_node[3][3], {compute_graph.l2_264.out[0], compute_graph.l2_264.out[1], compute_graph.l2_264.out[2], compute_graph.l2_264.out[3], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[3], compute_graph.l2_265.in[0], compute_graph.l2_265.in[1], compute_graph.l2_265.in[2], compute_graph.l2_265.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(987): mode(0), layer(291): {compute_graph.l2_265.out[0], compute_graph.l2_265.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1183): mode(0), layer(292): {compute_graph.l2l3_262_spill.out[2], compute_graph.l2l3_262_spill.out[3], compute_graph.templated_graph_266.ifm.in[0], compute_graph.templated_graph_266.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1311), layer(292): compute_graph.templated_graph_266.compute_node[0][0], compute_graph.templated_graph_266.compute_node[0][1], compute_graph.templated_graph_266.compute_node[0][2], compute_graph.templated_graph_266.compute_node[0][3], compute_graph.templated_graph_266.compute_node[1][0], compute_graph.templated_graph_266.compute_node[1][1], compute_graph.templated_graph_266.compute_node[1][2], compute_graph.templated_graph_266.compute_node[1][3], compute_graph.templated_graph_266.compute_node[2][0], compute_graph.templated_graph_266.compute_node[2][1], compute_graph.templated_graph_266.compute_node[2][2], compute_graph.templated_graph_266.compute_node[2][3], compute_graph.templated_graph_266.compute_node[3][0], compute_graph.templated_graph_266.compute_node[3][1], compute_graph.templated_graph_266.compute_node[3][2], compute_graph.templated_graph_266.compute_node[3][3], {compute_graph.templated_graph_266.ifm.out[0], compute_graph.templated_graph_266.ifm.out[1], compute_graph.templated_graph_266.ifm.out[2], compute_graph.templated_graph_266.ifm.out[3], compute_graph.templated_graph_266.ofm.in[0], compute_graph.templated_graph_266.ofm.in[1], compute_graph.templated_graph_266.ofm.in[2], compute_graph.templated_graph_266.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1310): mode(3), layer(292): {compute_graph.templated_graph_266.ofm.out[0], compute_graph.templated_graph_266.ofm.out[1], compute_graph.l2l3_266_spill.in[0], compute_graph.l2l3_266_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1310) is pipelined with KernelLayerNode(1311) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1188): mode(0), layer(293): {compute_graph.l2l3_266_spill.out[0], compute_graph.l2l3_266_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1180): mode(0), layer(293): {compute_graph.ifm_ddr_2.out[3], compute_graph.ifm_ddr_2.out[4], compute_graph.L2_IFM_Buffer_for_input2_2_port1.in[0], compute_graph.L2_IFM_Buffer_for_input2_2_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1544), layer(293): compute_graph.flexml_layers[216].compute_node[0][0], compute_graph.flexml_layers[216].compute_node[0][1], compute_graph.flexml_layers[216].compute_node[0][2], compute_graph.flexml_layers[216].compute_node[0][3], compute_graph.flexml_layers[216].compute_node[1][0], compute_graph.flexml_layers[216].compute_node[1][1], compute_graph.flexml_layers[216].compute_node[1][2], compute_graph.flexml_layers[216].compute_node[1][3], compute_graph.flexml_layers[216].compute_node[2][0], compute_graph.flexml_layers[216].compute_node[2][1], compute_graph.flexml_layers[216].compute_node[2][2], compute_graph.flexml_layers[216].compute_node[2][3], compute_graph.flexml_layers[216].compute_node[3][0], compute_graph.flexml_layers[216].compute_node[3][1], compute_graph.flexml_layers[216].compute_node[3][2], compute_graph.flexml_layers[216].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[0], compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[1], compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[2], compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[3], compute_graph.l2_267.in[0], compute_graph.l2_267.in[1], compute_graph.l2_267.in[2], compute_graph.l2_267.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(988): mode(0), layer(293): {compute_graph.l2_267.out[0], compute_graph.l2_267.out[1], compute_graph.l2l3_267_spill.in[0], compute_graph.l2l3_267_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1177): mode(0), layer(294): {compute_graph.l2l3_259_spill.out[1], compute_graph.templated_graph_268.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1189): mode(4), layer(294): {compute_graph.l2l3_267_spill.out[0], compute_graph.templated_graph_268.wts.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(1189) is pipelined with BufferToBufferNode(1177) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1313), layer(294): compute_graph.templated_graph_268.mk[0][0], compute_graph.templated_graph_268.mk[0][1], compute_graph.templated_graph_268.mk[0][2], compute_graph.templated_graph_268.mk[0][3], compute_graph.templated_graph_268.mk[1][0], compute_graph.templated_graph_268.mk[1][1], compute_graph.templated_graph_268.mk[1][2], compute_graph.templated_graph_268.mk[1][3], compute_graph.templated_graph_268.mk[2][0], compute_graph.templated_graph_268.mk[2][1], compute_graph.templated_graph_268.mk[2][2], compute_graph.templated_graph_268.mk[2][3], compute_graph.templated_graph_268.mk[3][0], compute_graph.templated_graph_268.mk[3][1], compute_graph.templated_graph_268.mk[3][2], compute_graph.templated_graph_268.mk[3][3], {compute_graph.templated_graph_268.ifm.out[0], compute_graph.templated_graph_268.ifm.out[1], compute_graph.templated_graph_268.ifm.out[2], compute_graph.templated_graph_268.ifm.out[3], compute_graph.templated_graph_268.wts.out[0], compute_graph.templated_graph_268.wts.out[1], compute_graph.templated_graph_268.wts.out[2], compute_graph.templated_graph_268.wts.out[3], compute_graph.templated_graph_268.ofm.in[0], compute_graph.templated_graph_268.ofm.in[1], compute_graph.templated_graph_268.ofm.in[2], compute_graph.templated_graph_268.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1312): mode(3), layer(294): {compute_graph.templated_graph_268.ofm.out[0], compute_graph.l2l3_268_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1312) is pipelined with KernelLayerNode(1313) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1190): mode(0), layer(295): {compute_graph.l2l3_268_spill.out[0], compute_graph.l2l3_268_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(890): mode(4), layer(295): {compute_graph.Layer_269_wts_ddr.out[0], compute_graph.Layer_269_wts_ddr.out[1], compute_graph.Layer_269_l2_wts.in[0], compute_graph.Layer_269_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(890) is pipelined with BufferToBufferNode(1190) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1545), layer(295): compute_graph.flexml_layers[217].compute_node[0][0], compute_graph.flexml_layers[217].compute_node[0][1], compute_graph.flexml_layers[217].compute_node[0][2], compute_graph.flexml_layers[217].compute_node[0][3], compute_graph.flexml_layers[217].compute_node[1][0], compute_graph.flexml_layers[217].compute_node[1][1], compute_graph.flexml_layers[217].compute_node[1][2], compute_graph.flexml_layers[217].compute_node[1][3], compute_graph.flexml_layers[217].compute_node[2][0], compute_graph.flexml_layers[217].compute_node[2][1], compute_graph.flexml_layers[217].compute_node[2][2], compute_graph.flexml_layers[217].compute_node[2][3], compute_graph.flexml_layers[217].compute_node[3][0], compute_graph.flexml_layers[217].compute_node[3][1], compute_graph.flexml_layers[217].compute_node[3][2], compute_graph.flexml_layers[217].compute_node[3][3], {compute_graph.Layer_269_l2_wts.out[0], compute_graph.Layer_269_l2_wts.out[1], compute_graph.Layer_269_l2_wts.out[2], compute_graph.Layer_269_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[3], compute_graph.l2_269.in[0], compute_graph.l2_269.in[1], compute_graph.l2_269.in[2], compute_graph.l2_269.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1546), layer(296): compute_graph.flexml_layers[218].compute_node[0][0], compute_graph.flexml_layers[218].compute_node[0][1], compute_graph.flexml_layers[218].compute_node[0][2], compute_graph.flexml_layers[218].compute_node[0][3], compute_graph.flexml_layers[218].compute_node[1][0], compute_graph.flexml_layers[218].compute_node[1][1], compute_graph.flexml_layers[218].compute_node[1][2], compute_graph.flexml_layers[218].compute_node[1][3], compute_graph.flexml_layers[218].compute_node[2][0], compute_graph.flexml_layers[218].compute_node[2][1], compute_graph.flexml_layers[218].compute_node[2][2], compute_graph.flexml_layers[218].compute_node[2][3], compute_graph.flexml_layers[218].compute_node[3][0], compute_graph.flexml_layers[218].compute_node[3][1], compute_graph.flexml_layers[218].compute_node[3][2], compute_graph.flexml_layers[218].compute_node[3][3], {compute_graph.l2_269.out[0], compute_graph.l2_269.out[1], compute_graph.l2_269.out[2], compute_graph.l2_269.out[3], compute_graph.l2_270.in[0], compute_graph.l2_270.in[1], compute_graph.l2_270.in[2], compute_graph.l2_270.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1185): mode(0), layer(297): {compute_graph.l2l3_263_spill.out[2], compute_graph.l2l3_263_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1547), layer(297): compute_graph.flexml_layers[219].compute_node[0][0], compute_graph.flexml_layers[219].compute_node[0][1], compute_graph.flexml_layers[219].compute_node[0][2], compute_graph.flexml_layers[219].compute_node[0][3], compute_graph.flexml_layers[219].compute_node[1][0], compute_graph.flexml_layers[219].compute_node[1][1], compute_graph.flexml_layers[219].compute_node[1][2], compute_graph.flexml_layers[219].compute_node[1][3], compute_graph.flexml_layers[219].compute_node[2][0], compute_graph.flexml_layers[219].compute_node[2][1], compute_graph.flexml_layers[219].compute_node[2][2], compute_graph.flexml_layers[219].compute_node[2][3], compute_graph.flexml_layers[219].compute_node[3][0], compute_graph.flexml_layers[219].compute_node[3][1], compute_graph.flexml_layers[219].compute_node[3][2], compute_graph.flexml_layers[219].compute_node[3][3], {compute_graph.l2_270.out[0], compute_graph.l2_270.out[1], compute_graph.l2_270.out[2], compute_graph.l2_270.out[3], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[3], compute_graph.l2_271.in[0], compute_graph.l2_271.in[1], compute_graph.l2_271.in[2], compute_graph.l2_271.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1187): mode(0), layer(298): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1548), layer(298): compute_graph.flexml_layers[220].compute_node[0][0], compute_graph.flexml_layers[220].compute_node[0][1], compute_graph.flexml_layers[220].compute_node[0][2], compute_graph.flexml_layers[220].compute_node[0][3], compute_graph.flexml_layers[220].compute_node[1][0], compute_graph.flexml_layers[220].compute_node[1][1], compute_graph.flexml_layers[220].compute_node[1][2], compute_graph.flexml_layers[220].compute_node[1][3], compute_graph.flexml_layers[220].compute_node[2][0], compute_graph.flexml_layers[220].compute_node[2][1], compute_graph.flexml_layers[220].compute_node[2][2], compute_graph.flexml_layers[220].compute_node[2][3], compute_graph.flexml_layers[220].compute_node[3][0], compute_graph.flexml_layers[220].compute_node[3][1], compute_graph.flexml_layers[220].compute_node[3][2], compute_graph.flexml_layers[220].compute_node[3][3], {compute_graph.l2_271.out[0], compute_graph.l2_271.out[1], compute_graph.l2_271.out[2], compute_graph.l2_271.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[3], compute_graph.l2_272.in[0], compute_graph.l2_272.in[1], compute_graph.l2_272.in[2], compute_graph.l2_272.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(989): mode(0), layer(298): {compute_graph.l2_272.out[0], compute_graph.l2_272.out[1], compute_graph.ofm_ddr_2_l2l3_272_spill.in[0], compute_graph.ofm_ddr_2_l2l3_272_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1175): mode(0), layer(299): {compute_graph.l2l3_258_spill.out[0], compute_graph.templated_graph_273.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1191): mode(4), layer(299): {compute_graph.ofm_ddr_2_l2l3_272_spill.out[0], compute_graph.templated_graph_273.wts.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(1191) is pipelined with BufferToBufferNode(1175) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1315), layer(299): compute_graph.templated_graph_273.mk[0][0], compute_graph.templated_graph_273.mk[0][1], compute_graph.templated_graph_273.mk[0][2], compute_graph.templated_graph_273.mk[0][3], compute_graph.templated_graph_273.mk[1][0], compute_graph.templated_graph_273.mk[1][1], compute_graph.templated_graph_273.mk[1][2], compute_graph.templated_graph_273.mk[1][3], compute_graph.templated_graph_273.mk[2][0], compute_graph.templated_graph_273.mk[2][1], compute_graph.templated_graph_273.mk[2][2], compute_graph.templated_graph_273.mk[2][3], compute_graph.templated_graph_273.mk[3][0], compute_graph.templated_graph_273.mk[3][1], compute_graph.templated_graph_273.mk[3][2], compute_graph.templated_graph_273.mk[3][3], {compute_graph.templated_graph_273.ifm.out[0], compute_graph.templated_graph_273.ifm.out[1], compute_graph.templated_graph_273.ifm.out[2], compute_graph.templated_graph_273.ifm.out[3], compute_graph.templated_graph_273.wts.out[0], compute_graph.templated_graph_273.wts.out[1], compute_graph.templated_graph_273.wts.out[2], compute_graph.templated_graph_273.wts.out[3], compute_graph.templated_graph_273.ofm.in[0], compute_graph.templated_graph_273.ofm.in[1], compute_graph.templated_graph_273.ofm.in[2], compute_graph.templated_graph_273.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1314): mode(3), layer(299): {compute_graph.templated_graph_273.ofm.out[0], compute_graph.l2l3_273_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1314) is pipelined with KernelLayerNode(1315) +INFO: [aiecompiler 77-23129] BufferToBufferNode 1316 will not be pipelined because it's in the same layer 300 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2l3_273_spill.out[0], compute_graph.l2l3_273_spill.out[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1192): mode(0), layer(300): {compute_graph.l2l3_273_spill.out[0], compute_graph.l2l3_273_spill.out[1], compute_graph.templated_graph_274.ifm_mem.in[0], compute_graph.templated_graph_274.ifm_mem.in[1]}, Scheduler computes number of sub-layer phases to be 12. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1317), layer(300): compute_graph.templated_graph_274.mk[0][0], compute_graph.templated_graph_274.mk[0][1], compute_graph.templated_graph_274.mk[0][2], compute_graph.templated_graph_274.mk[0][3], compute_graph.templated_graph_274.mk[1][0], compute_graph.templated_graph_274.mk[1][1], compute_graph.templated_graph_274.mk[1][2], compute_graph.templated_graph_274.mk[1][3], compute_graph.templated_graph_274.mk[2][0], compute_graph.templated_graph_274.mk[2][1], compute_graph.templated_graph_274.mk[2][2], compute_graph.templated_graph_274.mk[2][3], compute_graph.templated_graph_274.mk[3][0], compute_graph.templated_graph_274.mk[3][1], compute_graph.templated_graph_274.mk[3][2], compute_graph.templated_graph_274.mk[3][3], {compute_graph.templated_graph_274.ifm_mem.out[0], compute_graph.templated_graph_274.ifm_mem.out[1], compute_graph.templated_graph_274.ifm_mem.out[2], compute_graph.templated_graph_274.ifm_mem.out[3], compute_graph.templated_graph_274.ofm_mem.in[0], compute_graph.templated_graph_274.ofm_mem.in[1], compute_graph.templated_graph_274.ofm_mem.in[2], compute_graph.templated_graph_274.ofm_mem.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(891): mode(3), layer(301): {compute_graph.Layer_276_wts_ddr.out[0], compute_graph.Layer_276_wts_ddr.out[1], compute_graph.Layer_276_l2_wts.in[0], compute_graph.Layer_276_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(891) is pipelined with KernelLayerNode(1317) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.templated_graph_274.ofm_mem.out[0], compute_graph.templated_graph_274.ofm_mem.out[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1316): mode(3), layer(300): {compute_graph.templated_graph_274.ofm_mem.out[0], compute_graph.templated_graph_274.ofm_mem.out[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[1]}, Scheduler computes number of sub-layer phases to be 12. +INFO: [aiecompiler 77-23129] BufferToBufferNode 990 will not be pipelined because it's in the same layer 301 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.spill_L3_Concat_Buffer_layer_275.out[0], compute_graph.spill_L3_Concat_Buffer_layer_275.out[1], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1193): mode(0), layer(301): {compute_graph.spill_L3_Concat_Buffer_layer_275.out[0], compute_graph.spill_L3_Concat_Buffer_layer_275.out[1], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[1]}, Scheduler computes number of sub-layer phases to be 5. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1549), layer(301): compute_graph.flexml_layers[221].compute_node[0][0], compute_graph.flexml_layers[221].compute_node[0][1], compute_graph.flexml_layers[221].compute_node[0][2], compute_graph.flexml_layers[221].compute_node[0][3], compute_graph.flexml_layers[221].compute_node[1][0], compute_graph.flexml_layers[221].compute_node[1][1], compute_graph.flexml_layers[221].compute_node[1][2], compute_graph.flexml_layers[221].compute_node[1][3], compute_graph.flexml_layers[221].compute_node[2][0], compute_graph.flexml_layers[221].compute_node[2][1], compute_graph.flexml_layers[221].compute_node[2][2], compute_graph.flexml_layers[221].compute_node[2][3], compute_graph.flexml_layers[221].compute_node[3][0], compute_graph.flexml_layers[221].compute_node[3][1], compute_graph.flexml_layers[221].compute_node[3][2], compute_graph.flexml_layers[221].compute_node[3][3], {compute_graph.Layer_276_l2_wts.out[0], compute_graph.Layer_276_l2_wts.out[1], compute_graph.Layer_276_l2_wts.out[2], compute_graph.Layer_276_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[3], compute_graph.l2_276.in[0], compute_graph.l2_276.in[1], compute_graph.l2_276.in[2], compute_graph.l2_276.in[3]}, Scheduler computes number of sub-layer phases to be 5. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(990): mode(3), layer(301): {compute_graph.l2_276.out[0], compute_graph.l2_276.out[1], compute_graph.l2l3_276_spill.in[0], compute_graph.l2l3_276_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1194): mode(0), layer(302): {compute_graph.l2l3_276_spill.out[0], compute_graph.l2l3_276_spill.out[1], compute_graph.templated_graph_277.ifm.in[0], compute_graph.templated_graph_277.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1318): mode(0), layer(302): {compute_graph.templated_graph_277.ifm.out[0], compute_graph.templated_graph_277.ifm.out[1], compute_graph.l2l3_277_spill.in[0], compute_graph.l2l3_277_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1195): mode(0), layer(303): {compute_graph.l2l3_276_spill.out[2], compute_graph.l2l3_276_spill.out[3], compute_graph.templated_graph_278.ifm.in[0], compute_graph.templated_graph_278.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1319): mode(0), layer(303): {compute_graph.templated_graph_278.ifm.out[0], compute_graph.templated_graph_278.ifm.out[1], compute_graph.l2l3_278_spill.in[0], compute_graph.l2l3_278_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1198): mode(0), layer(304): {compute_graph.l2l3_278_spill.out[2], compute_graph.l2l3_278_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(998): mode(0), layer(304): {compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_287.in[0], compute_graph.spill_L3_Concat_Buffer_layer_287.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1197): mode(0), layer(305): {compute_graph.l2l3_278_spill.out[0], compute_graph.l2l3_278_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(991): mode(0), layer(305): {compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_279.in[0], compute_graph.spill_L3_Concat_Buffer_layer_279.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1202): mode(0), layer(306): {compute_graph.spill_L3_Concat_Buffer_layer_279.out[0], compute_graph.spill_L3_Concat_Buffer_layer_279.out[1], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(892): mode(4), layer(306): {compute_graph.Layer_280_wts_ddr.out[0], compute_graph.Layer_280_wts_ddr.out[1], compute_graph.Layer_280_l2_wts.in[0], compute_graph.Layer_280_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(892) is pipelined with BufferToBufferNode(1202) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[3], compute_graph.l2_280.in[0], compute_graph.l2_280.in[1], compute_graph.l2_280.in[2], compute_graph.l2_280.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1550), layer(306): compute_graph.flexml_layers[222].compute_node[0][0], compute_graph.flexml_layers[222].compute_node[0][1], compute_graph.flexml_layers[222].compute_node[0][2], compute_graph.flexml_layers[222].compute_node[0][3], compute_graph.flexml_layers[222].compute_node[1][0], compute_graph.flexml_layers[222].compute_node[1][1], compute_graph.flexml_layers[222].compute_node[1][2], compute_graph.flexml_layers[222].compute_node[1][3], compute_graph.flexml_layers[222].compute_node[2][0], compute_graph.flexml_layers[222].compute_node[2][1], compute_graph.flexml_layers[222].compute_node[2][2], compute_graph.flexml_layers[222].compute_node[2][3], compute_graph.flexml_layers[222].compute_node[3][0], compute_graph.flexml_layers[222].compute_node[3][1], compute_graph.flexml_layers[222].compute_node[3][2], compute_graph.flexml_layers[222].compute_node[3][3], {compute_graph.Layer_280_l2_wts.out[0], compute_graph.Layer_280_l2_wts.out[1], compute_graph.Layer_280_l2_wts.out[2], compute_graph.Layer_280_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[3], compute_graph.l2_280.in[0], compute_graph.l2_280.in[1], compute_graph.l2_280.in[2], compute_graph.l2_280.in[3]}, Scheduler computes number of sub-layer phases to be 6. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(993): mode(3), layer(306): {compute_graph.l2_280.out[0], compute_graph.l2_280.out[1], compute_graph.l2l3_280_spill.in[0], compute_graph.l2l3_280_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(993) is pipelined with KernelLayerNode(1550) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1203): mode(0), layer(307): {compute_graph.l2l3_280_spill.out[0], compute_graph.l2l3_280_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1551), layer(307): compute_graph.flexml_layers[223].compute_node[0][0], compute_graph.flexml_layers[223].compute_node[0][1], compute_graph.flexml_layers[223].compute_node[0][2], compute_graph.flexml_layers[223].compute_node[0][3], compute_graph.flexml_layers[223].compute_node[1][0], compute_graph.flexml_layers[223].compute_node[1][1], compute_graph.flexml_layers[223].compute_node[1][2], compute_graph.flexml_layers[223].compute_node[1][3], compute_graph.flexml_layers[223].compute_node[2][0], compute_graph.flexml_layers[223].compute_node[2][1], compute_graph.flexml_layers[223].compute_node[2][2], compute_graph.flexml_layers[223].compute_node[2][3], compute_graph.flexml_layers[223].compute_node[3][0], compute_graph.flexml_layers[223].compute_node[3][1], compute_graph.flexml_layers[223].compute_node[3][2], compute_graph.flexml_layers[223].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[3], compute_graph.l2_281.in[0], compute_graph.l2_281.in[1], compute_graph.l2_281.in[2], compute_graph.l2_281.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(994): mode(3), layer(307): {compute_graph.l2_281.out[0], compute_graph.l2_281.out[1], compute_graph.l2l3_281_spill.in[0], compute_graph.l2l3_281_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(994) is pipelined with KernelLayerNode(1551) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1204): mode(0), layer(308): {compute_graph.l2l3_281_spill.out[0], compute_graph.l2l3_281_spill.out[1], compute_graph.templated_graph_282.ifm.in[0], compute_graph.templated_graph_282.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1320): mode(0), layer(308): {compute_graph.templated_graph_282.ifm.out[0], compute_graph.templated_graph_282.ifm.out[1], compute_graph.l2l3_282_spill.in[0], compute_graph.l2l3_282_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 995 will not be pipelined because it's in the same layer 309 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1613), layer(309): {compute_graph.l2l3_282_spill.out[0], compute_graph.const_ifm_ddr.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1614), layer(309): {compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1615), layer(309): {compute_graph.l2l3_282_spill.out[1], compute_graph.const_ifm_ddr.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1616), layer(309): {compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1552), layer(309): compute_graph.flexml_layers[224].compute_node[0][0], compute_graph.flexml_layers[224].compute_node[0][1], compute_graph.flexml_layers[224].compute_node[0][2], compute_graph.flexml_layers[224].compute_node[0][3], compute_graph.flexml_layers[224].compute_node[1][0], compute_graph.flexml_layers[224].compute_node[1][1], compute_graph.flexml_layers[224].compute_node[1][2], compute_graph.flexml_layers[224].compute_node[1][3], compute_graph.flexml_layers[224].compute_node[2][0], compute_graph.flexml_layers[224].compute_node[2][1], compute_graph.flexml_layers[224].compute_node[2][2], compute_graph.flexml_layers[224].compute_node[2][3], compute_graph.flexml_layers[224].compute_node[3][0], compute_graph.flexml_layers[224].compute_node[3][1], compute_graph.flexml_layers[224].compute_node[3][2], compute_graph.flexml_layers[224].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[3], compute_graph.l2_283.in[0], compute_graph.l2_283.in[1], compute_graph.l2_283.in[2], compute_graph.l2_283.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(995): mode(3), layer(309): {compute_graph.l2_283.out[0], compute_graph.l2_283.out[1], compute_graph.l2l3_283_spill.in[0], compute_graph.l2l3_283_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 996 will not be pipelined because it's in the same layer 310 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1617), layer(310): {compute_graph.ifm_ddr_1.out[2], compute_graph.l2l3_283_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1618), layer(310): {compute_graph.L2_IFM_Buffer_for_input1_1_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1619), layer(310): {compute_graph.ifm_ddr_1.out[3], compute_graph.l2l3_283_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1620), layer(310): {compute_graph.L2_IFM_Buffer_for_input1_1_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1553), layer(310): compute_graph.flexml_layers[225].compute_node[0][0], compute_graph.flexml_layers[225].compute_node[0][1], compute_graph.flexml_layers[225].compute_node[0][2], compute_graph.flexml_layers[225].compute_node[0][3], compute_graph.flexml_layers[225].compute_node[1][0], compute_graph.flexml_layers[225].compute_node[1][1], compute_graph.flexml_layers[225].compute_node[1][2], compute_graph.flexml_layers[225].compute_node[1][3], compute_graph.flexml_layers[225].compute_node[2][0], compute_graph.flexml_layers[225].compute_node[2][1], compute_graph.flexml_layers[225].compute_node[2][2], compute_graph.flexml_layers[225].compute_node[2][3], compute_graph.flexml_layers[225].compute_node[3][0], compute_graph.flexml_layers[225].compute_node[3][1], compute_graph.flexml_layers[225].compute_node[3][2], compute_graph.flexml_layers[225].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[3], compute_graph.l2_284.in[0], compute_graph.l2_284.in[1], compute_graph.l2_284.in[2], compute_graph.l2_284.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(996): mode(3), layer(310): {compute_graph.l2_284.out[0], compute_graph.l2_284.out[1], compute_graph.l2l3_284_spill.in[0], compute_graph.l2l3_284_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1205): mode(0), layer(311): {compute_graph.l2l3_281_spill.out[2], compute_graph.l2l3_281_spill.out[3], compute_graph.templated_graph_285.ifm.in[0], compute_graph.templated_graph_285.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1321): mode(0), layer(311): {compute_graph.templated_graph_285.ifm.out[0], compute_graph.templated_graph_285.ifm.out[1], compute_graph.l2l3_285_spill.in[0], compute_graph.l2l3_285_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 997 will not be pipelined because it's in the same layer 312 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1621), layer(312): {compute_graph.ifm_ddr_1.out[4], compute_graph.l2l3_285_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1622), layer(312): {compute_graph.L2_IFM_Buffer_for_input1_2_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1623), layer(312): {compute_graph.ifm_ddr_1.out[5], compute_graph.l2l3_285_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1624), layer(312): {compute_graph.L2_IFM_Buffer_for_input1_2_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[1], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[2], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1554), layer(312): compute_graph.flexml_layers[226].compute_node[0][0], compute_graph.flexml_layers[226].compute_node[0][1], compute_graph.flexml_layers[226].compute_node[0][2], compute_graph.flexml_layers[226].compute_node[0][3], compute_graph.flexml_layers[226].compute_node[1][0], compute_graph.flexml_layers[226].compute_node[1][1], compute_graph.flexml_layers[226].compute_node[1][2], compute_graph.flexml_layers[226].compute_node[1][3], compute_graph.flexml_layers[226].compute_node[2][0], compute_graph.flexml_layers[226].compute_node[2][1], compute_graph.flexml_layers[226].compute_node[2][2], compute_graph.flexml_layers[226].compute_node[2][3], compute_graph.flexml_layers[226].compute_node[3][0], compute_graph.flexml_layers[226].compute_node[3][1], compute_graph.flexml_layers[226].compute_node[3][2], compute_graph.flexml_layers[226].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[1], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[2], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[3], compute_graph.l2_286.in[0], compute_graph.l2_286.in[1], compute_graph.l2_286.in[2], compute_graph.l2_286.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(893): mode(3), layer(313): {compute_graph.Layer_288_wts_ddr.out[0], compute_graph.Layer_288_wts_ddr.out[1], compute_graph.Layer_288_l2_wts.in[0], compute_graph.Layer_288_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(893) is pipelined with KernelLayerNode(1554) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(997): mode(3), layer(312): {compute_graph.l2_286.out[0], compute_graph.l2_286.out[1], compute_graph.spill_L3_Concat_Buffer_layer_287.in[2], compute_graph.spill_L3_Concat_Buffer_layer_287.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1212): mode(0), layer(313): {compute_graph.spill_L3_Concat_Buffer_layer_287.out[0], compute_graph.spill_L3_Concat_Buffer_layer_287.out[1], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[3], compute_graph.l2_288.in[0], compute_graph.l2_288.in[1], compute_graph.l2_288.in[2], compute_graph.l2_288.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1555), layer(313): compute_graph.flexml_layers[227].compute_node[0][0], compute_graph.flexml_layers[227].compute_node[0][1], compute_graph.flexml_layers[227].compute_node[0][2], compute_graph.flexml_layers[227].compute_node[0][3], compute_graph.flexml_layers[227].compute_node[1][0], compute_graph.flexml_layers[227].compute_node[1][1], compute_graph.flexml_layers[227].compute_node[1][2], compute_graph.flexml_layers[227].compute_node[1][3], compute_graph.flexml_layers[227].compute_node[2][0], compute_graph.flexml_layers[227].compute_node[2][1], compute_graph.flexml_layers[227].compute_node[2][2], compute_graph.flexml_layers[227].compute_node[2][3], compute_graph.flexml_layers[227].compute_node[3][0], compute_graph.flexml_layers[227].compute_node[3][1], compute_graph.flexml_layers[227].compute_node[3][2], compute_graph.flexml_layers[227].compute_node[3][3], {compute_graph.Layer_288_l2_wts.out[0], compute_graph.Layer_288_l2_wts.out[1], compute_graph.Layer_288_l2_wts.out[2], compute_graph.Layer_288_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[3], compute_graph.l2_288.in[0], compute_graph.l2_288.in[1], compute_graph.l2_288.in[2], compute_graph.l2_288.in[3]}, Scheduler computes number of sub-layer phases to be 6. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(999): mode(3), layer(313): {compute_graph.l2_288.out[0], compute_graph.l2_288.out[1], compute_graph.l2l3_288_spill.in[0], compute_graph.l2l3_288_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(999) is pipelined with KernelLayerNode(1555) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1213): mode(0), layer(314): {compute_graph.l2l3_288_spill.out[0], compute_graph.l2l3_288_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1556), layer(314): compute_graph.flexml_layers[228].compute_node[0][0], compute_graph.flexml_layers[228].compute_node[0][1], compute_graph.flexml_layers[228].compute_node[0][2], compute_graph.flexml_layers[228].compute_node[0][3], compute_graph.flexml_layers[228].compute_node[1][0], compute_graph.flexml_layers[228].compute_node[1][1], compute_graph.flexml_layers[228].compute_node[1][2], compute_graph.flexml_layers[228].compute_node[1][3], compute_graph.flexml_layers[228].compute_node[2][0], compute_graph.flexml_layers[228].compute_node[2][1], compute_graph.flexml_layers[228].compute_node[2][2], compute_graph.flexml_layers[228].compute_node[2][3], compute_graph.flexml_layers[228].compute_node[3][0], compute_graph.flexml_layers[228].compute_node[3][1], compute_graph.flexml_layers[228].compute_node[3][2], compute_graph.flexml_layers[228].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[3], compute_graph.l2_289.in[0], compute_graph.l2_289.in[1], compute_graph.l2_289.in[2], compute_graph.l2_289.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1000): mode(0), layer(314): {compute_graph.l2_289.out[0], compute_graph.l2_289.out[1], compute_graph.l2l3_289_spill.in[0], compute_graph.l2l3_289_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 1001 will not be pipelined because it's in the same layer 315 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1625), layer(315): {compute_graph.l2l3_282_spill.out[2], compute_graph.l2l3_289_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1626), layer(315): {compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1627), layer(315): {compute_graph.l2l3_282_spill.out[3], compute_graph.l2l3_289_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1628), layer(315): {compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1557), layer(315): compute_graph.flexml_layers[229].compute_node[0][0], compute_graph.flexml_layers[229].compute_node[0][1], compute_graph.flexml_layers[229].compute_node[0][2], compute_graph.flexml_layers[229].compute_node[0][3], compute_graph.flexml_layers[229].compute_node[1][0], compute_graph.flexml_layers[229].compute_node[1][1], compute_graph.flexml_layers[229].compute_node[1][2], compute_graph.flexml_layers[229].compute_node[1][3], compute_graph.flexml_layers[229].compute_node[2][0], compute_graph.flexml_layers[229].compute_node[2][1], compute_graph.flexml_layers[229].compute_node[2][2], compute_graph.flexml_layers[229].compute_node[2][3], compute_graph.flexml_layers[229].compute_node[3][0], compute_graph.flexml_layers[229].compute_node[3][1], compute_graph.flexml_layers[229].compute_node[3][2], compute_graph.flexml_layers[229].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[3], compute_graph.l2_290.in[0], compute_graph.l2_290.in[1], compute_graph.l2_290.in[2], compute_graph.l2_290.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1001): mode(3), layer(315): {compute_graph.l2_290.out[0], compute_graph.l2_290.out[1], compute_graph.l2l3_290_spill.in[0], compute_graph.l2l3_290_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1629), layer(316): {compute_graph.l2l3_284_spill.out[0], compute_graph.l2l3_290_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1630), layer(316): {compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1631), layer(316): {compute_graph.l2l3_284_spill.out[1], compute_graph.l2l3_290_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1632), layer(316): {compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1558), layer(316): compute_graph.flexml_layers[230].compute_node[0][0], compute_graph.flexml_layers[230].compute_node[0][1], compute_graph.flexml_layers[230].compute_node[0][2], compute_graph.flexml_layers[230].compute_node[0][3], compute_graph.flexml_layers[230].compute_node[1][0], compute_graph.flexml_layers[230].compute_node[1][1], compute_graph.flexml_layers[230].compute_node[1][2], compute_graph.flexml_layers[230].compute_node[1][3], compute_graph.flexml_layers[230].compute_node[2][0], compute_graph.flexml_layers[230].compute_node[2][1], compute_graph.flexml_layers[230].compute_node[2][2], compute_graph.flexml_layers[230].compute_node[2][3], compute_graph.flexml_layers[230].compute_node[3][0], compute_graph.flexml_layers[230].compute_node[3][1], compute_graph.flexml_layers[230].compute_node[3][2], compute_graph.flexml_layers[230].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[3], compute_graph.l2_291.in[0], compute_graph.l2_291.in[1], compute_graph.l2_291.in[2], compute_graph.l2_291.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1633), layer(316): {compute_graph.l2_291.out[0], compute_graph.l2_291.out[2]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1634), layer(316): {compute_graph.ofm_ddr_3_l2l3_291_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_292.in[2]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1635), layer(316): {compute_graph.l2_291.out[1], compute_graph.l2_291.out[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1636), layer(316): {compute_graph.ofm_ddr_3_l2l3_291_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_292.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1196): mode(0), layer(317): {compute_graph.l2l3_277_spill.out[0], compute_graph.l2l3_277_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1004): mode(0), layer(317): {compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_292.in[0], compute_graph.spill_L3_Concat_Buffer_layer_292.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 1322 will not be pipelined because it's in the same layer 318 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.spill_L3_Concat_Buffer_layer_292.out[0], compute_graph.spill_L3_Concat_Buffer_layer_292.out[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1216): mode(0), layer(318): {compute_graph.spill_L3_Concat_Buffer_layer_292.out[0], compute_graph.spill_L3_Concat_Buffer_layer_292.out[1], compute_graph.templated_graph_293.ifm_mem.in[0], compute_graph.templated_graph_293.ifm_mem.in[1]}, Scheduler computes number of sub-layer phases to be 23. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1323), layer(318): compute_graph.templated_graph_293.mk[0][0], compute_graph.templated_graph_293.mk[0][1], compute_graph.templated_graph_293.mk[0][2], compute_graph.templated_graph_293.mk[0][3], compute_graph.templated_graph_293.mk[1][0], compute_graph.templated_graph_293.mk[1][1], compute_graph.templated_graph_293.mk[1][2], compute_graph.templated_graph_293.mk[1][3], compute_graph.templated_graph_293.mk[2][0], compute_graph.templated_graph_293.mk[2][1], compute_graph.templated_graph_293.mk[2][2], compute_graph.templated_graph_293.mk[2][3], compute_graph.templated_graph_293.mk[3][0], compute_graph.templated_graph_293.mk[3][1], compute_graph.templated_graph_293.mk[3][2], compute_graph.templated_graph_293.mk[3][3], {compute_graph.templated_graph_293.ifm_mem.out[0], compute_graph.templated_graph_293.ifm_mem.out[1], compute_graph.templated_graph_293.ifm_mem.out[2], compute_graph.templated_graph_293.ifm_mem.out[3], compute_graph.templated_graph_293.ofm_mem.in[0], compute_graph.templated_graph_293.ofm_mem.in[1], compute_graph.templated_graph_293.ofm_mem.in[2], compute_graph.templated_graph_293.ofm_mem.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(894): mode(3), layer(319): {compute_graph.Layer_295_wts_ddr.out[0], compute_graph.Layer_295_wts_ddr.out[1], compute_graph.Layer_295_l2_wts.in[0], compute_graph.Layer_295_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(894) is pipelined with KernelLayerNode(1323) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.templated_graph_293.ofm_mem.out[0], compute_graph.templated_graph_293.ofm_mem.out[1], compute_graph.spill_L3_Concat_Buffer_layer_294.in[0], compute_graph.spill_L3_Concat_Buffer_layer_294.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1322): mode(3), layer(318): {compute_graph.templated_graph_293.ofm_mem.out[0], compute_graph.templated_graph_293.ofm_mem.out[1], compute_graph.spill_L3_Concat_Buffer_layer_294.in[0], compute_graph.spill_L3_Concat_Buffer_layer_294.in[1]}, Scheduler computes number of sub-layer phases to be 23. +INFO: [aiecompiler 77-23129] BufferToBufferNode 1006 will not be pipelined because it's in the same layer 319 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.spill_L3_Concat_Buffer_layer_294.out[0], compute_graph.spill_L3_Concat_Buffer_layer_294.out[1], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1217): mode(0), layer(319): {compute_graph.spill_L3_Concat_Buffer_layer_294.out[0], compute_graph.spill_L3_Concat_Buffer_layer_294.out[1], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[1]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1559), layer(319): compute_graph.flexml_layers[231].compute_node[0][0], compute_graph.flexml_layers[231].compute_node[0][1], compute_graph.flexml_layers[231].compute_node[0][2], compute_graph.flexml_layers[231].compute_node[0][3], compute_graph.flexml_layers[231].compute_node[1][0], compute_graph.flexml_layers[231].compute_node[1][1], compute_graph.flexml_layers[231].compute_node[1][2], compute_graph.flexml_layers[231].compute_node[1][3], compute_graph.flexml_layers[231].compute_node[2][0], compute_graph.flexml_layers[231].compute_node[2][1], compute_graph.flexml_layers[231].compute_node[2][2], compute_graph.flexml_layers[231].compute_node[2][3], compute_graph.flexml_layers[231].compute_node[3][0], compute_graph.flexml_layers[231].compute_node[3][1], compute_graph.flexml_layers[231].compute_node[3][2], compute_graph.flexml_layers[231].compute_node[3][3], {compute_graph.Layer_295_l2_wts.out[0], compute_graph.Layer_295_l2_wts.out[1], compute_graph.Layer_295_l2_wts.out[2], compute_graph.Layer_295_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[3], compute_graph.l2_295.in[0], compute_graph.l2_295.in[1], compute_graph.l2_295.in[2], compute_graph.l2_295.in[3]}, Scheduler computes number of sub-layer phases to be 6. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(895): mode(3), layer(320): {compute_graph.Layer_296_wts_ddr.out[0], compute_graph.Layer_296_wts_ddr.out[1], compute_graph.Layer_296_l2_wts.in[0], compute_graph.Layer_296_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(895) is pipelined with KernelLayerNode(1559) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1006): mode(3), layer(319): {compute_graph.l2_295.out[0], compute_graph.l2_295.out[1], compute_graph.l2l3_295_spill.in[0], compute_graph.l2l3_295_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 1007 will not be pipelined because it's in the same layer 320 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2l3_295_spill.out[0], compute_graph.l2l3_295_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1218): mode(0), layer(320): {compute_graph.l2l3_295_spill.out[0], compute_graph.l2l3_295_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[1]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1560), layer(320): compute_graph.flexml_layers[232].compute_node[0][0], compute_graph.flexml_layers[232].compute_node[0][1], compute_graph.flexml_layers[232].compute_node[0][2], compute_graph.flexml_layers[232].compute_node[0][3], compute_graph.flexml_layers[232].compute_node[1][0], compute_graph.flexml_layers[232].compute_node[1][1], compute_graph.flexml_layers[232].compute_node[1][2], compute_graph.flexml_layers[232].compute_node[1][3], compute_graph.flexml_layers[232].compute_node[2][0], compute_graph.flexml_layers[232].compute_node[2][1], compute_graph.flexml_layers[232].compute_node[2][2], compute_graph.flexml_layers[232].compute_node[2][3], compute_graph.flexml_layers[232].compute_node[3][0], compute_graph.flexml_layers[232].compute_node[3][1], compute_graph.flexml_layers[232].compute_node[3][2], compute_graph.flexml_layers[232].compute_node[3][3], {compute_graph.Layer_296_l2_wts.out[0], compute_graph.Layer_296_l2_wts.out[1], compute_graph.Layer_296_l2_wts.out[2], compute_graph.Layer_296_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[3], compute_graph.l2_296.in[0], compute_graph.l2_296.in[1], compute_graph.l2_296.in[2], compute_graph.l2_296.in[3]}, Scheduler computes number of sub-layer phases to be 9. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(896): mode(3), layer(321): {compute_graph.Layer_297_wts_ddr.out[0], compute_graph.Layer_297_wts_ddr.out[1], compute_graph.Layer_297_l2_wts.in[0], compute_graph.Layer_297_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(896) is pipelined with KernelLayerNode(1560) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1007): mode(3), layer(320): {compute_graph.l2_296.out[0], compute_graph.l2_296.out[1], compute_graph.l2l3_296_spill.in[0], compute_graph.l2l3_296_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1219): mode(0), layer(321): {compute_graph.l2l3_296_spill.out[0], compute_graph.l2l3_296_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1561), layer(321): compute_graph.flexml_layers[233].compute_node[0][0], compute_graph.flexml_layers[233].compute_node[0][1], compute_graph.flexml_layers[233].compute_node[0][2], compute_graph.flexml_layers[233].compute_node[0][3], compute_graph.flexml_layers[233].compute_node[1][0], compute_graph.flexml_layers[233].compute_node[1][1], compute_graph.flexml_layers[233].compute_node[1][2], compute_graph.flexml_layers[233].compute_node[1][3], compute_graph.flexml_layers[233].compute_node[2][0], compute_graph.flexml_layers[233].compute_node[2][1], compute_graph.flexml_layers[233].compute_node[2][2], compute_graph.flexml_layers[233].compute_node[2][3], compute_graph.flexml_layers[233].compute_node[3][0], compute_graph.flexml_layers[233].compute_node[3][1], compute_graph.flexml_layers[233].compute_node[3][2], compute_graph.flexml_layers[233].compute_node[3][3], {compute_graph.Layer_297_l2_wts.out[0], compute_graph.Layer_297_l2_wts.out[1], compute_graph.Layer_297_l2_wts.out[2], compute_graph.Layer_297_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[3], compute_graph.l2_297.in[0], compute_graph.l2_297.in[1], compute_graph.l2_297.in[2], compute_graph.l2_297.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1008): mode(3), layer(321): {compute_graph.l2_297.out[0], compute_graph.l2_297.out[1], compute_graph.l2l3_297_spill.in[0], compute_graph.l2l3_297_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1008) is pipelined with KernelLayerNode(1561) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1220): mode(0), layer(322): {compute_graph.l2l3_297_spill.out[0], compute_graph.l2l3_297_spill.out[1], compute_graph.templated_graph_298.ifm.in[0], compute_graph.templated_graph_298.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1325), layer(322): compute_graph.templated_graph_298.compute_node[0][0], compute_graph.templated_graph_298.compute_node[0][1], compute_graph.templated_graph_298.compute_node[0][2], compute_graph.templated_graph_298.compute_node[0][3], compute_graph.templated_graph_298.compute_node[1][0], compute_graph.templated_graph_298.compute_node[1][1], compute_graph.templated_graph_298.compute_node[1][2], compute_graph.templated_graph_298.compute_node[1][3], compute_graph.templated_graph_298.compute_node[2][0], compute_graph.templated_graph_298.compute_node[2][1], compute_graph.templated_graph_298.compute_node[2][2], compute_graph.templated_graph_298.compute_node[2][3], compute_graph.templated_graph_298.compute_node[3][0], compute_graph.templated_graph_298.compute_node[3][1], compute_graph.templated_graph_298.compute_node[3][2], compute_graph.templated_graph_298.compute_node[3][3], {compute_graph.templated_graph_298.ifm.out[0], compute_graph.templated_graph_298.ifm.out[1], compute_graph.templated_graph_298.ifm.out[2], compute_graph.templated_graph_298.ifm.out[3], compute_graph.templated_graph_298.ofm.in[0], compute_graph.templated_graph_298.ofm.in[1], compute_graph.templated_graph_298.ofm.in[2], compute_graph.templated_graph_298.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1324): mode(3), layer(322): {compute_graph.templated_graph_298.ofm.out[0], compute_graph.templated_graph_298.ofm.out[1], compute_graph.l2l3_298_spill.in[0], compute_graph.l2l3_298_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1324) is pipelined with KernelLayerNode(1325) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1222): mode(0), layer(323): {compute_graph.l2l3_298_spill.out[0], compute_graph.l2l3_298_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1562), layer(323): compute_graph.flexml_layers[234].compute_node[0][0], compute_graph.flexml_layers[234].compute_node[0][1], compute_graph.flexml_layers[234].compute_node[0][2], compute_graph.flexml_layers[234].compute_node[0][3], compute_graph.flexml_layers[234].compute_node[1][0], compute_graph.flexml_layers[234].compute_node[1][1], compute_graph.flexml_layers[234].compute_node[1][2], compute_graph.flexml_layers[234].compute_node[1][3], compute_graph.flexml_layers[234].compute_node[2][0], compute_graph.flexml_layers[234].compute_node[2][1], compute_graph.flexml_layers[234].compute_node[2][2], compute_graph.flexml_layers[234].compute_node[2][3], compute_graph.flexml_layers[234].compute_node[3][0], compute_graph.flexml_layers[234].compute_node[3][1], compute_graph.flexml_layers[234].compute_node[3][2], compute_graph.flexml_layers[234].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[3], compute_graph.L2_OFM_Buffer_layer_299.in[0], compute_graph.L2_OFM_Buffer_layer_299.in[1], compute_graph.L2_OFM_Buffer_layer_299.in[2], compute_graph.L2_OFM_Buffer_layer_299.in[3]}, Scheduler computes number of sub-layer phases to be 18. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1009): mode(3), layer(323): {compute_graph.L2_OFM_Buffer_layer_299.out[0], compute_graph.L2_OFM_Buffer_layer_299.out[1], compute_graph.ofm_ddr_4.in[0], compute_graph.ofm_ddr_4.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1009) is pipelined with KernelLayerNode(1562) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1221): mode(0), layer(324): {compute_graph.l2l3_297_spill.out[2], compute_graph.l2l3_297_spill.out[3], compute_graph.templated_graph_300.ifm.in[0], compute_graph.templated_graph_300.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1327), layer(324): compute_graph.templated_graph_300.compute_node[0][0], compute_graph.templated_graph_300.compute_node[0][1], compute_graph.templated_graph_300.compute_node[0][2], compute_graph.templated_graph_300.compute_node[0][3], compute_graph.templated_graph_300.compute_node[1][0], compute_graph.templated_graph_300.compute_node[1][1], compute_graph.templated_graph_300.compute_node[1][2], compute_graph.templated_graph_300.compute_node[1][3], compute_graph.templated_graph_300.compute_node[2][0], compute_graph.templated_graph_300.compute_node[2][1], compute_graph.templated_graph_300.compute_node[2][2], compute_graph.templated_graph_300.compute_node[2][3], compute_graph.templated_graph_300.compute_node[3][0], compute_graph.templated_graph_300.compute_node[3][1], compute_graph.templated_graph_300.compute_node[3][2], compute_graph.templated_graph_300.compute_node[3][3], {compute_graph.templated_graph_300.ifm.out[0], compute_graph.templated_graph_300.ifm.out[1], compute_graph.templated_graph_300.ifm.out[2], compute_graph.templated_graph_300.ifm.out[3], compute_graph.templated_graph_300.ofm.in[0], compute_graph.templated_graph_300.ofm.in[1], compute_graph.templated_graph_300.ofm.in[2], compute_graph.templated_graph_300.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1326): mode(3), layer(324): {compute_graph.templated_graph_300.ofm.out[0], compute_graph.templated_graph_300.ofm.out[1], compute_graph.l2l3_300_spill.in[0], compute_graph.l2l3_300_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1326) is pipelined with KernelLayerNode(1327) +INFO: [aiecompiler 77-23129] BufferToBufferNode 1010 will not be pipelined because it's in the same layer 325 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1637), layer(325): {compute_graph.l2l3_5_spill.out[6], compute_graph.l2l3_300_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1638), layer(325): {compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1639), layer(325): {compute_graph.l2l3_5_spill.out[7], compute_graph.l2l3_300_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1640), layer(325): {compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1563), layer(325): compute_graph.flexml_layers[235].compute_node[0][0], compute_graph.flexml_layers[235].compute_node[0][1], compute_graph.flexml_layers[235].compute_node[0][2], compute_graph.flexml_layers[235].compute_node[0][3], compute_graph.flexml_layers[235].compute_node[1][0], compute_graph.flexml_layers[235].compute_node[1][1], compute_graph.flexml_layers[235].compute_node[1][2], compute_graph.flexml_layers[235].compute_node[1][3], compute_graph.flexml_layers[235].compute_node[2][0], compute_graph.flexml_layers[235].compute_node[2][1], compute_graph.flexml_layers[235].compute_node[2][2], compute_graph.flexml_layers[235].compute_node[2][3], compute_graph.flexml_layers[235].compute_node[3][0], compute_graph.flexml_layers[235].compute_node[3][1], compute_graph.flexml_layers[235].compute_node[3][2], compute_graph.flexml_layers[235].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[3], compute_graph.l2_301.in[0], compute_graph.l2_301.in[1], compute_graph.l2_301.in[2], compute_graph.l2_301.in[3]}, Scheduler computes number of sub-layer phases to be 9. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1010): mode(3), layer(325): {compute_graph.l2_301.out[0], compute_graph.l2_301.out[1], compute_graph.l2l3_301_spill.in[0], compute_graph.l2l3_301_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1224): mode(0), layer(326): {compute_graph.l2l3_301_spill.out[0], compute_graph.l2l3_301_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1564), layer(326): compute_graph.flexml_layers[236].compute_node[0][0], compute_graph.flexml_layers[236].compute_node[0][1], compute_graph.flexml_layers[236].compute_node[0][2], compute_graph.flexml_layers[236].compute_node[0][3], compute_graph.flexml_layers[236].compute_node[1][0], compute_graph.flexml_layers[236].compute_node[1][1], compute_graph.flexml_layers[236].compute_node[1][2], compute_graph.flexml_layers[236].compute_node[1][3], compute_graph.flexml_layers[236].compute_node[2][0], compute_graph.flexml_layers[236].compute_node[2][1], compute_graph.flexml_layers[236].compute_node[2][2], compute_graph.flexml_layers[236].compute_node[2][3], compute_graph.flexml_layers[236].compute_node[3][0], compute_graph.flexml_layers[236].compute_node[3][1], compute_graph.flexml_layers[236].compute_node[3][2], compute_graph.flexml_layers[236].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[3], compute_graph.L2_OFM_Buffer_layer_302.in[0], compute_graph.L2_OFM_Buffer_layer_302.in[1], compute_graph.L2_OFM_Buffer_layer_302.in[2], compute_graph.L2_OFM_Buffer_layer_302.in[3]}, Scheduler computes number of sub-layer phases to be 18. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1011): mode(3), layer(326): {compute_graph.L2_OFM_Buffer_layer_302.out[0], compute_graph.L2_OFM_Buffer_layer_302.out[1], compute_graph.ofm_ddr_5.in[0], compute_graph.ofm_ddr_5.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1011) is pipelined with KernelLayerNode(1564) +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-22492' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[8].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[18].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[19].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[28].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[29].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[38].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[39].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[100].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[101].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[118].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[119].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[136].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[137].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[154].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[155].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[172].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[173].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[185].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[186].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.templated_graph_231.mk[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[203].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.templated_graph_254.mk[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[213].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.templated_graph_274.mk[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.templated_graph_298.compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[8].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[18].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[19].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[28].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[29].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[38].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[39].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[100].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[101].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[118].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[119].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[136].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[137].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[154].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[155].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[172].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[173].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[185].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[186].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.templated_graph_231.mk[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[203].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.templated_graph_254.mk[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[213].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.templated_graph_274.mk[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.templated_graph_298.compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[8].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[18].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[19].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[28].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[29].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[38].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[39].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[100].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[101].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[118].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[119].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[136].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[137].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[154].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[155].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[172].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[173].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[185].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[186].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.templated_graph_231.mk[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[203].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.templated_graph_254.mk[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[213].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.templated_graph_274.mk[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.templated_graph_298.compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[8].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[18].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[19].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[28].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[29].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[38].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[39].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[100].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[101].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[118].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[119].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[136].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[137].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[154].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[155].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[172].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[173].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[185].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[186].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.templated_graph_231.mk[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[203].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.templated_graph_254.mk[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[213].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.templated_graph_274.mk[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.templated_graph_298.compute_node[0][3] +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-22549' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-6272] Completing Scheduler pass +INFO: [aiecompiler 77-23111] Writing layer control parameter JSON file Work/aie/layer_control_parameters.json +INFO: [aiecompiler 77-6593] Max LCP size in 32-bit words: 225 +INFO: [aiecompiler 77-6497] runtime-opt stats: Avoided compilation of 418 cores out of 432 cores. +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0/scripts/0_0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1/scripts/0_1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2/scripts/0_2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3/scripts/0_3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (1,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/1_0/scripts/1_0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/1_0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (1,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/1_1/scripts/1_1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/1_1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (1,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/1_2/scripts/1_2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/1_2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (1,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/1_3/scripts/1_3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/1_3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (2,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/2_0/scripts/2_0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/2_0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (2,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/2_1/scripts/2_1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/2_1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (2,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/2_2/scripts/2_2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/2_2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (2,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/2_3/scripts/2_3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/2_3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (3,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/3_0/scripts/3_0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/3_0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (3,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/3_1/scripts/3_1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/3_1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (3,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/3_2/scripts/3_2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/3_2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (3,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/3_3/scripts/3_3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/3_3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable0/scripts/0_0_reloadable0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable1/scripts/0_0_reloadable1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable2/scripts/0_0_reloadable2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable3/scripts/0_0_reloadable3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable4/scripts/0_0_reloadable4.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable4/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable5/scripts/0_0_reloadable5.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable5/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable6/scripts/0_0_reloadable6.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable6/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable7/scripts/0_0_reloadable7.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable7/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable8/scripts/0_0_reloadable8.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable8/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable9/scripts/0_0_reloadable9.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable9/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable10/scripts/0_0_reloadable10.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable10/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable11/scripts/0_0_reloadable11.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable11/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable12/scripts/0_0_reloadable12.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable12/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable13/scripts/0_0_reloadable13.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable13/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable14/scripts/0_0_reloadable14.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable14/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable15/scripts/0_0_reloadable15.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable15/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable16/scripts/0_0_reloadable16.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable16/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable17/scripts/0_0_reloadable17.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable17/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable18/scripts/0_0_reloadable18.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable18/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable19/scripts/0_0_reloadable19.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable19/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable20/scripts/0_0_reloadable20.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable20/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable21/scripts/0_0_reloadable21.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable21/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable22/scripts/0_0_reloadable22.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable22/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable23/scripts/0_0_reloadable23.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable23/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable24/scripts/0_0_reloadable24.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable24/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable25/scripts/0_0_reloadable25.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable25/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable0/scripts/0_1_reloadable0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable1/scripts/0_1_reloadable1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable2/scripts/0_1_reloadable2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable3/scripts/0_1_reloadable3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable4/scripts/0_1_reloadable4.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable4/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable5/scripts/0_1_reloadable5.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable5/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable6/scripts/0_1_reloadable6.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable6/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable7/scripts/0_1_reloadable7.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable7/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable8/scripts/0_1_reloadable8.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable8/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable9/scripts/0_1_reloadable9.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable9/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable10/scripts/0_1_reloadable10.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable10/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable11/scripts/0_1_reloadable11.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable11/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable12/scripts/0_1_reloadable12.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable12/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable13/scripts/0_1_reloadable13.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable13/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable14/scripts/0_1_reloadable14.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable14/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable15/scripts/0_1_reloadable15.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable15/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable16/scripts/0_1_reloadable16.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable16/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable17/scripts/0_1_reloadable17.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable17/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable18/scripts/0_1_reloadable18.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable18/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable19/scripts/0_1_reloadable19.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable19/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable20/scripts/0_1_reloadable20.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable20/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable21/scripts/0_1_reloadable21.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable21/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable22/scripts/0_1_reloadable22.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable22/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable23/scripts/0_1_reloadable23.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable23/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable24/scripts/0_1_reloadable24.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable24/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable25/scripts/0_1_reloadable25.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable25/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable0/scripts/0_2_reloadable0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable1/scripts/0_2_reloadable1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable2/scripts/0_2_reloadable2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable3/scripts/0_2_reloadable3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable4/scripts/0_2_reloadable4.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable4/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable5/scripts/0_2_reloadable5.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable5/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable6/scripts/0_2_reloadable6.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable6/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable7/scripts/0_2_reloadable7.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable7/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable8/scripts/0_2_reloadable8.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable8/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable9/scripts/0_2_reloadable9.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable9/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable10/scripts/0_2_reloadable10.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable10/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable11/scripts/0_2_reloadable11.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable11/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable12/scripts/0_2_reloadable12.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable12/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable13/scripts/0_2_reloadable13.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable13/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable14/scripts/0_2_reloadable14.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable14/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable15/scripts/0_2_reloadable15.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable15/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable16/scripts/0_2_reloadable16.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable16/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable17/scripts/0_2_reloadable17.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable17/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable18/scripts/0_2_reloadable18.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable18/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable19/scripts/0_2_reloadable19.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable19/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable20/scripts/0_2_reloadable20.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable20/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable21/scripts/0_2_reloadable21.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable21/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable22/scripts/0_2_reloadable22.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable22/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable23/scripts/0_2_reloadable23.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable23/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable24/scripts/0_2_reloadable24.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable24/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable25/scripts/0_2_reloadable25.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable25/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable0/scripts/0_3_reloadable0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3_reloadable0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable1/scripts/0_3_reloadable1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3_reloadable1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable2/scripts/0_3_reloadable2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3_reloadable2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable3/scripts/0_3_reloadable3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3_reloadable3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable4/scripts/0_3_reloadable4.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3_reloadable4/timestamped_log +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-21796' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-6298' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable5/scripts/0_3_reloadable5.bcf +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-6299' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable0.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable1.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable2.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable3.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable5.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable17.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable19.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable20.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable21.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable22.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable23.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable24.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable25.llgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable0.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable1.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable2.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable3.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable5.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable17.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable19.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable20.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable21.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable22.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable23.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable24.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable25.elfgen.Makefile all 2>&1 + +INFO: [aiecompiler 77-22140] Generating Control Packet Binary +INFO: [aiecompiler 77-22511] Creating External Buffers Json Work/ps/c_rts/external_buffer_id.json +INFO: [aiecompiler 77-22141] Generating Transaction Binary +INFO: [aiecompiler 77-6311] Opened file : Work/ps/cdo/generated-source/cdo_main.cpp +INFO: [aiecompiler 77-6311] Opened file : Work/ps/cdo/generated-source/gen_cdo.cpp +INFO: [aiecompiler 77-6311] Opened file : Work/ps/cdo/generated-source/gen_cdo.h +INFO: [aiecompiler 77-6311] Opened file : Work/ps/cdo/Makefile +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/ps/cdo -f Makefile all +INFO: [aiecompiler 77-6439] Run completed. Find additional information in: + Guidance: Work/reports/guidance.html + +INFO: [aiecompiler 77-6440] Use the vitis_analyzer tool to visualize and navigate the relevant reports. Run: + vitis_analyzer Work/top.aiecompile_summary + +Compilation Complete +(WARNING:62, CRITICAL-WARNING:0, ERROR:0) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/.target b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/.target new file mode 100644 index 0000000000000000000000000000000000000000..04c6b38c182ba78ab9abf45c5b7604d20e6b06a4 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/.target @@ -0,0 +1 @@ +hw \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/0_0.log b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/0_0.log new file mode 100644 index 0000000000000000000000000000000000000000..471966c8ca3642c4e7815e6aa3938298f02953bd --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/0_0.log @@ -0,0 +1,59 @@ +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0.ll -o../Release/chesswork848/0_0.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork848 ../Release/chesswork848/0_0.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0 -L +chess-backend 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0 -L +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/include/adf/aie/tile_control.h", line 292, column 8: in "/usr/local/lib/python3.10/dist-packages/include/adf/aie/tile_control.h", line 292: (loop #8) + loop software pipelining (to 2 cycles) is feasible for a minimum loop count of 5, + but requires the creation of a post-amble, for which the loop was not prepared + ... consider annotating the loop with `chess_prepare_for_pipelining', as well as + increasing the current `chess_loop_range(1,)` annotation to `chess_loop_range(5,)', or remove it. + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0 -L --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "0_0/src/0_0.cc", line 37, column 12: in "0_0/src/0_0.cc", line 37: (loop #13) + loop software pipelining (to 8 cycles) is feasible but requires the creation of a post-amble, + for which the loop was not prepared + ... consider annotating the loop with `chess_prepare_for_pipelining' + +Warning: in "0_0/src/0_0.cc", line 12: (loop #3) + Non leaf loop was prepared for pipelining. But the pipelined solutions have not been selected. + Consider removing the chess_prepare_for_pipelining directive as it may improve results +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0 -L --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0.objlist -o../0_0.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0" +bridge -o../Release/0_0 ../Release/0_0.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork848 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (19 errors, 3 warnings) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0 b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0 new file mode 100644 index 0000000000000000000000000000000000000000..4b382d8360b3c882bd183d59e4ab847db514ec8e Binary files /dev/null and b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0 differ diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.# b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.# new file mode 100644 index 0000000000000000000000000000000000000000..3fe156513c3670054f2aa8355a219a6c55705dd2 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.# @@ -0,0 +1,2 @@ +3d25a3bfdb1b94e31ca421fe169265bb6b32498c +1a735f496f1284ab86135ede88939a8aeeb375a3 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.## b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.## new file mode 100644 index 0000000000000000000000000000000000000000..b9fb424512073e296004b111422bb87a047163c7 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.## @@ -0,0 +1,2 @@ +46f16ef751fb28f6dce00ef4d676df021c165b9a +92e103875c6a39bc8cdbe4a21c02c5a94ba3cc96 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.calltree b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.calltree new file mode 100644 index 0000000000000000000000000000000000000000..f8a868bf7b484d791433b0f8bebebdfc0725be44 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.calltree @@ -0,0 +1,32 @@ + +// File generated by bridge version V-2024.06#84922c0d9f#241219, Fri Mar 21 03:42:51 2025 +// Copyright 2014-2024 Synopsys, Inc. All rights reserved. +// bridge -o../Release/0_0 ../Release/0_0.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork848 -pme + + +// Release: ipp V-2024.06-TGT-241219 + +_main_init + _main + _ZN3adf11block_writeEPKNS_7reg_valEj + _Z13kernelWrapperPPvjjjj + __cxa_finalize + _fini (referenced text) + + +Call tree stack and functions sizes: + +stack stack stack call func func function name + desc level level desc +----- ----- ----- ----- ----- ----- -------------------------------------------------------------- + 0 320 0 0 224 1998 _main_init + 192 320 1 1 1108 1774 _main + 0 0 2 2 174 174 _ZN3adf11block_writeEPKNS_7reg_valEj + * * 2 2 * * _Z13kernelWrapperPPvjjjj (_extern) + 64 128 2 2 324 492 __cxa_finalize + 64 64 3 3 168 168 _fini + + +Maximum call level : 3 +Maximum stack level: 3 +Maximum stack size : 320 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.cmic2 b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.cmic2 new file mode 100644 index 0000000000000000000000000000000000000000..f5651c6a640c9df2e9a1eaae268b6cc9f28cc921 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.cmic2 @@ -0,0 +1,2640 @@ + +// File generated by darts version V-2024.06#84922c0d9f#241219, Fri Mar 21 03:42:52 2025 +// Copyright 2014-2024 Synopsys, Inc. All rights reserved. +// darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0 me + +// Release: ipp V-2024.06-TGT-241219 +.label __AIE_ARCH_MODEL_VERSION__21011100__inlined__1__me_basic___main_init_ +.label _main_init +.function _main_init _main_init +.src_ref 0 "me_basic.c" 91 4 first +.src_ref 0 "me_basic.c" 87 first +.function_start + 0 "01000100" // MOVXM sp, #506560 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1 "10000000" // /* MW 5 */ + 2 "11110101" // /* MW 4 */ + 3 "10111001" // /* MW 3 */ + 4 "00000111" // /* MW 2 */ + 5 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 69 8 +.src_ref 0 "me_basic.c" 69 41 + 6 "01000100" // MOVXM r8, #0 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 7 "00000000" // /* MW 5 */ + 8 "00100000" // /* MW 4 */ + 9 "00000100" // /* MW 3 */ + 10 "00000000" // /* MW 2 */ + 11 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 69 8 + 12 "01000100" // MOVXM r16, #0 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 13 "00000000" // /* MW 5 */ + 14 "00100000" // /* MW 4 */ + 15 "00001000" // /* MW 3 */ + 16 "00000000" // /* MW 2 */ + 17 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 69 8 first + 18 "10011000" // EQ r16, r8, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 19 "00000111" // /* MW 3 */ + 20 "00100001" // /* MW 2 */ + 21 "00010010" // /* MW 1 */ +.src_ref 0 "me_basic.c" 69 8 + 22 "10000100" // JNZ r16, #128 /* MW 6 */ /* control_operation: words=6 jump conditional cycles_taken=1 cycles_not_taken=0 direct absolute target_address=128 delay_slots=5 */ + 23 "00000001" // /* MW 5 */ + 24 "01000000" // /* MW 4 */ + 25 "01000000" // /* MW 3 */ + 26 "00000000" // /* MW 2 */ + 27 "10000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 28 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 29 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 30 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 31 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 32 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 33 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 34 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 35 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 36 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 37 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 98 11 +.src_ref 0 "me_basic.c" 98 11 + 38 "10111010" // NOPA; MOVS p7, p0; MOV r9, r1 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 39 "01110010" // /* MW 9 */ + 40 "01010000" // /* MW 8 */ + 41 "00101000" // /* MW 7 */ + 42 "00000001" // /* MW 6 */ + 43 "10001011" // /* MW 5 */ + 44 "10000000" // /* MW 4 */ + 45 "11110111" // /* MW 3 */ + 46 "00101100" // /* MW 2 */ + 47 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 69 41 +.src_ref 0 "me_basic.c" 70 13 + 48 "11100001" // NOPA; NOPB; NOPS; MOVXM p6, #-4; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 49 "00000000" // /* MW 15 */ + 50 "00000000" // /* MW 14 */ + 51 "00010000" // /* MW 13 */ + 52 "11111110" // /* MW 12 */ + 53 "00110111" // /* MW 11 */ + 54 "11111111" // /* MW 10 */ + 55 "11111111" // /* MW 9 */ + 56 "00111111" // /* MW 8 */ + 57 "01011011" // /* MW 7 */ + 58 "00000001" // /* MW 6 */ + 59 "00100000" // /* MW 5 */ + 60 "00000000" // /* MW 4 */ + 61 "11110000" // /* MW 3 */ + 62 "00101100" // /* MW 2 */ + 63 "00000000" // /* MW 1 */ +.label TGT_F_main_init_64 +.src_ref 0 "me_basic.c" 69 41 +.src_ref 0 "me_basic.c" 70 13 first +.loop_nesting 1 + 64 "11010100" // LDA p0, [p6], #-4; MOV r10, p6 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 65 "10000001" // /* MW 5 */ + 66 "00111001" // /* MW 4 */ + 67 "11010101" // /* MW 3 */ + 68 "10000011" // /* MW 2 */ + 69 "11011111" // /* MW 1 */ + 70 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 71 "00000000" // /* MW 1 */ + 72 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 73 "00000000" // /* MW 1 */ + 74 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 75 "00000000" // /* MW 1 */ + 76 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 77 "00000000" // /* MW 1 */ + 78 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 79 "00000000" // /* MW 1 */ + 80 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 81 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 70 16 +.no_stack_arguments + 82 "00011000" // JL p0 /* MW 4 */ /* control_operation: words=4 call unconditional cycles_taken=1 indirect absolute delay_slots=5 */ + 83 "00000000" // /* MW 3 */ + 84 "00110000" // /* MW 2 */ + 85 "00010000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 86 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 87 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 88 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 89 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 90 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 91 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 92 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 93 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 94 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 95 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 69 41 first +.return_address + 96 "10011000" // NE r16, r10, r8 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 97 "10001000" // /* MW 3 */ + 98 "10100000" // /* MW 2 */ + 99 "00010010" // /* MW 1 */ +.src_ref 0 "me_basic.c" 69 8 + 100 "10000100" // JNZ r16, #64 /* MW 6 */ /* control_operation: words=6 jump conditional cycles_taken=1 cycles_not_taken=0 direct absolute target_address=64 delay_slots=5 */ + 101 "00000001" // /* MW 5 */ + 102 "01000000" // /* MW 4 */ + 103 "00100000" // /* MW 3 */ + 104 "00000000" // /* MW 2 */ + 105 "10000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 106 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 107 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 108 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 109 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 110 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 111 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 112 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 113 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 114 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 115 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 98 11 +.src_ref 0 "me_basic.c" 98 11 +.loop_nesting 0 + 116 "11110110" // NOPA; NOPB; MOVS p0, p7; MOV r1, r9 /* MW 12 */ /* control_operation: words=12 cycles_taken=1 */ + 117 "01110000" // /* MW 11 */ + 118 "01010000" // /* MW 10 */ + 119 "00101010" // /* MW 9 */ + 120 "00000000" // /* MW 8 */ + 121 "10001011" // /* MW 7 */ + 122 "10011100" // /* MW 6 */ + 123 "00100000" // /* MW 5 */ + 124 "00000000" // /* MW 4 */ + 125 "11110000" // /* MW 3 */ + 126 "00101100" // /* MW 2 */ + 127 "00000000" // /* MW 1 */ +.label TGT_F_main_init_128 +.src_ref 0 "me_basic.c" 98 11 first +.no_stack_arguments + 128 "00000100" // JL #224 /* MW 6 */ /* control_operation: words=6 call unconditional cycles_taken=1 direct absolute target_address=224 delay_slots=5 */ + 129 "00000001" // /* MW 5 */ + 130 "00000000" // /* MW 4 */ + 131 "01110000" // /* MW 3 */ + 132 "00000000" // /* MW 2 */ + 133 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 134 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 135 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 136 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 137 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 138 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 139 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 140 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 141 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 142 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 143 "00000000" // /* MW 1 */ +.src_ref 1 "stdlib.h" 77 4 first +.return_address +.no_stack_arguments + 144 "00000100" // JL #1696 /* MW 6 */ /* control_operation: words=6 call unconditional cycles_taken=1 direct absolute target_address=1696 delay_slots=5 */ + 145 "00000001" // /* MW 5 */ + 146 "00000000" // /* MW 4 */ + 147 "01010000" // /* MW 3 */ + 148 "00000011" // /* MW 2 */ + 149 "00000000" // /* MW 1 */ +.src_ref 1 "stdlib.h" 77 4 +.delay_slot + 150 "10111000" // MOV p0, #0 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 151 "00000000" // /* MW 3 */ + 152 "01100000" // /* MW 2 */ + 153 "00011000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 154 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 155 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 156 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 157 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 158 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 159 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 160 "11100001" // NOPA; NOPB; NOPS; NOPX; NOPM; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 161 "00000000" // /* MW 15 */ + 162 "00000000" // /* MW 14 */ + 163 "01111000" // /* MW 13 */ + 164 "10100101" // /* MW 12 */ + 165 "00000001" // /* MW 11 */ + 166 "00000000" // /* MW 10 */ + 167 "00000000" // /* MW 9 */ + 168 "00000000" // /* MW 8 */ + 169 "01011011" // /* MW 7 */ + 170 "00000001" // /* MW 6 */ + 171 "00100000" // /* MW 5 */ + 172 "00000000" // /* MW 4 */ + 173 "11110000" // /* MW 3 */ + 174 "00101100" // /* MW 2 */ + 175 "00000000" // /* MW 1 */ +.return_address +.swstall chess_separator_scheduler + 176 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 177 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 178 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 179 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 180 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 181 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 182 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 183 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 184 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 185 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 186 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 187 "00000000" // /* MW 1 */ +.src_ref 1 "stdlib.h" 61 4 first + 188 "00011000" // DONE /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 189 "00000000" // /* MW 3 */ + 190 "00001000" // /* MW 2 */ + 191 "00010000" // /* MW 1 */ +.swstall chess_separator_scheduler + 192 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 193 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 194 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 195 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 196 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 197 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 198 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 199 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 200 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 201 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 202 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 203 "00000000" // /* MW 1 */ +.src_ref 1 "stdlib.h" 62 4 first +.swstall for_chess_exit +.exit + 204 "10011000" // NOPA /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 205 "01100111" // /* MW 3 */ + 206 "00000001" // /* MW 2 */ + 207 "00000000" // /* MW 1 */ +.label TGT_F_main_init_208 +.src_ref 1 "stdlib.h" 64 4 first +.loop_nesting 1 + 208 "10000100" // J #208 /* MW 6 */ /* control_operation: words=6 jump unconditional cycles_taken=1 direct absolute target_address=208 delay_slots=5 */ + 209 "00000000" // /* MW 5 */ + 210 "00000000" // /* MW 4 */ + 211 "01101000" // /* MW 3 */ + 212 "00000000" // /* MW 2 */ + 213 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 214 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 215 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 216 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 217 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 218 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 219 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 220 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 221 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 222 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ +.label _main_init__end + 223 "00000000" // /* MW 1 */ +.label _main___func_begin0 +.label _main +.function main _main +.src_ref 2 "0_0.cc" 12 +.src_ref 2 "0_0.cc" 12 first +.function_start + 224 "10111010" // MOVA m0, #-160; PADDXM [sp], #192 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 225 "01110000" // /* MW 9 */ + 226 "00000000" // /* MW 8 */ + 227 "00000000" // /* MW 7 */ + 228 "00000000" // /* MW 6 */ + 229 "00000110" // /* MW 5 */ + 230 "00000000" // /* MW 4 */ + 231 "10000000" // /* MW 3 */ + 232 "00000000" // /* MW 2 */ + 233 "11101100" // /* MW 1 */ +.src_ref 3 "tile_control.h" 147 12 +.src_ref 3 "tile_control.h" 147 14 +.src_ref 2 "0_0.cc" 49 38 +.src_ref 2 "0_0.cc" 50 33 +.src_ref 2 "0_0.cc" 59 50 +.src_ref 2 "0_0.cc" 59 63 +.src_ref 2 "0_0.cc" 63 43 +.src_ref 2 "0_0.cc" 64 63 +.src_ref 2 "0_0.cc" 67 46 + 234 "10111010" // MOVA m1, #-184; MOVX r12, #2; MOV p1, sp /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 235 "01111000" // /* MW 9 */ + 236 "11110000" // /* MW 8 */ + 237 "10110010" // /* MW 7 */ + 238 "01001000" // /* MW 6 */ + 239 "11000000" // /* MW 5 */ + 240 "00000000" // /* MW 4 */ + 241 "10000000" // /* MW 3 */ + 242 "00000100" // /* MW 2 */ + 243 "11101001" // /* MW 1 */ +.src_ref 3 "tile_control.h" 278 68 +.src_ref 3 "tile_control.h" 278 68 +.src_ref 3 "tile_control.h" 278 68 +.src_ref 4 "io_buffer_compiler.h" 566 27 +.src_ref 4 "io_buffer_compiler.h" 567 18 + 244 "01111110" // MOVA r25, #0; PADDB [p1], m0; MOVS p6, p1; MOVXM p0, #651488 /* MW 14 */ /* control_operation: words=14 cycles_taken=1 */ + 245 "01100000" // /* MW 13 */ + 246 "10010001" // /* MW 12 */ + 247 "11010000" // /* MW 11 */ + 248 "00000010" // /* MW 10 */ + 249 "00001110" // /* MW 9 */ + 250 "10000110" // /* MW 8 */ + 251 "01001111" // /* MW 7 */ + 252 "00000000" // /* MW 6 */ + 253 "00100000" // /* MW 5 */ + 254 "00010111" // /* MW 4 */ + 255 "00000010" // /* MW 3 */ + 256 "00011001" // /* MW 2 */ + 257 "00000000" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 564 18 +.src_ref 4 "io_buffer_compiler.h" 565 24 +.src_ref 4 "io_buffer_compiler.h" 572 18 +.src_ref 2 "0_0.cc" 19 8 + 258 "01111110" // NOPA; PADDB [p6], m1; ST p1, [sp, #-4]; MOVX r16, #1; MOV r24, #0 /* MW 14 */ /* control_operation: words=14 cycles_taken=1 */ + 259 "10110000" // /* MW 13 */ + 260 "10010011" // /* MW 12 */ + 261 "11111111" // /* MW 11 */ + 262 "00001011" // /* MW 10 */ + 263 "00000000" // /* MW 9 */ + 264 "01100001" // /* MW 8 */ + 265 "00000101" // /* MW 7 */ + 266 "00100000" // /* MW 6 */ + 267 "00100000" // /* MW 5 */ + 268 "01010111" // /* MW 4 */ + 269 "11111100" // /* MW 3 */ + 270 "00101100" // /* MW 2 */ + 271 "00000000" // /* MW 1 */ +.label TGT_F_main_48 +.src_ref 3 "tile_control.h" 278 68 first +.loop_nesting 1 + 272 "10011000" // ST.TM r25, [p0], #16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 273 "00111110" // /* MW 3 */ + 274 "01001111" // /* MW 2 */ + 275 "00001000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 278 68 + 276 "10011000" // ST.TM r25, [p0], #-16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 277 "00111110" // /* MW 3 */ + 278 "11001111" // /* MW 2 */ + 279 "00001000" // /* MW 1 */ + 280 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 281 "00000000" // /* MW 1 */ + 282 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 283 "00000000" // /* MW 1 */ + 284 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 285 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 19 8 first + 286 "00011000" // ACQ #62, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 287 "00001000" // /* MW 3 */ + 288 "11000011" // /* MW 2 */ + 289 "00010111" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 566 27 +.src_ref 2 "0_0.cc" 29 31 + 290 "10111010" // MOVA m7, #-92; MOVXM p7, #504448 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 291 "00010000" // /* MW 9 */ + 292 "01000000" // /* MW 8 */ + 293 "10110001" // /* MW 7 */ + 294 "11101111" // /* MW 6 */ + 295 "00000001" // /* MW 5 */ + 296 "00000000" // /* MW 4 */ + 297 "10000000" // /* MW 3 */ + 298 "10011100" // /* MW 2 */ + 299 "11110100" // /* MW 1 */ + 300 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 301 "00000000" // /* MW 1 */ + 302 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 303 "00000000" // /* MW 1 */ + 304 "10011000" // ST p0, [sp, #-8] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 305 "00011101" // /* MW 3 */ + 306 "11111000" // /* MW 2 */ + 307 "00001111" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 564 18 first + 308 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 309 "00010001" // /* MW 3 */ + 310 "00011111" // /* MW 2 */ + 311 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 565 24 first + 312 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 313 "00010001" // /* MW 3 */ + 314 "00011111" // /* MW 2 */ + 315 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 565 24 + 316 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 317 "00010001" // /* MW 3 */ + 318 "00011111" // /* MW 2 */ + 319 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 567 18 first + 320 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 321 "00110001" // /* MW 3 */ + 322 "00011111" // /* MW 2 */ + 323 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 567 18 + 324 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 325 "00110001" // /* MW 3 */ + 326 "00011111" // /* MW 2 */ + 327 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 566 27 first + 328 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 329 "00110001" // /* MW 3 */ + 330 "00011111" // /* MW 2 */ + 331 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 564 18 first + 332 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 333 "00010001" // /* MW 3 */ + 334 "00011111" // /* MW 2 */ + 335 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 565 24 first + 336 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 337 "00010001" // /* MW 3 */ + 338 "00011111" // /* MW 2 */ + 339 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 565 24 + 340 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 341 "00010001" // /* MW 3 */ + 342 "00011111" // /* MW 2 */ + 343 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 567 18 first + 344 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 345 "00110001" // /* MW 3 */ + 346 "00011111" // /* MW 2 */ + 347 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 567 18 + 348 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 349 "00110001" // /* MW 3 */ + 350 "00011111" // /* MW 2 */ + 351 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 566 27 first + 352 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 353 "00110001" // /* MW 3 */ + 354 "00011111" // /* MW 2 */ + 355 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 564 18 first + 356 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 357 "00010001" // /* MW 3 */ + 358 "00011111" // /* MW 2 */ + 359 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 565 24 first + 360 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 361 "00010001" // /* MW 3 */ + 362 "00011111" // /* MW 2 */ + 363 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 565 24 + 364 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 365 "00010001" // /* MW 3 */ + 366 "00011111" // /* MW 2 */ + 367 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 567 18 first + 368 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 369 "00110001" // /* MW 3 */ + 370 "00011111" // /* MW 2 */ + 371 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 567 18 + 372 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 373 "00110001" // /* MW 3 */ + 374 "00011111" // /* MW 2 */ + 375 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 566 27 first + 376 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 377 "00110001" // /* MW 3 */ + 378 "00011111" // /* MW 2 */ + 379 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 564 18 first + 380 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 381 "00010001" // /* MW 3 */ + 382 "00011111" // /* MW 2 */ + 383 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 565 24 first + 384 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 385 "00010001" // /* MW 3 */ + 386 "00011111" // /* MW 2 */ + 387 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 565 24 + 388 "10011000" // ST r24, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 389 "00010001" // /* MW 3 */ + 390 "00011111" // /* MW 2 */ + 391 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 567 18 first + 392 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 393 "00110001" // /* MW 3 */ + 394 "00011111" // /* MW 2 */ + 395 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 567 18 + 396 "10011000" // ST r25, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 397 "00110001" // /* MW 3 */ + 398 "00011111" // /* MW 2 */ + 399 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 566 27 first + 400 "10011000" // ST r25, [p1], m7 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 401 "00110001" // /* MW 3 */ + 402 "11101011" // /* MW 2 */ + 403 "00001001" // /* MW 1 */ + 404 "00110110" // NOPA; NOPB; ST p1, [sp, #-12]; NOPX /* MW 12 */ /* control_operation: words=12 cycles_taken=1 */ + 405 "11000001" // /* MW 11 */ + 406 "01001110" // /* MW 10 */ + 407 "11111010" // /* MW 9 */ + 408 "00000011" // /* MW 8 */ + 409 "00000000" // /* MW 7 */ + 410 "00000000" // /* MW 6 */ + 411 "00100000" // /* MW 5 */ + 412 "00000000" // /* MW 4 */ + 413 "11110000" // /* MW 3 */ + 414 "00101100" // /* MW 2 */ + 415 "00000000" // /* MW 1 */ +.label TGT_F_main_192 +.src_ref 2 "0_0.cc" 29 31 first +.src_ref 2 "0_0.cc" 37 12 first +.loop_nesting 2 + 416 "10111010" // LDA r17, [p7], #4; MOVXM ls, #496 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 417 "00010000" // /* MW 9 */ + 418 "11111000" // /* MW 8 */ + 419 "01111000" // /* MW 7 */ + 420 "00000000" // /* MW 6 */ + 421 "00000000" // /* MW 5 */ + 422 "00000000" // /* MW 4 */ + 423 "11010000" // /* MW 3 */ + 424 "11000110" // /* MW 2 */ + 425 "11100011" // /* MW 1 */ +.src_ref 2 "0_0.cc" 30 36 first +.src_ref 2 "0_0.cc" 37 12 + 426 "10111010" // LDA r10, [p7], #4; MOVXM le, #592 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 427 "00010000" // /* MW 9 */ + 428 "00101000" // /* MW 8 */ + 429 "10111001" // /* MW 7 */ + 430 "00000001" // /* MW 6 */ + 431 "00000000" // /* MW 5 */ + 432 "00000000" // /* MW 4 */ + 433 "11010000" // /* MW 3 */ + 434 "10101010" // /* MW 2 */ + 435 "11100011" // /* MW 1 */ +.src_ref 2 "0_0.cc" 31 37 first + 436 "10011000" // LDA r9, [p7], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 437 "00110110" // /* MW 3 */ + 438 "00011101" // /* MW 2 */ + 439 "00000111" // /* MW 1 */ +.src_ref 2 "0_0.cc" 32 37 first + 440 "10011000" // LDA r8, [p7], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 441 "00010110" // /* MW 3 */ + 442 "00011101" // /* MW 2 */ + 443 "00000111" // /* MW 1 */ +.src_ref 2 "0_0.cc" 33 38 first + 444 "10011000" // LDA r19, [p7], #8 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 445 "01110110" // /* MW 3 */ + 446 "00101110" // /* MW 2 */ + 447 "00000111" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 572 18 +.src_ref 2 "0_0.cc" 40 79 +.src_ref 2 "0_0.cc" 40 86 + 448 "11010100" // LDA p1, [sp, #-4]; MOV p2, p7 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 449 "10000001" // /* MW 5 */ + 450 "11011101" // /* MW 4 */ + 451 "00100100" // /* MW 3 */ + 452 "10010011" // /* MW 2 */ + 453 "11111111" // /* MW 1 */ + 454 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 455 "00000000" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 572 18 +.src_ref 4 "io_buffer_compiler.h" 575 18 +.src_ref 2 "0_0.cc" 58 20 + 456 "11100100" // MOVX r16, #0; MOV el7, r24 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 457 "00111001" // /* MW 5 */ + 458 "00110001" // /* MW 4 */ + 459 "00100111" // /* MW 3 */ + 460 "00000000" // /* MW 2 */ + 461 "00000100" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 575 18 + 462 "11111000" // MOV el9, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 463 "10011100" // /* MW 3 */ + 464 "10010000" // /* MW 2 */ + 465 "00011100" // /* MW 1 */ + 466 "01011000" // ADD.NC r20, r9, r10 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 467 "10101001" // /* MW 3 */ + 468 "00010100" // /* MW 2 */ + 469 "00011101" // /* MW 1 */ +.src_ref 2 "0_0.cc" 41 30 + 470 "10111010" // NOPA; MOVS p0, p6; ADD.NC r20, r20, r8 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 471 "10100010" // /* MW 9 */ + 472 "00010000" // /* MW 8 */ + 473 "10001101" // /* MW 7 */ + 474 "00000010" // /* MW 6 */ + 475 "10001011" // /* MW 5 */ + 476 "10011000" // /* MW 4 */ + 477 "11110000" // /* MW 3 */ + 478 "00101100" // /* MW 2 */ + 479 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 37 12 first +.src_ref 2 "0_0.cc" 42 24 +.src_ref 2 "0_0.cc" 43 20 + 480 "11100001" // MOVA r18, #6; NOPB; NOPS; MOVX r19, #0; ADD.NC lc, r19, r20; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 481 "00000000" // /* MW 15 */ + 482 "00000000" // /* MW 14 */ + 483 "10101000" // /* MW 13 */ + 484 "11101000" // /* MW 12 */ + 485 "10111100" // /* MW 11 */ + 486 "00001010" // /* MW 10 */ + 487 "00110000" // /* MW 9 */ + 488 "00000001" // /* MW 8 */ + 489 "01011011" // /* MW 7 */ + 490 "00000001" // /* MW 6 */ + 491 "00100000" // /* MW 5 */ + 492 "00000000" // /* MW 4 */ + 493 "00000000" // /* MW 3 */ + 494 "11010010" // /* MW 2 */ + 495 "00000000" // /* MW 1 */ +.label ZLS_F_main_272 +.src_ref 4 "io_buffer_compiler.h" 572 18 first +.src_ref 2 "0_0.cc" 40 79 first +.src_ref 2 "0_0.cc" 40 86 first +.src_ref 2 "0_0.cc" 42 24 first +.begin_of_loop +.loop_nesting 3 + 496 "10111010" // LDA dn6, [p2], #4; ST el7, [p1], #4; ADD.NC r19, r19, #1 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 497 "01000010" // /* MW 9 */ + 498 "11000000" // /* MW 8 */ + 499 "01101100" // /* MW 7 */ + 500 "10000010" // /* MW 6 */ + 501 "11101001" // /* MW 5 */ + 502 "00011101" // /* MW 4 */ + 503 "11010001" // /* MW 3 */ + 504 "11100100" // /* MW 2 */ + 505 "01000011" // /* MW 1 */ +.src_ref 2 "0_0.cc" 40 98 +.src_ref 2 "0_0.cc" 43 20 first + 506 "00010100" // LDA r20, [p2], #4; ADD.NC r18, r18, #4 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 507 "00000100" // /* MW 5 */ + 508 "00110010" // /* MW 4 */ + 509 "11011001" // /* MW 3 */ + 510 "11010010" // /* MW 2 */ + 511 "01000011" // /* MW 1 */ +.src_ref 2 "0_0.cc" 40 110 first +.src_ref 2 "0_0.cc" 40 117 first + 512 "10011000" // LDA dc7, [p2], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 513 "11100110" // /* MW 3 */ + 514 "00011111" // /* MW 2 */ + 515 "00000010" // /* MW 1 */ +.src_ref 2 "0_0.cc" 40 129 + 516 "10011000" // LDA el11, [p2], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 517 "11101110" // /* MW 3 */ + 518 "00011110" // /* MW 2 */ + 519 "00000010" // /* MW 1 */ + 520 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 521 "00000000" // /* MW 1 */ + 522 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 523 "00000000" // /* MW 1 */ + 524 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 525 "00000000" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 573 24 first + 526 "10011000" // ST dn6, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 527 "00100001" // /* MW 3 */ + 528 "00011111" // /* MW 2 */ + 529 "00001001" // /* MW 1 */ + 530 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 531 "00000000" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 573 24 + 532 "10011000" // ST dc7, [p1], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 533 "11100001" // /* MW 3 */ + 534 "00011111" // /* MW 2 */ + 535 "00001001" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 575 18 first + 536 "00000010" // ST el9, [p1], #4; NOPM /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 537 "01110000" // /* MW 7 */ + 538 "10100101" // /* MW 6 */ + 539 "00000001" // /* MW 5 */ + 540 "00000000" // /* MW 4 */ + 541 "00110000" // /* MW 3 */ + 542 "11001101" // /* MW 2 */ + 543 "00100011" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 574 27 first + 544 "11100001" // NOPA; NOPB; ST r20, [p1], #4; NOPX; NOPM; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 545 "00000000" // /* MW 15 */ + 546 "00000000" // /* MW 14 */ + 547 "01111000" // /* MW 13 */ + 548 "10100101" // /* MW 12 */ + 549 "00000001" // /* MW 11 */ + 550 "00000000" // /* MW 10 */ + 551 "00000000" // /* MW 9 */ + 552 "10000000" // /* MW 8 */ + 553 "10010001" // /* MW 7 */ + 554 "00011110" // /* MW 6 */ + 555 "00100001" // /* MW 5 */ + 556 "00000000" // /* MW 4 */ + 557 "11110000" // /* MW 3 */ + 558 "00101100" // /* MW 2 */ + 559 "00000000" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 574 27 + 560 "11100001" // NOPA; NOPB; ST el11, [p1], #-20; NOPX; NOPM; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 561 "00000000" // /* MW 15 */ + 562 "00000000" // /* MW 14 */ + 563 "01111000" // /* MW 13 */ + 564 "10100101" // /* MW 12 */ + 565 "00000001" // /* MW 11 */ + 566 "00000000" // /* MW 10 */ + 567 "00000000" // /* MW 9 */ + 568 "10000000" // /* MW 8 */ + 569 "11101001" // /* MW 7 */ + 570 "10111110" // /* MW 6 */ + 571 "00100001" // /* MW 5 */ + 572 "00000000" // /* MW 4 */ + 573 "11110000" // /* MW 3 */ + 574 "00101100" // /* MW 2 */ + 575 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 41 30 + 576 "11100001" // NOPA; NOPB; NOPS; NOPX; MOV r21, p1; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 577 "00000000" // /* MW 15 */ + 578 "00000000" // /* MW 14 */ + 579 "01111000" // /* MW 13 */ + 580 "01100000" // /* MW 12 */ + 581 "10101001" // /* MW 11 */ + 582 "00000010" // /* MW 10 */ + 583 "00000000" // /* MW 9 */ + 584 "00000000" // /* MW 8 */ + 585 "01011011" // /* MW 7 */ + 586 "00000001" // /* MW 6 */ + 587 "00100000" // /* MW 5 */ + 588 "00000000" // /* MW 4 */ + 589 "11110000" // /* MW 3 */ + 590 "00101100" // /* MW 2 */ + 591 "00000000" // /* MW 1 */ +.label ZLE_F_main_368 +.src_ref 2 "0_0.cc" 41 30 first +.end_of_loop + 592 "11100001" // NOPA; NOPB; ST r21, [p0], #4; NOPX; ADD.NC p1, r21, #24; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 593 "00000000" // /* MW 15 */ + 594 "00000000" // /* MW 14 */ + 595 "00001000" // /* MW 13 */ + 596 "01000110" // /* MW 12 */ + 597 "10110101" // /* MW 11 */ + 598 "00000000" // /* MW 10 */ + 599 "00000000" // /* MW 9 */ + 600 "10000000" // /* MW 8 */ + 601 "10110001" // /* MW 7 */ + 602 "00011110" // /* MW 6 */ + 603 "00100000" // /* MW 5 */ + 604 "00000000" // /* MW 4 */ + 605 "11110000" // /* MW 3 */ + 606 "00101100" // /* MW 2 */ + 607 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 70 32 +.loop_nesting 2 + 608 "10111010" // MOVA r11, #0; LSHL r20, r19, r12; MOV r19, p7 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 609 "01111000" // /* MW 9 */ + 610 "01100000" // /* MW 8 */ + 611 "01101111" // /* MW 7 */ + 612 "01101110" // /* MW 6 */ + 613 "01000110" // /* MW 5 */ + 614 "00100111" // /* MW 4 */ + 615 "00000000" // /* MW 3 */ + 616 "00001011" // /* MW 2 */ + 617 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 50 40 + 618 "00100100" // ADD r13, r19, #-24; ADD.NC r19, r18, #1 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 619 "00000001" // /* MW 5 */ + 620 "10110010" // /* MW 4 */ + 621 "01101001" // /* MW 3 */ + 622 "01110100" // /* MW 2 */ + 623 "10011011" // /* MW 1 */ +.src_ref 2 "0_0.cc" 49 38 +.src_ref 2 "0_0.cc" 49 38 first +.src_ref 2 "0_0.cc" 50 30 + 624 "00111010" // MOVS p7, r13; LSHL r18, r18, r12; MOV dj6, r20 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 625 "01111001" // /* MW 9 */ + 626 "00010000" // /* MW 8 */ + 627 "01000101" // /* MW 7 */ + 628 "01101111" // /* MW 6 */ + 629 "00100110" // /* MW 5 */ + 630 "00100101" // /* MW 4 */ + 631 "01100000" // /* MW 3 */ + 632 "10100001" // /* MW 2 */ + 633 "11110001" // /* MW 1 */ +.src_ref 2 "0_0.cc" 49 38 +.src_ref 2 "0_0.cc" 50 33 first + 634 "11100100" // LSHL r18, r19, r12; MOV dj5, r18 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 635 "01000001" // /* MW 5 */ + 636 "00010010" // /* MW 4 */ + 637 "10111011" // /* MW 3 */ + 638 "10011001" // /* MW 2 */ + 639 "10011100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 49 38 first +.src_ref 2 "0_0.cc" 50 33 + 640 "10010100" // LDA r18, [p7, dj5]; ADD.NC dn7, r13, r18 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 641 "10010010" // /* MW 5 */ + 642 "10001101" // /* MW 4 */ + 643 "11011110" // /* MW 3 */ + 644 "01001010" // /* MW 2 */ + 645 "11110100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 50 30 first + 646 "00000010" // ST dn7, [p6, dj6]; ADD.NC r17, r17, #-1 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 647 "11000000" // /* MW 7 */ + 648 "01111111" // /* MW 6 */ + 649 "00101100" // /* MW 5 */ + 650 "00000010" // /* MW 4 */ + 651 "00110000" // /* MW 3 */ + 652 "01110100" // /* MW 2 */ + 653 "11011000" // /* MW 1 */ + 654 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 655 "00000000" // /* MW 1 */ + 656 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 657 "00000000" // /* MW 1 */ + 658 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 659 "00000000" // /* MW 1 */ + 660 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 661 "00000000" // /* MW 1 */ + 662 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 663 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 52 20 first + 664 "00000010" // NOPS; ADD.NC r18, r19, r18 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 665 "10100000" // /* MW 7 */ + 666 "11100100" // /* MW 6 */ + 667 "01001100" // /* MW 5 */ + 668 "00000010" // /* MW 4 */ + 669 "01100000" // /* MW 3 */ + 670 "00101011" // /* MW 2 */ + 671 "00000000" // /* MW 1 */ +.label TGT_F_main_448 +.src_ref 2 "0_0.cc" 58 20 first +.loop_nesting 3 + 672 "10011000" // SUB r16, r18, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 673 "00000001" // /* MW 3 */ + 674 "10100001" // /* MW 2 */ + 675 "00010100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 59 63 first + 676 "10011000" // LSHL r15, r16, r12 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 677 "11001101" // /* MW 3 */ + 678 "00011110" // /* MW 2 */ + 679 "00010100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 59 63 +.src_ref 2 "0_0.cc" 59 63 +.src_ref 2 "0_0.cc" 60 40 +.src_ref 2 "0_0.cc" 63 43 +.src_ref 2 "0_0.cc" 64 63 +.src_ref 2 "0_0.cc" 65 40 +.src_ref 2 "0_0.cc" 67 46 + 680 "00000010" // MOVS p7, r13; MOV dj7, r15 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 681 "01110000" // /* MW 7 */ + 682 "11010000" // /* MW 6 */ + 683 "11000011" // /* MW 5 */ + 684 "00000011" // /* MW 4 */ + 685 "01100000" // /* MW 3 */ + 686 "10100001" // /* MW 2 */ + 687 "11110001" // /* MW 1 */ +.src_ref 2 "0_0.cc" 59 63 + 688 "10011000" // LDA r0, [p7, dj7] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 689 "00010110" // /* MW 3 */ + 690 "11100000" // /* MW 2 */ + 691 "00000111" // /* MW 1 */ +.src_ref 2 "0_0.cc" 59 16 +.no_stack_arguments + 692 "00000100" // JL #1344 /* MW 6 */ /* control_operation: words=6 call unconditional cycles_taken=1 direct absolute target_address=1344 delay_slots=5 */ + 693 "00000001" // /* MW 5 */ + 694 "00000000" // /* MW 4 */ + 695 "10100000" // /* MW 3 */ + 696 "00000010" // /* MW 2 */ + 697 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 59 57 +.delay_slot + 698 "10011000" // ADD.NC r14, r16, #1 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 699 "00000000" // /* MW 3 */ + 700 "10011000" // /* MW 2 */ + 701 "00011011" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 702 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 703 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 59 50 +.delay_slot + 704 "10011000" // LSHL r16, r14, r12 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 705 "11001101" // /* MW 3 */ + 706 "10100000" // /* MW 2 */ + 707 "00010011" // /* MW 1 */ +.src_ref 2 "0_0.cc" 59 50 +.delay_slot + 708 "01011000" // ADD.NC p0, r13, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 709 "11000001" // /* MW 3 */ + 710 "01100110" // /* MW 2 */ + 711 "00011000" // /* MW 1 */ +.delay_slot + 712 "00000010" // ST r17, [sp, #-16]; NOPM /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 713 "01110000" // /* MW 7 */ + 714 "10100101" // /* MW 6 */ + 715 "00000001" // /* MW 5 */ + 716 "00000000" // /* MW 4 */ + 717 "10110000" // /* MW 3 */ + 718 "01000110" // /* MW 2 */ + 719 "11111110" // /* MW 1 */ +.src_ref 2 "0_0.cc" 60 40 +.src_ref 2 "0_0.cc" 60 49 +.return_address + 720 "11100100" // MOVX r17, #1; MOV dj0, r15 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 721 "01000001" // /* MW 5 */ + 722 "00001111" // /* MW 4 */ + 723 "10100001" // /* MW 3 */ + 724 "01000000" // /* MW 2 */ + 725 "00000100" // /* MW 1 */ +.src_ref 3 "tile_control.h" 147 12 +.src_ref 2 "0_0.cc" 60 40 first + 726 "10111010" // LDA r18, [p7, dj0]; MOVXM r20, #30656 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 727 "00010000" // /* MW 9 */ + 728 "11100000" // /* MW 8 */ + 729 "10001011" // /* MW 7 */ + 730 "00011110" // /* MW 6 */ + 731 "00000000" // /* MW 5 */ + 732 "00000000" // /* MW 4 */ + 733 "11010000" // /* MW 3 */ + 734 "01001010" // /* MW 2 */ + 735 "11100000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 147 12 + 736 "01000100" // MOVXM r21, #30658 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 737 "10000100" // /* MW 5 */ + 738 "10101111" // /* MW 4 */ + 739 "01111010" // /* MW 3 */ + 740 "00000000" // /* MW 2 */ + 741 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 147 12 +.src_ref 3 "tile_control.h" 260 15 + 742 "01000100" // MOVXM p0, #524288 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 743 "00000000" // /* MW 5 */ + 744 "11000000" // /* MW 4 */ + 745 "00000000" // /* MW 3 */ + 746 "00001000" // /* MW 2 */ + 747 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 440 26 + 748 "01000100" // MOVXM r16, #7340035 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 749 "00000110" // /* MW 5 */ + 750 "00100000" // /* MW 4 */ + 751 "00001000" // /* MW 3 */ + 752 "01110000" // /* MW 2 */ + 753 "00000000" // /* MW 1 */ + 754 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 755 "00000000" // /* MW 1 */ + 756 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 757 "00000000" // /* MW 1 */ + 758 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 759 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 60 49 + 760 "10011000" // LSHL r17, r18, r17 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 761 "00011101" // /* MW 3 */ + 762 "10100011" // /* MW 2 */ + 763 "00010100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 61 20 first + 764 "01011000" // ADD.NC r18, r17, r14 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 765 "10111001" // /* MW 3 */ + 766 "10011000" // /* MW 2 */ + 767 "00011100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 63 43 first + 768 "10011000" // LSHL r19, r18, r12 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 769 "11001101" // /* MW 3 */ + 770 "10100110" // /* MW 2 */ + 771 "00010100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 63 43 + 772 "11111000" // MOV dj0, r19 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 773 "10100000" // /* MW 3 */ + 774 "10001001" // /* MW 2 */ + 775 "00011000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 63 43 + 776 "10011000" // LDA r22, [p7, dj0] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 777 "11010110" // /* MW 3 */ + 778 "00000010" // /* MW 2 */ + 779 "00000111" // /* MW 1 */ + 780 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 781 "00000000" // /* MW 1 */ + 782 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 783 "00000000" // /* MW 1 */ + 784 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 785 "00000000" // /* MW 1 */ + 786 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 787 "00000000" // /* MW 1 */ + 788 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 789 "00000000" // /* MW 1 */ + 790 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 791 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 147 14 first + 792 "10011000" // LTU r27, r22, r12 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 793 "11001100" // /* MW 3 */ + 794 "10110110" // /* MW 2 */ + 795 "00010101" // /* MW 1 */ +.src_ref 3 "tile_control.h" 147 12 + 796 "00011000" // SEL.EQZ r20, r21, r20, r27 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 797 "01000010" // /* MW 3 */ + 798 "01101001" // /* MW 2 */ + 799 "00010101" // /* MW 1 */ +.src_ref 3 "tile_control.h" 147 12 + 800 "01011000" // ADD.NC r20, r22, r20 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 801 "01010001" // /* MW 3 */ + 802 "00011011" // /* MW 2 */ + 803 "00011101" // /* MW 1 */ +.src_ref 3 "tile_control.h" 147 12 + 804 "10011000" // LSHL r20, r20, r12 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 805 "11001101" // /* MW 3 */ + 806 "00101000" // /* MW 2 */ + 807 "00010101" // /* MW 1 */ +.src_ref 3 "tile_control.h" 147 12 +.src_ref 3 "tile_control.h" 260 15 + 808 "00000010" // NOPS; MOV dj0, r20 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 809 "01110000" // /* MW 7 */ + 810 "00010000" // /* MW 6 */ + 811 "01000101" // /* MW 5 */ + 812 "00000000" // /* MW 4 */ + 813 "01100000" // /* MW 3 */ + 814 "00101011" // /* MW 2 */ + 815 "00000000" // /* MW 1 */ +.label TGT_F_main_592 +.src_ref 3 "tile_control.h" 147 12 +.src_ref 3 "tile_control.h" 260 15 first +.loop_nesting 4 + 816 "10011000" // LDA.TM r20, [p0, dj0] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 817 "10010011" // /* MW 3 */ + 818 "00000010" // /* MW 2 */ + 819 "00000000" // /* MW 1 */ + 820 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 821 "00000000" // /* MW 1 */ + 822 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 823 "00000000" // /* MW 1 */ + 824 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 825 "00000000" // /* MW 1 */ + 826 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 827 "00000000" // /* MW 1 */ + 828 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 829 "00000000" // /* MW 1 */ + 830 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 831 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 440 26 first + 832 "10011000" // AND r21, r20, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 833 "00000100" // /* MW 3 */ + 834 "00101011" // /* MW 2 */ + 835 "00010101" // /* MW 1 */ +.src_ref 3 "tile_control.h" 440 8 +.src_ref 3 "tile_control.h" 440 61 + 836 "10000100" // JNZ r21, #816 /* MW 6 */ /* control_operation: words=6 jump conditional cycles_taken=1 cycles_not_taken=0 direct absolute target_address=816 delay_slots=5 */ + 837 "00000001" // /* MW 5 */ + 838 "01000000" // /* MW 4 */ + 839 "10011000" // /* MW 3 */ + 840 "00000001" // /* MW 2 */ + 841 "10101000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 842 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 843 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 844 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 845 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 846 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 847 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 848 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 849 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 850 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 851 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 63 50 +.loop_nesting 3 + 852 "10011000" // ADD.NC r15, r18, #1 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 853 "00000000" // /* MW 3 */ + 854 "11011001" // /* MW 2 */ + 855 "00011011" // /* MW 1 */ +.src_ref 2 "0_0.cc" 64 63 first + 856 "10011000" // LSHL r14, r15, r12 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 857 "11001101" // /* MW 3 */ + 858 "11011100" // /* MW 2 */ + 859 "00010011" // /* MW 1 */ +.src_ref 2 "0_0.cc" 64 63 + 860 "11111000" // MOV dj1, r14 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 861 "00100000" // /* MW 3 */ + 862 "10000111" // /* MW 2 */ + 863 "00011001" // /* MW 1 */ +.src_ref 2 "0_0.cc" 64 63 + 864 "10011000" // LDA r0, [p7, dj1] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 865 "00010110" // /* MW 3 */ + 866 "00100000" // /* MW 2 */ + 867 "00000111" // /* MW 1 */ +.src_ref 2 "0_0.cc" 64 16 +.no_stack_arguments + 868 "00000100" // JL #1344 /* MW 6 */ /* control_operation: words=6 call unconditional cycles_taken=1 direct absolute target_address=1344 delay_slots=5 */ + 869 "00000001" // /* MW 5 */ + 870 "00000000" // /* MW 4 */ + 871 "10100000" // /* MW 3 */ + 872 "00000010" // /* MW 2 */ + 873 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 874 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 875 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 876 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 877 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 40 117 +.delay_slot + 878 "00011000" // ADD r13, r13, #8 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 879 "00100011" // /* MW 3 */ + 880 "01011010" // /* MW 2 */ + 881 "00010011" // /* MW 1 */ +.src_ref 2 "0_0.cc" 64 50 +.delay_slot + 882 "01011000" // ADD.NC p0, r19, r13 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 883 "10110101" // /* MW 3 */ + 884 "01101001" // /* MW 2 */ + 885 "00011000" // /* MW 1 */ +.delay_slot + 886 "01111010" // NOPA; ST r17, [sp, #-24]; NOPX /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 887 "00000000" // /* MW 9 */ + 888 "00000000" // /* MW 8 */ + 889 "00000000" // /* MW 7 */ + 890 "10000000" // /* MW 6 */ + 891 "00110101" // /* MW 5 */ + 892 "11101010" // /* MW 4 */ + 893 "11110111" // /* MW 3 */ + 894 "00101100" // /* MW 2 */ + 895 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 65 40 +.src_ref 2 "0_0.cc" 65 49 +.src_ref 2 "0_0.cc" 65 53 +.src_ref 2 "0_0.cc" 70 32 +.return_address + 896 "10111010" // MOVA r14, #0; MOVX r16, #1; MOV dj0, r14 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 897 "01111000" // /* MW 9 */ + 898 "10010000" // /* MW 8 */ + 899 "01000011" // /* MW 7 */ + 900 "00101000" // /* MW 6 */ + 901 "00000000" // /* MW 5 */ + 902 "00000001" // /* MW 4 */ + 903 "00000000" // /* MW 3 */ + 904 "00001110" // /* MW 2 */ + 905 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 40 98 +.src_ref 2 "0_0.cc" 65 40 first + 906 "00010100" // LDA r18, [p7, dj0]; ADD.NC r17, r13, #-4 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 907 "11111100" // /* MW 5 */ + 908 "10101101" // /* MW 4 */ + 909 "11011000" // /* MW 3 */ + 910 "01001010" // /* MW 2 */ + 911 "11100000" // /* MW 1 */ + 912 "10011000" // ST r17, [sp, #-20] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 913 "00110101" // /* MW 3 */ + 914 "11101110" // /* MW 2 */ + 915 "00001111" // /* MW 1 */ + 916 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 917 "00000000" // /* MW 1 */ + 918 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 919 "00000000" // /* MW 1 */ + 920 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 921 "00000000" // /* MW 1 */ + 922 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 923 "00000000" // /* MW 1 */ + 924 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 925 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 65 49 + 926 "10011000" // LSHL r18, r18, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 927 "00001101" // /* MW 3 */ + 928 "10100101" // /* MW 2 */ + 929 "00010100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 65 53 + 930 "10011000" // OR r16, r18, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 931 "00000101" // /* MW 3 */ + 932 "10100001" // /* MW 2 */ + 933 "00010100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 66 20 first + 934 "01011000" // ADD.NC r18, r15, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 935 "11000001" // /* MW 3 */ + 936 "10010111" // /* MW 2 */ + 937 "00011100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 67 46 +.src_ref 2 "0_0.cc" 67 46 first + 938 "00111010" // ST r16, [sp, #-32]; LSHL r12, r18, r12; MOV p0, p7 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 939 "01111001" // /* MW 9 */ + 940 "01100000" // /* MW 8 */ + 941 "00110111" // /* MW 7 */ + 942 "01101100" // /* MW 6 */ + 943 "11000110" // /* MW 5 */ + 944 "00100100" // /* MW 4 */ + 945 "10110000" // /* MW 3 */ + 946 "01000010" // /* MW 2 */ + 947 "11111100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 67 46 +.src_ref 2 "0_0.cc" 67 46 + 948 "00111010" // ST r18, [sp, #-28]; ADD r16, r17, r12; MOV dj0, r12 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 949 "01111001" // /* MW 9 */ + 950 "00010000" // /* MW 8 */ + 951 "01000011" // /* MW 7 */ + 952 "00000100" // /* MW 6 */ + 953 "00000110" // /* MW 5 */ + 954 "00100011" // /* MW 4 */ + 955 "10110000" // /* MW 3 */ + 956 "11001010" // /* MW 2 */ + 957 "11111100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 67 16 +.src_ref 2 "0_0.cc" 67 46 + 958 "01110110" // LDA r16, [p0, dj0]; ST r16, [sp, #-36]; MOVXM p7, #992 /* MW 12 */ /* control_operation: words=12 cycles_taken=1 */ + 959 "00010000" // /* MW 11 */ + 960 "11110000" // /* MW 10 */ + 961 "10110001" // /* MW 9 */ + 962 "00000011" // /* MW 8 */ + 963 "00000000" // /* MW 7 */ + 964 "10000000" // /* MW 6 */ + 965 "00010101" // /* MW 5 */ + 966 "11011110" // /* MW 4 */ + 967 "11010111" // /* MW 3 */ + 968 "01000010" // /* MW 2 */ + 969 "00000000" // /* MW 1 */ + 970 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 971 "00000000" // /* MW 1 */ + 972 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 973 "00000000" // /* MW 1 */ + 974 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 975 "00000000" // /* MW 1 */ + 976 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 977 "00000000" // /* MW 1 */ + 978 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 979 "00000000" // /* MW 1 */ + 980 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 981 "00000000" // /* MW 1 */ + 982 "10111010" // NOPA; NOPB; ADD.NC r15, r16, #-1 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 983 "11001110" // /* MW 9 */ + 984 "00111111" // /* MW 8 */ + 985 "11101100" // /* MW 7 */ + 986 "00000001" // /* MW 6 */ + 987 "00010000" // /* MW 5 */ + 988 "00000000" // /* MW 4 */ + 989 "11110000" // /* MW 3 */ + 990 "00101100" // /* MW 2 */ + 991 "00000000" // /* MW 1 */ +.label TGT_F_main_768 +.src_ref 2 "0_0.cc" 70 32 first +.loop_nesting 4 + 992 "10011000" // OR r16, r11, r14 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 993 "11100101" // /* MW 3 */ + 994 "11100000" // /* MW 2 */ + 995 "00010010" // /* MW 1 */ +.src_ref 2 "0_0.cc" 70 32 + 996 "10000100" // JNZ r16, #1104 /* MW 6 */ /* control_operation: words=6 jump conditional cycles_taken=1 cycles_not_taken=0 direct absolute target_address=1104 delay_slots=5 */ + 997 "00000001" // /* MW 5 */ + 998 "01000000" // /* MW 4 */ + 999 "00101000" // /* MW 3 */ + 1000 "00000010" // /* MW 2 */ + 1001 "10000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1002 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1003 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1004 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1005 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1006 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1007 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1008 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1009 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 67 46 +.src_ref 2 "0_0.cc" 67 46 +.src_ref 2 "0_0.cc" 72 40 +.src_ref 2 "0_0.cc" 72 40 +.delay_slot + 1010 "00000010" // MOVS p0, r13; MOV dj1, r12 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1011 "01110000" // /* MW 7 */ + 1012 "00010000" // /* MW 6 */ + 1013 "11000011" // /* MW 5 */ + 1014 "00000000" // /* MW 4 */ + 1015 "01100000" // /* MW 3 */ + 1016 "10100001" // /* MW 2 */ + 1017 "00010001" // /* MW 1 */ +.src_ref 2 "0_0.cc" 70 47 + 1018 "00011000" // LDA p1, [sp, #-36] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1019 "10011001" // /* MW 3 */ + 1020 "11011100" // /* MW 2 */ + 1021 "00000111" // /* MW 1 */ + 1022 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1023 "00000000" // /* MW 1 */ + 1024 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1025 "00000000" // /* MW 1 */ + 1026 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1027 "00000000" // /* MW 1 */ + 1028 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1029 "00000000" // /* MW 1 */ + 1030 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1031 "00000000" // /* MW 1 */ + 1032 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1033 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 70 47 + 1034 "10011000" // LDA r16, [p1] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1035 "00010110" // /* MW 3 */ + 1036 "00000110" // /* MW 2 */ + 1037 "00000001" // /* MW 1 */ + 1038 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1039 "00000000" // /* MW 1 */ + 1040 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1041 "00000000" // /* MW 1 */ + 1042 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1043 "00000000" // /* MW 1 */ + 1044 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1045 "00000000" // /* MW 1 */ + 1046 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1047 "00000000" // /* MW 1 */ + 1048 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1049 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 70 23 +.src_ref 2 "0_0.cc" 70 47 + 1050 "10000100" // JZ r16, #1104 /* MW 6 */ /* control_operation: words=6 jump conditional cycles_taken=1 cycles_not_taken=0 direct absolute target_address=1104 delay_slots=5 */ + 1051 "00000001" // /* MW 5 */ + 1052 "00000000" // /* MW 4 */ + 1053 "00101000" // /* MW 3 */ + 1054 "00000010" // /* MW 2 */ + 1055 "10000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1056 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1057 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1058 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1059 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1060 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1061 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1062 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1063 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1064 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1065 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1066 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1067 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1068 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1069 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1070 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1071 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1072 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1073 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1074 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1075 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1076 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1077 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 71 24 first + 1078 "00011000" // DONE /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1079 "00000000" // /* MW 3 */ + 1080 "00001000" // /* MW 2 */ + 1081 "00010000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1082 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1083 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1084 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1085 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1086 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1087 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1088 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1089 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1090 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1091 "00000000" // /* MW 1 */ +.swstall chess_separator_scheduler + 1092 "00110110" // NOPA; NOPB; NOPS; NOPX /* MW 12 */ /* control_operation: words=12 cycles_taken=1 */ + 1093 "10000001" // /* MW 11 */ + 1094 "10101101" // /* MW 10 */ + 1095 "00000000" // /* MW 9 */ + 1096 "00000000" // /* MW 8 */ + 1097 "00000000" // /* MW 7 */ + 1098 "00000000" // /* MW 6 */ + 1099 "00100000" // /* MW 5 */ + 1100 "00000000" // /* MW 4 */ + 1101 "11110000" // /* MW 3 */ + 1102 "00101100" // /* MW 2 */ + 1103 "00000000" // /* MW 1 */ +.label TGT_F_main_880 +.src_ref 2 "0_0.cc" 67 46 first +.src_ref 2 "0_0.cc" 72 20 +.src_ref 2 "0_0.cc" 72 40 first + 1104 "11010100" // LDA r0, [p0, dj1]; MOV p0, p6 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1105 "10000001" // /* MW 5 */ + 1106 "11011001" // /* MW 4 */ + 1107 "11010000" // /* MW 3 */ + 1108 "00000010" // /* MW 2 */ + 1109 "00000100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 72 20 +.no_stack_arguments + 1110 "00000100" // JL #2352 /* MW 6 */ /* control_operation: words=6 call unconditional cycles_taken=1 direct absolute target_address=2352 delay_slots=5 */ + 1111 "00000001" // /* MW 5 */ + 1112 "00000000" // /* MW 4 */ + 1113 "10011000" // /* MW 3 */ + 1114 "00000100" // /* MW 2 */ + 1115 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1116 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1117 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1118 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1119 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 72 20 +.delay_slot + 1120 "11111000" // MOV r1, r10 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1121 "00100000" // /* MW 3 */ + 1122 "01010101" // /* MW 2 */ + 1123 "00011000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 72 20 +.delay_slot + 1124 "11111000" // MOV r2, r9 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1125 "10100000" // /* MW 3 */ + 1126 "10010100" // /* MW 2 */ + 1127 "00011000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 72 20 +.delay_slot + 1128 "00000010" // NOPS; MOV r3, r8 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1129 "01110000" // /* MW 7 */ + 1130 "00010000" // /* MW 6 */ + 1131 "01101010" // /* MW 5 */ + 1132 "00000000" // /* MW 4 */ + 1133 "01100000" // /* MW 3 */ + 1134 "00101011" // /* MW 2 */ + 1135 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 67 16 +.return_address + 1136 "00011000" // JNZD r15, r15, p7 /* MW 4 */ /* control_operation: words=4 jump conditional cycles_taken=1 cycles_not_taken=0 indirect absolute delay_slots=5 */ + 1137 "11100000" // /* MW 3 */ + 1138 "11011111" // /* MW 2 */ + 1139 "00010011" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1140 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1141 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1142 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1143 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1144 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1145 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1146 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1147 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 67 76 +.delay_slot + 1148 "10011000" // ADD.NC r14, r14, #1 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1149 "00000000" // /* MW 3 */ + 1150 "10010111" // /* MW 2 */ + 1151 "00011011" // /* MW 1 */ +.src_ref 3 "tile_control.h" 147 12 +.src_ref 3 "tile_control.h" 147 14 +.src_ref 2 "0_0.cc" 49 38 +.src_ref 2 "0_0.cc" 50 33 +.src_ref 2 "0_0.cc" 55 12 +.src_ref 2 "0_0.cc" 55 64 +.src_ref 2 "0_0.cc" 59 50 +.src_ref 2 "0_0.cc" 59 63 +.src_ref 2 "0_0.cc" 63 43 +.src_ref 2 "0_0.cc" 64 63 +.src_ref 2 "0_0.cc" 67 46 +.src_ref 2 "0_0.cc" 78 31 +.loop_nesting 3 + 1152 "10111010" // LDA r17, [sp, #-16]; MOVX r12, #2; ADD.NC r11, r11, #1 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1153 "01001000" // /* MW 9 */ + 1154 "11000000" // /* MW 8 */ + 1155 "01101010" // /* MW 7 */ + 1156 "01001001" // /* MW 6 */ + 1157 "11000000" // /* MW 5 */ + 1158 "00000000" // /* MW 4 */ + 1159 "00100000" // /* MW 3 */ + 1160 "01000110" // /* MW 2 */ + 1161 "11111110" // /* MW 1 */ +.src_ref 2 "0_0.cc" 55 12 + 1162 "10111010" // LDA r18, [sp, #-20]; MOVXM p7, #672 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1163 "00010000" // /* MW 9 */ + 1164 "01010000" // /* MW 8 */ + 1165 "10110001" // /* MW 7 */ + 1166 "00000011" // /* MW 6 */ + 1167 "00000000" // /* MW 5 */ + 1168 "00000000" // /* MW 4 */ + 1169 "00100000" // /* MW 3 */ + 1170 "11001010" // /* MW 2 */ + 1171 "11111101" // /* MW 1 */ +.src_ref 2 "0_0.cc" 74 20 first + 1172 "00011000" // LDA r19, [sp, #-28] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1173 "01110001" // /* MW 3 */ + 1174 "11100110" // /* MW 2 */ + 1175 "00000111" // /* MW 1 */ +.src_ref 2 "0_0.cc" 75 43 + 1176 "00011000" // LDA r16, [sp, #-32] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1177 "00010001" // /* MW 3 */ + 1178 "11100010" // /* MW 2 */ + 1179 "00000111" // /* MW 1 */ +.src_ref 2 "0_0.cc" 75 43 + 1180 "00011000" // LDA r20, [sp, #-24] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1181 "10010001" // /* MW 3 */ + 1182 "11101010" // /* MW 2 */ + 1183 "00000111" // /* MW 1 */ + 1184 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1185 "00000000" // /* MW 1 */ + 1186 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1187 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 55 12 first + 1188 "00011000" // JNZD r17, r17, p7 /* MW 4 */ /* control_operation: words=4 jump conditional cycles_taken=1 cycles_not_taken=0 indirect absolute delay_slots=5 */ + 1189 "11100000" // /* MW 3 */ + 1190 "01100011" // /* MW 2 */ + 1191 "00010100" // /* MW 1 */ +.delay_slot + 1192 "00011000" // ADD r13, r18, #-4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1193 "11110011" // /* MW 3 */ + 1194 "10011011" // /* MW 2 */ + 1195 "00010100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 74 20 +.delay_slot + 1196 "10011000" // ADD.NC r18, r19, #3 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1197 "10000001" // /* MW 3 */ + 1198 "10011001" // /* MW 2 */ + 1199 "00011100" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1200 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1201 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 75 43 first +.delay_slot + 1202 "01011000" // ADD.NC r16, r20, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1203 "01000001" // /* MW 3 */ + 1204 "00011010" // /* MW 2 */ + 1205 "00011100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 75 60 +.delay_slot + 1206 "10011000" // ADD.NC r16, r16, #5 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1207 "00000010" // /* MW 3 */ + 1208 "00011000" // /* MW 2 */ + 1209 "00011100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 78 31 +.src_ref 2 "0_0.cc" 78 31 first +.src_ref 2 "0_0.cc" 79 21 +.src_ref 2 "0_0.cc" 79 21 +.src_ref 2 "0_0.cc" 80 27 +.src_ref 2 "0_0.cc" 80 27 +.loop_nesting 2 + 1210 "01110110" // MOVA r18, #62; MOVS p7, r13; LSHL r16, r18, r12; MOV r20, #63 /* MW 12 */ /* control_operation: words=12 cycles_taken=1 */ + 1211 "01011000" // /* MW 11 */ + 1212 "00111111" // /* MW 10 */ + 1213 "10001000" // /* MW 9 */ + 1214 "01101110" // /* MW 8 */ + 1215 "00000110" // /* MW 7 */ + 1216 "00100101" // /* MW 6 */ + 1217 "00001011" // /* MW 5 */ + 1218 "10001101" // /* MW 4 */ + 1219 "00000111" // /* MW 3 */ + 1220 "11010010" // /* MW 2 */ + 1221 "00000111" // /* MW 1 */ +.src_ref 2 "0_0.cc" 78 31 +.src_ref 2 "0_0.cc" 79 12 +.src_ref 2 "0_0.cc" 80 12 + 1222 "10111010" // MOVA r16, #1; MOVX r22, #-1; MOV dj0, r16 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1223 "01111000" // /* MW 9 */ + 1224 "00010000" // /* MW 8 */ + 1225 "01000100" // /* MW 7 */ + 1226 "11101000" // /* MW 6 */ + 1227 "01100111" // /* MW 5 */ + 1228 "00111111" // /* MW 4 */ + 1229 "00000000" // /* MW 3 */ + 1230 "00110000" // /* MW 2 */ + 1231 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 78 31 +.src_ref 2 "0_0.cc" 79 25 +.src_ref 2 "0_0.cc" 81 19 + 1232 "10111010" // LDA r17, [p7, dj0]; MOVXM r19, #504448 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1233 "00010000" // /* MW 9 */ + 1234 "01000000" // /* MW 8 */ + 1235 "01101001" // /* MW 7 */ + 1236 "11101110" // /* MW 6 */ + 1237 "00000001" // /* MW 5 */ + 1238 "00000000" // /* MW 4 */ + 1239 "11010000" // /* MW 3 */ + 1240 "01000110" // /* MW 2 */ + 1241 "11100000" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 564 18 +.src_ref 4 "io_buffer_compiler.h" 565 24 +.src_ref 4 "io_buffer_compiler.h" 572 18 +.src_ref 2 "0_0.cc" 79 25 first + 1242 "01100100" // EQ r27, r19, r13; MOV r24, #0 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1243 "00000001" // /* MW 5 */ + 1244 "00100000" // /* MW 4 */ + 1245 "11111100" // /* MW 3 */ + 1246 "11011010" // /* MW 2 */ + 1247 "10011110" // /* MW 1 */ +.src_ref 2 "0_0.cc" 79 21 + 1248 "00011000" // SEL.EQZ r23, r20, r18, r27 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1249 "00100010" // /* MW 3 */ + 1250 "00101111" // /* MW 2 */ + 1251 "00010101" // /* MW 1 */ +.src_ref 2 "0_0.cc" 81 19 + 1252 "01000100" // MOVXM r21, #505472 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1253 "00000000" // /* MW 5 */ + 1254 "10101101" // /* MW 4 */ + 1255 "10111010" // /* MW 3 */ + 1256 "00000111" // /* MW 2 */ + 1257 "00000000" // /* MW 1 */ + 1258 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1259 "00000000" // /* MW 1 */ + 1260 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1261 "00000000" // /* MW 1 */ + 1262 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1263 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 78 31 first + 1264 "00011000" // EQZ r26, r17 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1265 "11010000" // /* MW 3 */ + 1266 "01110100" // /* MW 2 */ + 1267 "00010100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 79 12 first + 1268 "00011000" // REL r23, r22 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1269 "01101000" // /* MW 3 */ + 1270 "11010001" // /* MW 2 */ + 1271 "00010101" // /* MW 1 */ +.src_ref 2 "0_0.cc" 80 27 first + 1272 "00011000" // SEL.EQZ r18, r18, r20, r27 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1273 "01000010" // /* MW 3 */ + 1274 "10100101" // /* MW 2 */ + 1275 "00010100" // /* MW 1 */ + 1276 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1277 "00000000" // /* MW 1 */ + 1278 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1279 "00000000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 80 12 + 1280 "00011000" // ACQ.COND r18, r16, r26 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1281 "00001000" // /* MW 3 */ + 1282 "10010111" // /* MW 2 */ + 1283 "00010100" // /* MW 1 */ + 1284 "10000100" // JZ r17, #416 /* MW 6 */ /* control_operation: words=6 jump conditional cycles_taken=1 cycles_not_taken=0 direct absolute target_address=416 delay_slots=5 */ + 1285 "00000001" // /* MW 5 */ + 1286 "00000000" // /* MW 4 */ + 1287 "11010000" // /* MW 3 */ + 1288 "00000000" // /* MW 2 */ + 1289 "10001000" // /* MW 1 */ +.src_ref 2 "0_0.cc" 81 19 first +.delay_slot + 1290 "00011000" // SEL.EQZ r16, r19, r21, r27 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1291 "01010010" // /* MW 3 */ + 1292 "11100001" // /* MW 2 */ + 1293 "00010100" // /* MW 1 */ +.src_ref 2 "0_0.cc" 29 31 +.delay_slot + 1294 "11111000" // MOV p7, r16 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1295 "00100000" // /* MW 3 */ + 1296 "01101000" // /* MW 2 */ + 1297 "00011111" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1298 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1299 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1300 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1301 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1302 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1303 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 278 68 +.src_ref 3 "tile_control.h" 278 68 +.src_ref 3 "tile_control.h" 278 68 +.src_ref 4 "io_buffer_compiler.h" 566 27 +.src_ref 4 "io_buffer_compiler.h" 567 18 +.src_ref 2 "0_0.cc" 19 8 +.loop_nesting 1 + 1304 "10111010" // LDA p0, [sp, #-8]; MOVX r25, #0; MOV r16, #1 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1305 "01011000" // /* MW 9 */ + 1306 "00000001" // /* MW 8 */ + 1307 "00001000" // /* MW 7 */ + 1308 "00001010" // /* MW 6 */ + 1309 "10010000" // /* MW 5 */ + 1310 "00000001" // /* MW 4 */ + 1311 "00100000" // /* MW 3 */ + 1312 "00000011" // /* MW 2 */ + 1313 "11111111" // /* MW 1 */ + 1314 "10000100" // J #272 /* MW 6 */ /* control_operation: words=6 jump unconditional cycles_taken=1 direct absolute target_address=272 delay_slots=5 */ + 1315 "00000000" // /* MW 5 */ + 1316 "00000000" // /* MW 4 */ + 1317 "10001000" // /* MW 3 */ + 1318 "00000000" // /* MW 2 */ + 1319 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1320 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1321 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1322 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1323 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1324 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1325 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1326 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1327 "00000000" // /* MW 1 */ +.src_ref 4 "io_buffer_compiler.h" 564 18 +.delay_slot + 1328 "00011000" // LDA p1, [sp, #-12] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1329 "10011001" // /* MW 3 */ + 1330 "11110100" // /* MW 2 */ +.label _main__end +.label _main___func_end0 + 1331 "00000111" // /* MW 1 */ +.label __ZN3adf11block_writeEPKNS_7reg_valEj___func_begin0 +.label _ZN3adf11block_writeEPKNS_7reg_valEj +.function block_write _ZN3adf11block_writeEPKNS_7reg_valEj +.src_ref 3 "tile_control.h" 288 first +.src_ref 3 "tile_control.h" 292 8 +.src_ref 3 "tile_control.h" 292 25 +.function_start + 1344 "10000100" // JZ r0, #1504 /* MW 6 */ /* control_operation: words=6 jump conditional cycles_taken=1 cycles_not_taken=0 direct absolute target_address=1504 delay_slots=5 */ + 1345 "00000001" // /* MW 5 */ + 1346 "00000000" // /* MW 4 */ + 1347 "11110000" // /* MW 3 */ + 1348 "00000010" // /* MW 2 */ + 1349 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1350 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1351 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1352 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1353 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1354 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1355 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1356 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1357 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1358 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1359 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 278 34 +.src_ref 3 "tile_control.h" 292 8 first + 1360 "11100100" // MOVX r0, #-4; MOV lc, r0 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1361 "01000001" // /* MW 5 */ + 1362 "11100000" // /* MW 4 */ + 1363 "00101010" // /* MW 3 */ + 1364 "00011110" // /* MW 2 */ + 1365 "11111000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 292 8 + 1366 "01000100" // MOVXM ls, #1392 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1367 "11100000" // /* MW 5 */ + 1368 "11101010" // /* MW 4 */ + 1369 "00000001" // /* MW 3 */ + 1370 "00000000" // /* MW 2 */ + 1371 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 292 8 + 1372 "01000100" // MOVXM le, #1488 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1373 "10100000" // /* MW 5 */ + 1374 "11101011" // /* MW 4 */ + 1375 "00000110" // /* MW 3 */ + 1376 "00000000" // /* MW 2 */ + 1377 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 278 34 +.src_ref 3 "tile_control.h" 278 68 + 1378 "01111110" // NOPA; NOPB; NOPS; MOVXM p1, #524288 /* MW 14 */ /* control_operation: words=14 cycles_taken=1 */ + 1379 "01100000" // /* MW 13 */ + 1380 "00101011" // /* MW 12 */ + 1381 "00000000" // /* MW 11 */ + 1382 "00000010" // /* MW 10 */ + 1383 "00000000" // /* MW 9 */ + 1384 "00010110" // /* MW 8 */ + 1385 "01000000" // /* MW 7 */ + 1386 "00000000" // /* MW 6 */ + 1387 "00100000" // /* MW 5 */ + 1388 "00000000" // /* MW 4 */ + 1389 "11110000" // /* MW 3 */ + 1390 "00101100" // /* MW 2 */ + 1391 "00000000" // /* MW 1 */ +.label ZLS_F_ZN3adf11block_writeEPKNS_7reg_valEj_48 +.src_ref 3 "tile_control.h" 292 44 +.begin_of_loop +.loop_nesting 1 + 1392 "10011000" // LDA r2, [p0], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1393 "01010110" // /* MW 3 */ + 1394 "00011100" // /* MW 2 */ + 1395 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 292 44 + 1396 "10011000" // LDA r1, [p0], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1397 "00110110" // /* MW 3 */ + 1398 "00011100" // /* MW 2 */ + 1399 "00000000" // /* MW 1 */ + 1400 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1401 "00000000" // /* MW 1 */ + 1402 "00111100" // NOPA; NOPB /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1403 "00100000" // /* MW 5 */ + 1404 "00000000" // /* MW 4 */ + 1405 "11110000" // /* MW 3 */ + 1406 "00101100" // /* MW 2 */ + 1407 "00000000" // /* MW 1 */ + 1408 "11100001" // NOPA; NOPB; NOPS; NOPX; NOPM; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 1409 "00000000" // /* MW 15 */ + 1410 "00000000" // /* MW 14 */ + 1411 "01111000" // /* MW 13 */ + 1412 "10100101" // /* MW 12 */ + 1413 "00000001" // /* MW 11 */ + 1414 "00000000" // /* MW 10 */ + 1415 "00000000" // /* MW 9 */ + 1416 "00000000" // /* MW 8 */ + 1417 "01011011" // /* MW 7 */ + 1418 "00000001" // /* MW 6 */ + 1419 "00100000" // /* MW 5 */ + 1420 "00000000" // /* MW 4 */ + 1421 "11110000" // /* MW 3 */ + 1422 "00101100" // /* MW 2 */ + 1423 "00000000" // /* MW 1 */ + 1424 "11100001" // NOPA; NOPB; NOPS; NOPX; NOPM; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 1425 "00000000" // /* MW 15 */ + 1426 "00000000" // /* MW 14 */ + 1427 "01111000" // /* MW 13 */ + 1428 "10100101" // /* MW 12 */ + 1429 "00000001" // /* MW 11 */ + 1430 "00000000" // /* MW 10 */ + 1431 "00000000" // /* MW 9 */ + 1432 "00000000" // /* MW 8 */ + 1433 "01011011" // /* MW 7 */ + 1434 "00000001" // /* MW 6 */ + 1435 "00100000" // /* MW 5 */ + 1436 "00000000" // /* MW 4 */ + 1437 "11110000" // /* MW 3 */ + 1438 "00101100" // /* MW 2 */ + 1439 "00000000" // /* MW 1 */ + 1440 "11100001" // NOPA; NOPB; NOPS; NOPX; NOPM; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 1441 "00000000" // /* MW 15 */ + 1442 "00000000" // /* MW 14 */ + 1443 "01111000" // /* MW 13 */ + 1444 "10100101" // /* MW 12 */ + 1445 "00000001" // /* MW 11 */ + 1446 "00000000" // /* MW 10 */ + 1447 "00000000" // /* MW 9 */ + 1448 "00000000" // /* MW 8 */ + 1449 "01011011" // /* MW 7 */ + 1450 "00000001" // /* MW 6 */ + 1451 "00100000" // /* MW 5 */ + 1452 "00000000" // /* MW 4 */ + 1453 "11110000" // /* MW 3 */ + 1454 "00101100" // /* MW 2 */ + 1455 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 278 34 first + 1456 "11100001" // NOPA; NOPB; NOPS; AND r3, r2, r0; NOPM; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 1457 "00000000" // /* MW 15 */ + 1458 "00000000" // /* MW 14 */ + 1459 "01111000" // /* MW 13 */ + 1460 "10100101" // /* MW 12 */ + 1461 "00000001" // /* MW 11 */ + 1462 "00100100" // /* MW 10 */ + 1463 "00110000" // /* MW 9 */ + 1464 "00000100" // /* MW 8 */ + 1465 "01011011" // /* MW 7 */ + 1466 "00000001" // /* MW 6 */ + 1467 "00100000" // /* MW 5 */ + 1468 "00000000" // /* MW 4 */ + 1469 "11110000" // /* MW 3 */ + 1470 "00101100" // /* MW 2 */ + 1471 "00000000" // /* MW 1 */ +.src_ref 3 "tile_control.h" 278 34 +.src_ref 3 "tile_control.h" 278 68 + 1472 "11100001" // NOPA; NOPB; NOPS; NOPX; MOV dj0, r3; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 1473 "00000000" // /* MW 15 */ + 1474 "00000000" // /* MW 14 */ + 1475 "01111000" // /* MW 13 */ + 1476 "11010000" // /* MW 12 */ + 1477 "01000000" // /* MW 11 */ + 1478 "00000000" // /* MW 10 */ + 1479 "00000000" // /* MW 9 */ + 1480 "00000000" // /* MW 8 */ + 1481 "01011011" // /* MW 7 */ + 1482 "00000001" // /* MW 6 */ + 1483 "00100000" // /* MW 5 */ + 1484 "00000000" // /* MW 4 */ + 1485 "11110000" // /* MW 3 */ + 1486 "00101100" // /* MW 2 */ + 1487 "00000000" // /* MW 1 */ +.label ZLE_F_ZN3adf11block_writeEPKNS_7reg_valEj_144 +.src_ref 3 "tile_control.h" 278 34 +.src_ref 3 "tile_control.h" 278 68 +.end_of_loop + 1488 "11100001" // NOPA; NOPB; ST.TM r1, [p1, dj0]; NOPX; NOPM; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 1489 "00000000" // /* MW 15 */ + 1490 "00000000" // /* MW 14 */ + 1491 "01111000" // /* MW 13 */ + 1492 "10100101" // /* MW 12 */ + 1493 "00000001" // /* MW 11 */ + 1494 "00000000" // /* MW 10 */ + 1495 "00000000" // /* MW 9 */ + 1496 "10000000" // /* MW 8 */ + 1497 "00111110" // /* MW 7 */ + 1498 "00000000" // /* MW 6 */ + 1499 "00100001" // /* MW 5 */ + 1500 "00000000" // /* MW 4 */ + 1501 "11110000" // /* MW 3 */ + 1502 "00101100" // /* MW 2 */ + 1503 "00000000" // /* MW 1 */ +.label TGT_F_ZN3adf11block_writeEPKNS_7reg_valEj_160 +.src_ref 3 "tile_control.h" 293 4 first +.loop_nesting 0 + 1504 "00011000" // RET lr /* MW 4 */ /* control_operation: words=4 rts unconditional cycles_taken=1 delay_slots=5 */ + 1505 "00000000" // /* MW 3 */ + 1506 "00101000" // /* MW 2 */ + 1507 "00010000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1508 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1509 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1510 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1511 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1512 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1513 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1514 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1515 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1516 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ +.label _ZN3adf11block_writeEPKNS_7reg_valEj__end +.label __ZN3adf11block_writeEPKNS_7reg_valEj___func_end0 + 1517 "00000000" // /* MW 1 */ +.label _fini +.function _fini _fini +.src_ref 0 "me_basic.c" 73 4 first +.src_ref 0 "me_basic.c" 73 9 +.function_start + 1520 "11000100" // PADDXM [sp], #64 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1521 "00000001" // /* MW 5 */ + 1522 "00000000" // /* MW 4 */ + 1523 "00000000" // /* MW 3 */ + 1524 "00001000" // /* MW 2 */ + 1525 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 75 8 +.src_ref 0 "me_basic.c" 76 13 + 1526 "00111010" // ST r14, [sp, #-12]; MOVXM r16, #0 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1527 "00010001" // /* MW 9 */ + 1528 "00000000" // /* MW 8 */ + 1529 "00001000" // /* MW 7 */ + 1530 "00000010" // /* MW 6 */ + 1531 "00000000" // /* MW 5 */ + 1532 "00000000" // /* MW 4 */ + 1533 "10110000" // /* MW 3 */ + 1534 "10111010" // /* MW 2 */ + 1535 "11111110" // /* MW 1 */ +.src_ref 0 "me_basic.c" 75 8 +.src_ref 0 "me_basic.c" 75 41 + 1536 "00111010" // ST p7, [sp, #-8]; MOVXM r14, #0 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1537 "00010001" // /* MW 9 */ + 1538 "00000000" // /* MW 8 */ + 1539 "11001000" // /* MW 7 */ + 1540 "00000001" // /* MW 6 */ + 1541 "00000000" // /* MW 5 */ + 1542 "00000000" // /* MW 4 */ + 1543 "10110000" // /* MW 3 */ + 1544 "01110011" // /* MW 2 */ + 1545 "11111111" // /* MW 1 */ +.src_ref 0 "me_basic.c" 75 8 first +.src_ref 0 "me_basic.c" 76 13 + 1546 "11100100" // EQ r16, r14, r16; MOV p7, r16 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1547 "01000001" // /* MW 5 */ + 1548 "11010000" // /* MW 4 */ + 1549 "11111110" // /* MW 3 */ + 1550 "00100000" // /* MW 2 */ + 1551 "01110100" // /* MW 1 */ +.src_ref 0 "me_basic.c" 75 8 + 1552 "10000100" // JNZ r16, #1648 /* MW 6 */ /* control_operation: words=6 jump conditional cycles_taken=1 cycles_not_taken=0 direct absolute target_address=1648 delay_slots=5 */ + 1553 "00000001" // /* MW 5 */ + 1554 "01000000" // /* MW 4 */ + 1555 "00111000" // /* MW 3 */ + 1556 "00000011" // /* MW 2 */ + 1557 "10000000" // /* MW 1 */ +.delay_slot + 1558 "10011000" // ST r15, [sp, #-4] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1559 "11110101" // /* MW 3 */ + 1560 "11111101" // /* MW 2 */ + 1561 "00001111" // /* MW 1 */ +.delay_slot + 1562 "10011000" // ST lr, [sp, #-16] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1563 "00111101" // /* MW 3 */ + 1564 "11110000" // /* MW 2 */ + 1565 "00001111" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1566 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1567 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1568 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1569 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1570 "00101110" // NOPA; NOPS; NOPM; NOPV /* MW 14 */ /* control_operation: words=14 cycles_taken=1 */ + 1571 "00011100" // /* MW 13 */ + 1572 "00000000" // /* MW 12 */ + 1573 "00000000" // /* MW 11 */ + 1574 "01010111" // /* MW 10 */ + 1575 "00011010" // /* MW 9 */ + 1576 "01000000" // /* MW 8 */ + 1577 "00000000" // /* MW 7 */ + 1578 "00000000" // /* MW 6 */ + 1579 "10110110" // /* MW 5 */ + 1580 "00000010" // /* MW 4 */ + 1581 "11110000" // /* MW 3 */ + 1582 "00101100" // /* MW 2 */ + 1583 "00000000" // /* MW 1 */ +.label TGT_F_fini_64 +.src_ref 0 "me_basic.c" 76 13 first +.loop_nesting 1 + 1584 "10011000" // LDA p0, [p7], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1585 "00011110" // /* MW 3 */ + 1586 "00011100" // /* MW 2 */ + 1587 "00000111" // /* MW 1 */ + 1588 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1589 "00000000" // /* MW 1 */ + 1590 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1591 "00000000" // /* MW 1 */ + 1592 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1593 "00000000" // /* MW 1 */ + 1594 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1595 "00000000" // /* MW 1 */ + 1596 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1597 "00000000" // /* MW 1 */ + 1598 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1599 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 76 16 +.no_stack_arguments + 1600 "00011000" // JL p0 /* MW 4 */ /* control_operation: words=4 call unconditional cycles_taken=1 indirect absolute delay_slots=5 */ + 1601 "00000000" // /* MW 3 */ + 1602 "00110000" // /* MW 2 */ + 1603 "00010000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 75 41 +.delay_slot + 1604 "11111000" // MOV r15, p7 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1605 "11000000" // /* MW 3 */ + 1606 "11011110" // /* MW 2 */ + 1607 "00011011" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1608 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1609 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1610 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1611 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1612 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1613 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1614 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1615 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 75 41 first +.return_address + 1616 "10011000" // NE r16, r15, r14 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1617 "11101000" // /* MW 3 */ + 1618 "11100000" // /* MW 2 */ + 1619 "00010011" // /* MW 1 */ +.src_ref 0 "me_basic.c" 75 8 + 1620 "10000100" // JNZ r16, #1584 /* MW 6 */ /* control_operation: words=6 jump conditional cycles_taken=1 cycles_not_taken=0 direct absolute target_address=1584 delay_slots=5 */ + 1621 "00000001" // /* MW 5 */ + 1622 "01000000" // /* MW 4 */ + 1623 "00011000" // /* MW 3 */ + 1624 "00000011" // /* MW 2 */ + 1625 "10000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1626 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1627 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1628 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1629 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1630 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1631 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1632 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1633 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1634 "00101110" // NOPA; NOPS; NOPM; NOPV /* MW 14 */ /* control_operation: words=14 cycles_taken=1 */ + 1635 "00011100" // /* MW 13 */ + 1636 "00000000" // /* MW 12 */ + 1637 "00000000" // /* MW 11 */ + 1638 "01010111" // /* MW 10 */ + 1639 "00011010" // /* MW 9 */ + 1640 "01000000" // /* MW 8 */ + 1641 "00000000" // /* MW 7 */ + 1642 "00000000" // /* MW 6 */ + 1643 "10110110" // /* MW 5 */ + 1644 "00000010" // /* MW 4 */ + 1645 "11110000" // /* MW 3 */ + 1646 "00101100" // /* MW 2 */ + 1647 "00000000" // /* MW 1 */ +.label TGT_F_fini_128 +.src_ref 0 "me_basic.c" 77 4 +.loop_nesting 0 + 1648 "00011000" // LDA lr, [sp, #-16] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1649 "00111001" // /* MW 3 */ + 1650 "11110000" // /* MW 2 */ + 1651 "00000111" // /* MW 1 */ + 1652 "00011000" // LDA r14, [sp, #-12] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1653 "11010001" // /* MW 3 */ + 1654 "11110101" // /* MW 2 */ + 1655 "00000111" // /* MW 1 */ + 1656 "00011000" // LDA r15, [sp, #-4] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1657 "11110001" // /* MW 3 */ + 1658 "11111101" // /* MW 2 */ + 1659 "00000111" // /* MW 1 */ + 1660 "00011000" // LDA p7, [sp, #-8] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1661 "10011001" // /* MW 3 */ + 1662 "11111011" // /* MW 2 */ + 1663 "00000111" // /* MW 1 */ + 1664 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1665 "00000000" // /* MW 1 */ + 1666 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1667 "00000000" // /* MW 1 */ + 1668 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1669 "00000000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 77 4 first + 1670 "00011000" // RET lr /* MW 4 */ /* control_operation: words=4 rts unconditional cycles_taken=1 delay_slots=5 */ + 1671 "00000000" // /* MW 3 */ + 1672 "00101000" // /* MW 2 */ + 1673 "00010000" // /* MW 1 */ +.src_ref 0 "me_basic.c" 77 4 +.delay_slot + 1674 "11000100" // PADDXM [sp], #-64 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1675 "00000001" // /* MW 5 */ + 1676 "00000000" // /* MW 4 */ + 1677 "00000000" // /* MW 3 */ + 1678 "11111000" // /* MW 2 */ + 1679 "11111111" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1680 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1681 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1682 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1683 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1684 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1685 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1686 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ +.label _fini__end + 1687 "00000000" // /* MW 1 */ +.label __cxa_finalize +.function __cxa_finalize __cxa_finalize +.src_ref 5 "atexit.c" 47 first +.src_ref 5 "atexit.c" 47 5 +.src_ref 5 "atexit.c" 55 15 +.function_start + 1696 "10111010" // MOVA r1, #-3; PADDXM [sp], #64 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1697 "01110000" // /* MW 9 */ + 1698 "00000000" // /* MW 8 */ + 1699 "00000000" // /* MW 7 */ + 1700 "00000000" // /* MW 6 */ + 1701 "00000010" // /* MW 5 */ + 1702 "00000000" // /* MW 4 */ + 1703 "00000000" // /* MW 3 */ + 1704 "10100001" // /* MW 2 */ + 1705 "11111111" // /* MW 1 */ +.src_ref 5 "atexit.c" 53 4 + 1706 "00111010" // ST lr, [sp, #-40]; MOVX r4, #8; MOV r3, packSign1 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1707 "01111001" // /* MW 9 */ + 1708 "11100000" // /* MW 8 */ + 1709 "01101101" // /* MW 7 */ + 1710 "00001000" // /* MW 6 */ + 1711 "01000001" // /* MW 5 */ + 1712 "00000000" // /* MW 4 */ + 1713 "10110000" // /* MW 3 */ + 1714 "00000111" // /* MW 2 */ + 1715 "11111011" // /* MW 1 */ + 1716 "00000010" // ST r3, [sp, #-44]; MOV r3, packSign0 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1717 "01110000" // /* MW 7 */ + 1718 "11100000" // /* MW 6 */ + 1719 "01101001" // /* MW 5 */ + 1720 "00000000" // /* MW 4 */ + 1721 "10110000" // /* MW 3 */ + 1722 "10001110" // /* MW 2 */ + 1723 "11111010" // /* MW 1 */ + 1724 "00000010" // ST r3, [sp, #-32]; MOV r3, unpackSign1 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1725 "01110000" // /* MW 7 */ + 1726 "00110000" // /* MW 6 */ + 1727 "01101110" // /* MW 5 */ + 1728 "00000000" // /* MW 4 */ + 1729 "10110000" // /* MW 3 */ + 1730 "00001110" // /* MW 2 */ + 1731 "11111100" // /* MW 1 */ + 1732 "00000010" // ST r3, [sp, #-28]; MOV r3, unpackSign0 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1733 "01110000" // /* MW 7 */ + 1734 "00110000" // /* MW 6 */ + 1735 "01101010" // /* MW 5 */ + 1736 "00000000" // /* MW 4 */ + 1737 "10110000" // /* MW 3 */ + 1738 "10001110" // /* MW 2 */ + 1739 "11111100" // /* MW 1 */ +.src_ref 5 "atexit.c" 52 14 +.src_ref 5 "atexit.c" 53 4 + 1740 "00111010" // ST r3, [sp, #-20]; MOVXM p0, #508960 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1741 "00010001" // /* MW 9 */ + 1742 "00010000" // /* MW 8 */ + 1743 "00110010" // /* MW 7 */ + 1744 "11110000" // /* MW 6 */ + 1745 "00000001" // /* MW 5 */ + 1746 "00000000" // /* MW 4 */ + 1747 "10110000" // /* MW 3 */ + 1748 "10001110" // /* MW 2 */ + 1749 "11111101" // /* MW 1 */ +.src_ref 5 "atexit.c" 52 14 first + 1750 "11010100" // LDA r2, [p0]; MOV r3, crSRSMode /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1751 "11000001" // /* MW 5 */ + 1752 "10110001" // /* MW 4 */ + 1753 "11010001" // /* MW 3 */ + 1754 "10001010" // /* MW 2 */ + 1755 "00000000" // /* MW 1 */ + 1756 "00000010" // ST r3, [sp, #-16]; MOV r3, crPackSize /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1757 "01110000" // /* MW 7 */ + 1758 "10110000" // /* MW 6 */ + 1759 "01101011" // /* MW 5 */ + 1760 "00000000" // /* MW 4 */ + 1761 "10110000" // /* MW 3 */ + 1762 "00001110" // /* MW 2 */ + 1763 "11111110" // /* MW 1 */ + 1764 "00000010" // ST r3, [sp, #-12]; MOV r3, crSat /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1765 "01110000" // /* MW 7 */ + 1766 "01110000" // /* MW 6 */ + 1767 "01101010" // /* MW 5 */ + 1768 "00000000" // /* MW 4 */ + 1769 "10110000" // /* MW 3 */ + 1770 "10001110" // /* MW 2 */ + 1771 "11111110" // /* MW 1 */ +.src_ref 5 "atexit.c" 53 4 first + 1772 "00000010" // ST r4, [p0]; MOV r9, upsSign1 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1773 "01110000" // /* MW 7 */ + 1774 "00110000" // /* MW 6 */ + 1775 "00101100" // /* MW 5 */ + 1776 "00000001" // /* MW 4 */ + 1777 "00110000" // /* MW 3 */ + 1778 "10010010" // /* MW 2 */ + 1779 "00000000" // /* MW 1 */ + 1780 "00000010" // ST r3, [sp, #-8]; MOV r8, upsSign0 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1781 "01110000" // /* MW 7 */ + 1782 "00110000" // /* MW 6 */ + 1783 "00001000" // /* MW 5 */ + 1784 "00000001" // /* MW 4 */ + 1785 "10110000" // /* MW 3 */ + 1786 "00001110" // /* MW 2 */ + 1787 "11111111" // /* MW 1 */ + 1788 "11111000" // MOV r11, vaddSign1 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1789 "01100000" // /* MW 3 */ + 1790 "11011010" // /* MW 2 */ + 1791 "00011010" // /* MW 1 */ + 1792 "11111000" // MOV r10, vaddSign0 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1793 "01100000" // /* MW 3 */ + 1794 "10010010" // /* MW 2 */ + 1795 "00011010" // /* MW 1 */ +.src_ref 5 "atexit.c" 54 8 first + 1796 "11100100" // ADD r2, r2, #-8; MOV r12, srsSign1 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1797 "10000001" // /* MW 5 */ + 1798 "00111111" // /* MW 4 */ + 1799 "01100110" // /* MW 3 */ + 1800 "10111100" // /* MW 2 */ + 1801 "00010000" // /* MW 1 */ +.src_ref 5 "atexit.c" 55 15 first + 1802 "11100100" // ASHL r13, r2, r1; MOV r1, crUnpackSize /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1803 "11000001" // /* MW 5 */ + 1804 "10100101" // /* MW 4 */ + 1805 "11010000" // /* MW 3 */ + 1806 "01000011" // /* MW 2 */ + 1807 "00010011" // /* MW 1 */ + 1808 "00000010" // ST r1, [sp, #-4]; MOV r1, crRnd /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1809 "01110000" // /* MW 7 */ + 1810 "10110000" // /* MW 6 */ + 1811 "00101111" // /* MW 5 */ + 1812 "00000000" // /* MW 4 */ + 1813 "10110000" // /* MW 3 */ + 1814 "10000110" // /* MW 2 */ + 1815 "11111111" // /* MW 1 */ + 1816 "00000010" // ST r1, [sp, #-24]; MOV r1, crUPSMode /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1817 "01110000" // /* MW 7 */ + 1818 "01110000" // /* MW 6 */ + 1819 "00101110" // /* MW 5 */ + 1820 "00000000" // /* MW 4 */ + 1821 "10110000" // /* MW 3 */ + 1822 "00000110" // /* MW 2 */ + 1823 "11111101" // /* MW 1 */ + 1824 "00000010" // ST r1, [sp, #-36]; MOV r14, srsSign0 /* MW 8 */ /* control_operation: words=8 cycles_taken=1 */ + 1825 "01110000" // /* MW 7 */ + 1826 "11100000" // /* MW 6 */ + 1827 "11001011" // /* MW 5 */ + 1828 "00000001" // /* MW 4 */ + 1829 "10110000" // /* MW 3 */ + 1830 "10000110" // /* MW 2 */ + 1831 "11111011" // /* MW 1 */ +.src_ref 5 "atexit.c" 56 37 + 1832 "01000100" // MOVXM r1, #508928 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1833 "00000000" // /* MW 5 */ + 1834 "10101000" // /* MW 4 */ + 1835 "11000000" // /* MW 3 */ + 1836 "00000111" // /* MW 2 */ + 1837 "00000000" // /* MW 1 */ +.src_ref 5 "atexit.c" 56 37 first + 1838 "01011000" // ADD.NC p6, r1, r2 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1839 "10001001" // /* MW 3 */ + 1840 "01100000" // /* MW 2 */ + 1841 "00011110" // /* MW 1 */ +.src_ref 5 "atexit.c" 60 4 + 1842 "01111110" // NOPA; NOPB; NOPS; MOVXM p7, #1856 /* MW 14 */ /* control_operation: words=14 cycles_taken=1 */ + 1843 "01100000" // /* MW 13 */ + 1844 "00101011" // /* MW 12 */ + 1845 "00000000" // /* MW 11 */ + 1846 "00000010" // /* MW 10 */ + 1847 "01110100" // /* MW 9 */ + 1848 "01110110" // /* MW 8 */ + 1849 "00000000" // /* MW 7 */ + 1850 "00000000" // /* MW 6 */ + 1851 "00100000" // /* MW 5 */ + 1852 "00000000" // /* MW 4 */ + 1853 "11110000" // /* MW 3 */ + 1854 "00101100" // /* MW 2 */ + 1855 "00000000" // /* MW 1 */ +.label TGT_F__cxa_finalize_160 +.src_ref 5 "atexit.c" 63 18 first +.loop_nesting 1 + 1856 "10011000" // LDA p1, [p6], #4 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1857 "10011110" // /* MW 3 */ + 1858 "00011100" // /* MW 2 */ + 1859 "00000110" // /* MW 1 */ +.src_ref 5 "atexit.c" 63 35 + 1860 "10011000" // LDA p0, [p6], #-12 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1861 "00011110" // /* MW 3 */ + 1862 "11011100" // /* MW 2 */ + 1863 "00000110" // /* MW 1 */ + 1864 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1865 "00000000" // /* MW 1 */ + 1866 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1867 "00000000" // /* MW 1 */ + 1868 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1869 "00000000" // /* MW 1 */ + 1870 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1871 "00000000" // /* MW 1 */ + 1872 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1873 "00000000" // /* MW 1 */ +.src_ref 5 "atexit.c" 63 24 +.no_stack_arguments + 1874 "00011000" // JL p1 /* MW 4 */ /* control_operation: words=4 call unconditional cycles_taken=1 indirect absolute delay_slots=5 */ + 1875 "01000000" // /* MW 3 */ + 1876 "00110000" // /* MW 2 */ + 1877 "00010000" // /* MW 1 */ +.delay_slot + 1878 "11111000" // MOV r15, r0 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1879 "00100000" // /* MW 3 */ + 1880 "11010000" // /* MW 2 */ + 1881 "00011011" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1882 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1883 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1884 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1885 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1886 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1887 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1888 "11100001" // NOPA; NOPB; NOPS; NOPX; NOPM; NOPV /* MW 16 */ /* control_operation: words=16 cycles_taken=1 */ + 1889 "00000000" // /* MW 15 */ + 1890 "00000000" // /* MW 14 */ + 1891 "01111000" // /* MW 13 */ + 1892 "10100101" // /* MW 12 */ + 1893 "00000001" // /* MW 11 */ + 1894 "00000000" // /* MW 10 */ + 1895 "00000000" // /* MW 9 */ + 1896 "00000000" // /* MW 8 */ + 1897 "01011011" // /* MW 7 */ + 1898 "00000001" // /* MW 6 */ + 1899 "00100000" // /* MW 5 */ + 1900 "00000000" // /* MW 4 */ + 1901 "11110000" // /* MW 3 */ + 1902 "00101100" // /* MW 2 */ + 1903 "00000000" // /* MW 1 */ +.src_ref 5 "atexit.c" 60 4 first +.return_address + 1904 "00011000" // JNZD r13, r13, p7 /* MW 4 */ /* control_operation: words=4 jump conditional cycles_taken=1 cycles_not_taken=0 indirect absolute delay_slots=5 */ + 1905 "11100000" // /* MW 3 */ + 1906 "01011011" // /* MW 2 */ + 1907 "00010011" // /* MW 1 */ +.delay_slot + 1908 "11111000" // MOV r0, r15 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1909 "10100000" // /* MW 3 */ + 1910 "00010111" // /* MW 2 */ + 1911 "00011000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1912 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1913 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1914 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1915 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1916 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1917 "00000000" // /* MW 1 */ +.delay_slot +.swstall delay_slot + 1918 "00000000" // NOPX /* MW 2 */ /* control_operation: words=2 cycles_taken=1 */ + 1919 "00000000" // /* MW 1 */ +.src_ref 5 "atexit.c" 66 +.loop_nesting 0 + 1920 "10111010" // LDA lr, [sp, #-40]; MOVX upsSign1, r9; MOV vaddSign1, r11 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1921 "01111000" // /* MW 9 */ + 1922 "11010000" // /* MW 8 */ + 1923 "10011010" // /* MW 7 */ + 1924 "00000010" // /* MW 6 */ + 1925 "11101010" // /* MW 5 */ + 1926 "00010011" // /* MW 4 */ + 1927 "00100000" // /* MW 3 */ + 1928 "00000111" // /* MW 2 */ + 1929 "11111011" // /* MW 1 */ + 1930 "10111010" // LDA r1, [sp, #-44]; MOVX upsSign0, r8; MOV vaddSign0, r10 /* MW 10 */ /* control_operation: words=10 cycles_taken=1 */ + 1931 "01111000" // /* MW 9 */ + 1932 "10010000" // /* MW 8 */ + 1933 "10011010" // /* MW 7 */ + 1934 "00000000" // /* MW 6 */ + 1935 "11001010" // /* MW 5 */ + 1936 "00010001" // /* MW 4 */ + 1937 "00100000" // /* MW 3 */ + 1938 "10000110" // /* MW 2 */ + 1939 "11111010" // /* MW 1 */ + 1940 "00101100" // LDA r2, [sp, #-32]; MOVX srsSign1, r12 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1941 "10000000" // /* MW 5 */ + 1942 "01111001" // /* MW 4 */ + 1943 "00100110" // /* MW 3 */ + 1944 "00001010" // /* MW 2 */ + 1945 "11111100" // /* MW 1 */ + 1946 "00101100" // LDA r3, [sp, #-28]; MOVX srsSign0, r14 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1947 "10000000" // /* MW 5 */ + 1948 "01110001" // /* MW 4 */ + 1949 "00100111" // /* MW 3 */ + 1950 "10001110" // /* MW 2 */ + 1951 "11111100" // /* MW 1 */ + 1952 "00011000" // LDA r7, [sp, #-20] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1953 "11110001" // /* MW 3 */ + 1954 "11101100" // /* MW 2 */ + 1955 "00000111" // /* MW 1 */ + 1956 "00011000" // LDA r4, [sp, #-16] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1957 "10010001" // /* MW 3 */ + 1958 "11110000" // /* MW 2 */ + 1959 "00000111" // /* MW 1 */ + 1960 "00011000" // LDA r5, [sp, #-12] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1961 "10110001" // /* MW 3 */ + 1962 "11110100" // /* MW 2 */ + 1963 "00000111" // /* MW 1 */ + 1964 "00011000" // LDA r6, [sp, #-8] /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 1965 "11010001" // /* MW 3 */ + 1966 "11111000" // /* MW 2 */ + 1967 "00000111" // /* MW 1 */ + 1968 "00101100" // LDA r1, [sp, #-4]; MOVX packSign1, r1 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1969 "10000000" // /* MW 5 */ + 1970 "11111000" // /* MW 4 */ + 1971 "00100000" // /* MW 3 */ + 1972 "10000110" // /* MW 2 */ + 1973 "11111111" // /* MW 1 */ + 1974 "00101100" // LDA r2, [sp, #-24]; MOVX packSign0, r2 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1975 "10000000" // /* MW 5 */ + 1976 "01110000" // /* MW 4 */ + 1977 "00100001" // /* MW 3 */ + 1978 "00001010" // /* MW 2 */ + 1979 "11111101" // /* MW 1 */ + 1980 "00101100" // LDA r3, [sp, #-36]; MOVX unpackSign1, r3 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1981 "10000000" // /* MW 5 */ + 1982 "11111011" // /* MW 4 */ + 1983 "00100001" // /* MW 3 */ + 1984 "10001110" // /* MW 2 */ + 1985 "11111011" // /* MW 1 */ +.src_ref 5 "atexit.c" 66 first + 1986 "11000100" // PADDXM [sp], #-64 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1987 "00000001" // /* MW 5 */ + 1988 "00000000" // /* MW 4 */ + 1989 "00000000" // /* MW 3 */ + 1990 "11111000" // /* MW 2 */ + 1991 "11111111" // /* MW 1 */ +.src_ref 5 "atexit.c" 66 + 1992 "11100100" // RET lr; MOV unpackSign0, r7 /* MW 6 */ /* control_operation: words=6 rts unconditional cycles_taken=1 delay_slots=5 */ + 1993 "01000001" // /* MW 5 */ + 1994 "01100111" // /* MW 4 */ + 1995 "00000100" // /* MW 3 */ + 1996 "00000000" // /* MW 2 */ + 1997 "00000101" // /* MW 1 */ +.delay_slot + 1998 "11100100" // MOVX crSRSMode, r4; MOV crPackSize, r5 /* MW 6 */ /* control_operation: words=6 cycles_taken=1 */ + 1999 "01000001" // /* MW 5 */ + 2000 "01100101" // /* MW 4 */ + 2001 "00000111" // /* MW 3 */ + 2002 "01100000" // /* MW 2 */ + 2003 "00100111" // /* MW 1 */ +.delay_slot + 2004 "00011000" // MOVX crSat, r6 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 2005 "10000000" // /* MW 3 */ + 2006 "10111011" // /* MW 2 */ + 2007 "00010001" // /* MW 1 */ +.delay_slot + 2008 "00011000" // MOVX crUnpackSize, r1 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 2009 "10000000" // /* MW 3 */ + 2010 "01111100" // /* MW 2 */ + 2011 "00010000" // /* MW 1 */ +.delay_slot + 2012 "00011000" // MOVX crRnd, r2 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 2013 "10000000" // /* MW 3 */ + 2014 "10111010" // /* MW 2 */ + 2015 "00010000" // /* MW 1 */ +.delay_slot + 2016 "00011000" // MOVX crUPSMode, r3 /* MW 4 */ /* control_operation: words=4 cycles_taken=1 */ + 2017 "00000000" // /* MW 3 */ + 2018 "11111100" // /* MW 2 */ +.label __cxa_finalize__end + 2019 "00010000" // /* MW 1 */ +.dir 0 "/scratch/sw_component_pipelines/continuous/gradle_simmodels_workspaces/21708/HEAD/build/Aie2p_core_model/stage-src/core_model/sipp/lib" +.dir 1 "/scratch/sw_component_pipelines/continuous/gradle_simmodels_workspaces/21708/HEAD/build/Aie2p_core_model/stage-src/core_model/sipp/lib/runtime/include" +.dir 2 "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/src" +.dir 3 "/usr/local/lib/python3.10/dist-packages/include/adf/aie" +.dir 4 "/usr/local/lib/python3.10/dist-packages/include/adf/io_buffer" +.dir 5 "/scratch/sw_component_pipelines/continuous/gradle_simmodels_workspaces/21708/HEAD/build/Aie2p_core_model/stage-src/core_model/sipp/lib/runtime/src" diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.cmico b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.cmico new file mode 100644 index 0000000000000000000000000000000000000000..f377058758269f564988080a1597f499edc1b997 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.cmico @@ -0,0 +1 @@ ++Mdec diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.lst b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.lst new file mode 100644 index 0000000000000000000000000000000000000000..6bc437b066d486a8c158397663e147dbc4c275a0 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.lst @@ -0,0 +1,824 @@ + +// File generated by darts version V-2024.06#84922c0d9f#241219, Fri Mar 21 03:42:52 2025 +// Copyright 2014-2024 Synopsys, Inc. All rights reserved. +// darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0 me + +// Release: ipp V-2024.06-TGT-241219 + +.text_segment PM 0 +.entry_point +.label __AIE_ARCH_MODEL_VERSION__21011100__inlined__1__me_basic___main_init_ +.label _main_init +.function_start + 0 0x00 0x07 0xb9 0xf5 0x80 0x44 MOVXM sp, #506560 + 6 0x00 0x00 0x04 0x20 0x00 0x44 MOVXM r8, #0 + 12 0x00 0x00 0x08 0x20 0x00 0x44 MOVXM r16, #0 + 18 0x12 0x21 0x07 0x98 EQ r16, r8, r16 + 22 0x80 0x00 0x40 0x40 0x01 0x84 JNZ r16, #128 +.delay_slot +.swstall delay_slot + 28 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 30 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 32 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 34 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 36 0x00 0x00 NOPX + 38 0x00 0x2c 0xf7 0x80 0x8b 0x01 0x28 0x50 0x72 0xba NOPA; MOVS p7, p0; MOV r9, r1 + 48 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x3f 0xff 0xff 0x37 0xfe 0x10 0x00 0x00 0xe1 NOPA; NOPB; NOPS; MOVXM p6, #-4; NOPV +.label TGT_F_main_init_64 +.loop_nesting 1 + 64 0xdf 0x83 0xd5 0x39 0x81 0xd4 LDA p0, [p6], #-4; MOV r10, p6 + 70 0x00 0x00 NOPX + 72 0x00 0x00 NOPX + 74 0x00 0x00 NOPX + 76 0x00 0x00 NOPX + 78 0x00 0x00 NOPX + 80 0x00 0x00 NOPX +.no_stack_arguments + 82 0x10 0x30 0x00 0x18 JL p0 +.delay_slot +.swstall delay_slot + 86 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 88 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 90 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 92 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 94 0x00 0x00 NOPX +.return_address + 96 0x12 0xa0 0x88 0x98 NE r16, r10, r8 + 100 0x80 0x00 0x20 0x40 0x01 0x84 JNZ r16, #64 +.delay_slot +.swstall delay_slot + 106 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 108 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 110 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 112 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 114 0x00 0x00 NOPX +.loop_nesting 0 + 116 0x00 0x2c 0xf0 0x00 0x20 0x9c 0x8b 0x00 0x2a 0x50 0x70 0xf6 NOPA; NOPB; MOVS p0, p7; MOV r1, r9 +.label TGT_F_main_init_128 +.no_stack_arguments + 128 0x00 0x00 0x70 0x00 0x01 0x04 JL #224 +.delay_slot +.swstall delay_slot + 134 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 136 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 138 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 140 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 142 0x00 0x00 NOPX +.return_address +.no_stack_arguments + 144 0x00 0x03 0x50 0x00 0x01 0x04 JL #1696 +.delay_slot + 150 0x18 0x60 0x00 0xb8 MOV p0, #0 +.delay_slot +.swstall delay_slot + 154 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 156 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 158 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 160 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; NOPM; NOPV +.return_address +.swstall chess_separator_scheduler + 176 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 178 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 180 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 182 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 184 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 186 0x00 0x00 NOPX + 188 0x10 0x08 0x00 0x18 DONE +.swstall chess_separator_scheduler + 192 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 194 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 196 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 198 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 200 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 202 0x00 0x00 NOPX +.swstall for_chess_exit +.exit + 204 0x00 0x01 0x67 0x98 NOPA +.label TGT_F_main_init_208 +.loop_nesting 1 + 208 0x00 0x00 0x68 0x00 0x00 0x84 J #208 +.delay_slot +.swstall delay_slot + 214 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 216 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 218 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 220 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 222 0x00 0x00 NOPX +.label _main_init__end +.label _main___func_begin0 +.label _main +.function_start + 224 0xec 0x00 0x80 0x00 0x06 0x00 0x00 0x00 0x70 0xba MOVA m0, #-160; PADDXM [sp], #192 + 234 0xe9 0x04 0x80 0x00 0xc0 0x48 0xb2 0xf0 0x78 0xba MOVA m1, #-184; MOVX r12, #2; MOV p1, sp + 244 0x00 0x19 0x02 0x17 0x20 0x00 0x4f 0x86 0x0e 0x02 0xd0 0x91 0x60 0x7e MOVA r25, #0; PADDB [p1], m0; MOVS p6, p1; MOVXM p0, #651488 + 258 0x00 0x2c 0xfc 0x57 0x20 0x20 0x05 0x61 0x00 0x0b 0xff 0x93 0xb0 0x7e NOPA; PADDB [p6], m1; ST p1, [sp, #-4]; MOVX r16, #1; MOV r24, #0 +.label TGT_F_main_48 +.loop_nesting 1 + 272 0x08 0x4f 0x3e 0x98 ST.TM r25, [p0], #16 + 276 0x08 0xcf 0x3e 0x98 ST.TM r25, [p0], #-16 + 280 0x00 0x00 NOPX + 282 0x00 0x00 NOPX + 284 0x00 0x00 NOPX + 286 0x17 0xc3 0x08 0x18 ACQ #62, r16 + 290 0xf4 0x9c 0x80 0x00 0x01 0xef 0xb1 0x40 0x10 0xba MOVA m7, #-92; MOVXM p7, #504448 + 300 0x00 0x00 NOPX + 302 0x00 0x00 NOPX + 304 0x0f 0xf8 0x1d 0x98 ST p0, [sp, #-8] + 308 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 312 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 316 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 320 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 324 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 328 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 332 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 336 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 340 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 344 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 348 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 352 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 356 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 360 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 364 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 368 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 372 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 376 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 380 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 384 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 388 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 392 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 396 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 400 0x09 0xeb 0x31 0x98 ST r25, [p1], m7 + 404 0x00 0x2c 0xf0 0x00 0x20 0x00 0x00 0x03 0xfa 0x4e 0xc1 0x36 NOPA; NOPB; ST p1, [sp, #-12]; NOPX +.label TGT_F_main_192 +.loop_nesting 2 + 416 0xe3 0xc6 0xd0 0x00 0x00 0x00 0x78 0xf8 0x10 0xba LDA r17, [p7], #4; MOVXM ls, #496 + 426 0xe3 0xaa 0xd0 0x00 0x00 0x01 0xb9 0x28 0x10 0xba LDA r10, [p7], #4; MOVXM le, #592 + 436 0x07 0x1d 0x36 0x98 LDA r9, [p7], #4 + 440 0x07 0x1d 0x16 0x98 LDA r8, [p7], #4 + 444 0x07 0x2e 0x76 0x98 LDA r19, [p7], #8 + 448 0xff 0x93 0x24 0xdd 0x81 0xd4 LDA p1, [sp, #-4]; MOV p2, p7 + 454 0x00 0x00 NOPX + 456 0x04 0x00 0x27 0x31 0x39 0xe4 MOVX r16, #0; MOV el7, r24 + 462 0x1c 0x90 0x9c 0xf8 MOV el9, r16 + 466 0x1d 0x14 0xa9 0x58 ADD.NC r20, r9, r10 + 470 0x00 0x2c 0xf0 0x98 0x8b 0x02 0x8d 0x10 0xa2 0xba NOPA; MOVS p0, p6; ADD.NC r20, r20, r8 + 480 0x00 0xd2 0x00 0x00 0x20 0x01 0x5b 0x01 0x30 0x0a 0xbc 0xe8 0xa8 0x00 0x00 0xe1 MOVA r18, #6; NOPB; NOPS; MOVX r19, #0; ADD.NC lc, r19, r20; NOPV +.label ZLS_F_main_272 +.loop_nesting 3 +.begin_of_loop + 496 0x43 0xe4 0xd1 0x1d 0xe9 0x82 0x6c 0xc0 0x42 0xba LDA dn6, [p2], #4; ST el7, [p1], #4; ADD.NC r19, r19, #1 + 506 0x43 0xd2 0xd9 0x32 0x04 0x14 LDA r20, [p2], #4; ADD.NC r18, r18, #4 + 512 0x02 0x1f 0xe6 0x98 LDA dc7, [p2], #4 + 516 0x02 0x1e 0xee 0x98 LDA el11, [p2], #4 + 520 0x00 0x00 NOPX + 522 0x00 0x00 NOPX + 524 0x00 0x00 NOPX + 526 0x09 0x1f 0x21 0x98 ST dn6, [p1], #4 + 530 0x00 0x00 NOPX + 532 0x09 0x1f 0xe1 0x98 ST dc7, [p1], #4 + 536 0x23 0xcd 0x30 0x00 0x01 0xa5 0x70 0x02 ST el9, [p1], #4; NOPM + 544 0x00 0x2c 0xf0 0x00 0x21 0x1e 0x91 0x80 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; ST r20, [p1], #4; NOPX; NOPM; NOPV + 560 0x00 0x2c 0xf0 0x00 0x21 0xbe 0xe9 0x80 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; ST el11, [p1], #-20; NOPX; NOPM; NOPV + 576 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x02 0xa9 0x60 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; MOV r21, p1; NOPV +.label ZLE_F_main_368 +.end_of_loop + 592 0x00 0x2c 0xf0 0x00 0x20 0x1e 0xb1 0x80 0x00 0x00 0xb5 0x46 0x08 0x00 0x00 0xe1 NOPA; NOPB; ST r21, [p0], #4; NOPX; ADD.NC p1, r21, #24; NOPV +.loop_nesting 2 + 608 0x00 0x0b 0x00 0x27 0x46 0x6e 0x6f 0x60 0x78 0xba MOVA r11, #0; LSHL r20, r19, r12; MOV r19, p7 + 618 0x9b 0x74 0x69 0xb2 0x01 0x24 ADD r13, r19, #-24; ADD.NC r19, r18, #1 + 624 0xf1 0xa1 0x60 0x25 0x26 0x6f 0x45 0x10 0x79 0x3a MOVS p7, r13; LSHL r18, r18, r12; MOV dj6, r20 + 634 0x9c 0x99 0xbb 0x12 0x41 0xe4 LSHL r18, r19, r12; MOV dj5, r18 + 640 0xf4 0x4a 0xde 0x8d 0x92 0x94 LDA r18, [p7, dj5]; ADD.NC dn7, r13, r18 + 646 0xd8 0x74 0x30 0x02 0x2c 0x7f 0xc0 0x02 ST dn7, [p6, dj6]; ADD.NC r17, r17, #-1 + 654 0x00 0x00 NOPX + 656 0x00 0x00 NOPX + 658 0x00 0x00 NOPX + 660 0x00 0x00 NOPX + 662 0x00 0x00 NOPX + 664 0x00 0x2b 0x60 0x02 0x4c 0xe4 0xa0 0x02 NOPS; ADD.NC r18, r19, r18 +.label TGT_F_main_448 +.loop_nesting 3 + 672 0x14 0xa1 0x01 0x98 SUB r16, r18, r16 + 676 0x14 0x1e 0xcd 0x98 LSHL r15, r16, r12 + 680 0xf1 0xa1 0x60 0x03 0xc3 0xd0 0x70 0x02 MOVS p7, r13; MOV dj7, r15 + 688 0x07 0xe0 0x16 0x98 LDA r0, [p7, dj7] +.no_stack_arguments + 692 0x00 0x02 0xa0 0x00 0x01 0x04 JL #1344 +.delay_slot + 698 0x1b 0x98 0x00 0x98 ADD.NC r14, r16, #1 +.delay_slot +.swstall delay_slot + 702 0x00 0x00 NOPX +.delay_slot + 704 0x13 0xa0 0xcd 0x98 LSHL r16, r14, r12 +.delay_slot + 708 0x18 0x66 0xc1 0x58 ADD.NC p0, r13, r16 +.delay_slot + 712 0xfe 0x46 0xb0 0x00 0x01 0xa5 0x70 0x02 ST r17, [sp, #-16]; NOPM +.return_address + 720 0x04 0x40 0xa1 0x0f 0x41 0xe4 MOVX r17, #1; MOV dj0, r15 + 726 0xe0 0x4a 0xd0 0x00 0x00 0x1e 0x8b 0xe0 0x10 0xba LDA r18, [p7, dj0]; MOVXM r20, #30656 + 736 0x00 0x00 0x7a 0xaf 0x84 0x44 MOVXM r21, #30658 + 742 0x00 0x08 0x00 0xc0 0x00 0x44 MOVXM p0, #524288 + 748 0x00 0x70 0x08 0x20 0x06 0x44 MOVXM r16, #7340035 + 754 0x00 0x00 NOPX + 756 0x00 0x00 NOPX + 758 0x00 0x00 NOPX + 760 0x14 0xa3 0x1d 0x98 LSHL r17, r18, r17 + 764 0x1c 0x98 0xb9 0x58 ADD.NC r18, r17, r14 + 768 0x14 0xa6 0xcd 0x98 LSHL r19, r18, r12 + 772 0x18 0x89 0xa0 0xf8 MOV dj0, r19 + 776 0x07 0x02 0xd6 0x98 LDA r22, [p7, dj0] + 780 0x00 0x00 NOPX + 782 0x00 0x00 NOPX + 784 0x00 0x00 NOPX + 786 0x00 0x00 NOPX + 788 0x00 0x00 NOPX + 790 0x00 0x00 NOPX + 792 0x15 0xb6 0xcc 0x98 LTU r27, r22, r12 + 796 0x15 0x69 0x42 0x18 SEL.EQZ r20, r21, r20, r27 + 800 0x1d 0x1b 0x51 0x58 ADD.NC r20, r22, r20 + 804 0x15 0x28 0xcd 0x98 LSHL r20, r20, r12 + 808 0x00 0x2b 0x60 0x00 0x45 0x10 0x70 0x02 NOPS; MOV dj0, r20 +.label TGT_F_main_592 +.loop_nesting 4 + 816 0x00 0x02 0x93 0x98 LDA.TM r20, [p0, dj0] + 820 0x00 0x00 NOPX + 822 0x00 0x00 NOPX + 824 0x00 0x00 NOPX + 826 0x00 0x00 NOPX + 828 0x00 0x00 NOPX + 830 0x00 0x00 NOPX + 832 0x15 0x2b 0x04 0x98 AND r21, r20, r16 + 836 0xa8 0x01 0x98 0x40 0x01 0x84 JNZ r21, #816 +.delay_slot +.swstall delay_slot + 842 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 844 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 846 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 848 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 850 0x00 0x00 NOPX +.loop_nesting 3 + 852 0x1b 0xd9 0x00 0x98 ADD.NC r15, r18, #1 + 856 0x13 0xdc 0xcd 0x98 LSHL r14, r15, r12 + 860 0x19 0x87 0x20 0xf8 MOV dj1, r14 + 864 0x07 0x20 0x16 0x98 LDA r0, [p7, dj1] +.no_stack_arguments + 868 0x00 0x02 0xa0 0x00 0x01 0x04 JL #1344 +.delay_slot +.swstall delay_slot + 874 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 876 0x00 0x00 NOPX +.delay_slot + 878 0x13 0x5a 0x23 0x18 ADD r13, r13, #8 +.delay_slot + 882 0x18 0x69 0xb5 0x58 ADD.NC p0, r19, r13 +.delay_slot + 886 0x00 0x2c 0xf7 0xea 0x35 0x80 0x00 0x00 0x00 0x7a NOPA; ST r17, [sp, #-24]; NOPX +.return_address + 896 0x00 0x0e 0x00 0x01 0x00 0x28 0x43 0x90 0x78 0xba MOVA r14, #0; MOVX r16, #1; MOV dj0, r14 + 906 0xe0 0x4a 0xd8 0xad 0xfc 0x14 LDA r18, [p7, dj0]; ADD.NC r17, r13, #-4 + 912 0x0f 0xee 0x35 0x98 ST r17, [sp, #-20] + 916 0x00 0x00 NOPX + 918 0x00 0x00 NOPX + 920 0x00 0x00 NOPX + 922 0x00 0x00 NOPX + 924 0x00 0x00 NOPX + 926 0x14 0xa5 0x0d 0x98 LSHL r18, r18, r16 + 930 0x14 0xa1 0x05 0x98 OR r16, r18, r16 + 934 0x1c 0x97 0xc1 0x58 ADD.NC r18, r15, r16 + 938 0xfc 0x42 0xb0 0x24 0xc6 0x6c 0x37 0x60 0x79 0x3a ST r16, [sp, #-32]; LSHL r12, r18, r12; MOV p0, p7 + 948 0xfc 0xca 0xb0 0x23 0x06 0x04 0x43 0x10 0x79 0x3a ST r18, [sp, #-28]; ADD r16, r17, r12; MOV dj0, r12 + 958 0x00 0x42 0xd7 0xde 0x15 0x80 0x00 0x03 0xb1 0xf0 0x10 0x76 LDA r16, [p0, dj0]; ST r16, [sp, #-36]; MOVXM p7, #992 + 970 0x00 0x00 NOPX + 972 0x00 0x00 NOPX + 974 0x00 0x00 NOPX + 976 0x00 0x00 NOPX + 978 0x00 0x00 NOPX + 980 0x00 0x00 NOPX + 982 0x00 0x2c 0xf0 0x00 0x10 0x01 0xec 0x3f 0xce 0xba NOPA; NOPB; ADD.NC r15, r16, #-1 +.label TGT_F_main_768 +.loop_nesting 4 + 992 0x12 0xe0 0xe5 0x98 OR r16, r11, r14 + 996 0x80 0x02 0x28 0x40 0x01 0x84 JNZ r16, #1104 +.delay_slot +.swstall delay_slot + 1002 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1004 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1006 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1008 0x00 0x00 NOPX +.delay_slot + 1010 0x11 0xa1 0x60 0x00 0xc3 0x10 0x70 0x02 MOVS p0, r13; MOV dj1, r12 + 1018 0x07 0xdc 0x99 0x18 LDA p1, [sp, #-36] + 1022 0x00 0x00 NOPX + 1024 0x00 0x00 NOPX + 1026 0x00 0x00 NOPX + 1028 0x00 0x00 NOPX + 1030 0x00 0x00 NOPX + 1032 0x00 0x00 NOPX + 1034 0x01 0x06 0x16 0x98 LDA r16, [p1] + 1038 0x00 0x00 NOPX + 1040 0x00 0x00 NOPX + 1042 0x00 0x00 NOPX + 1044 0x00 0x00 NOPX + 1046 0x00 0x00 NOPX + 1048 0x00 0x00 NOPX + 1050 0x80 0x02 0x28 0x00 0x01 0x84 JZ r16, #1104 +.delay_slot +.swstall delay_slot + 1056 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1058 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1060 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1062 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1064 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1066 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1068 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1070 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1072 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1074 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1076 0x00 0x00 NOPX + 1078 0x10 0x08 0x00 0x18 DONE +.swstall chess_separator_scheduler + 1082 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1084 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1086 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1088 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1090 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 1092 0x00 0x2c 0xf0 0x00 0x20 0x00 0x00 0x00 0x00 0xad 0x81 0x36 NOPA; NOPB; NOPS; NOPX +.label TGT_F_main_880 + 1104 0x04 0x02 0xd0 0xd9 0x81 0xd4 LDA r0, [p0, dj1]; MOV p0, p6 +.no_stack_arguments + 1110 0x00 0x04 0x98 0x00 0x01 0x04 JL #2352 +.delay_slot +.swstall delay_slot + 1116 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1118 0x00 0x00 NOPX +.delay_slot + 1120 0x18 0x55 0x20 0xf8 MOV r1, r10 +.delay_slot + 1124 0x18 0x94 0xa0 0xf8 MOV r2, r9 +.delay_slot + 1128 0x00 0x2b 0x60 0x00 0x6a 0x10 0x70 0x02 NOPS; MOV r3, r8 +.return_address + 1136 0x13 0xdf 0xe0 0x18 JNZD r15, r15, p7 +.delay_slot +.swstall delay_slot + 1140 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1142 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1144 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1146 0x00 0x00 NOPX +.delay_slot + 1148 0x1b 0x97 0x00 0x98 ADD.NC r14, r14, #1 +.loop_nesting 3 + 1152 0xfe 0x46 0x20 0x00 0xc0 0x49 0x6a 0xc0 0x48 0xba LDA r17, [sp, #-16]; MOVX r12, #2; ADD.NC r11, r11, #1 + 1162 0xfd 0xca 0x20 0x00 0x00 0x03 0xb1 0x50 0x10 0xba LDA r18, [sp, #-20]; MOVXM p7, #672 + 1172 0x07 0xe6 0x71 0x18 LDA r19, [sp, #-28] + 1176 0x07 0xe2 0x11 0x18 LDA r16, [sp, #-32] + 1180 0x07 0xea 0x91 0x18 LDA r20, [sp, #-24] + 1184 0x00 0x00 NOPX + 1186 0x00 0x00 NOPX + 1188 0x14 0x63 0xe0 0x18 JNZD r17, r17, p7 +.delay_slot + 1192 0x14 0x9b 0xf3 0x18 ADD r13, r18, #-4 +.delay_slot + 1196 0x1c 0x99 0x81 0x98 ADD.NC r18, r19, #3 +.delay_slot +.swstall delay_slot + 1200 0x00 0x00 NOPX +.delay_slot + 1202 0x1c 0x1a 0x41 0x58 ADD.NC r16, r20, r16 +.delay_slot + 1206 0x1c 0x18 0x02 0x98 ADD.NC r16, r16, #5 +.loop_nesting 2 + 1210 0x07 0xd2 0x07 0x8d 0x0b 0x25 0x06 0x6e 0x88 0x3f 0x58 0x76 MOVA r18, #62; MOVS p7, r13; LSHL r16, r18, r12; MOV r20, #63 + 1222 0x00 0x30 0x00 0x3f 0x67 0xe8 0x44 0x10 0x78 0xba MOVA r16, #1; MOVX r22, #-1; MOV dj0, r16 + 1232 0xe0 0x46 0xd0 0x00 0x01 0xee 0x69 0x40 0x10 0xba LDA r17, [p7, dj0]; MOVXM r19, #504448 + 1242 0x9e 0xda 0xfc 0x20 0x01 0x64 EQ r27, r19, r13; MOV r24, #0 + 1248 0x15 0x2f 0x22 0x18 SEL.EQZ r23, r20, r18, r27 + 1252 0x00 0x07 0xba 0xad 0x00 0x44 MOVXM r21, #505472 + 1258 0x00 0x00 NOPX + 1260 0x00 0x00 NOPX + 1262 0x00 0x00 NOPX + 1264 0x14 0x74 0xd0 0x18 EQZ r26, r17 + 1268 0x15 0xd1 0x68 0x18 REL r23, r22 + 1272 0x14 0xa5 0x42 0x18 SEL.EQZ r18, r18, r20, r27 + 1276 0x00 0x00 NOPX + 1278 0x00 0x00 NOPX + 1280 0x14 0x97 0x08 0x18 ACQ.COND r18, r16, r26 + 1284 0x88 0x00 0xd0 0x00 0x01 0x84 JZ r17, #416 +.delay_slot + 1290 0x14 0xe1 0x52 0x18 SEL.EQZ r16, r19, r21, r27 +.delay_slot + 1294 0x1f 0x68 0x20 0xf8 MOV p7, r16 +.delay_slot +.swstall delay_slot + 1298 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1300 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1302 0x00 0x00 NOPX +.loop_nesting 1 + 1304 0xff 0x03 0x20 0x01 0x90 0x0a 0x08 0x01 0x58 0xba LDA p0, [sp, #-8]; MOVX r25, #0; MOV r16, #1 + 1314 0x00 0x00 0x88 0x00 0x00 0x84 J #272 +.delay_slot +.swstall delay_slot + 1320 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1322 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1324 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1326 0x00 0x00 NOPX +.delay_slot + 1328 0x07 0xf4 0x99 0x18 LDA p1, [sp, #-12] +.label _main__end +.label _main___func_end0 + +.text_segment PM 1344 +.label __ZN3adf11block_writeEPKNS_7reg_valEj___func_begin0 +.label _ZN3adf11block_writeEPKNS_7reg_valEj +.function_start + 1344 0x00 0x02 0xf0 0x00 0x01 0x84 JZ r0, #1504 +.delay_slot +.swstall delay_slot + 1350 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1352 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1354 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1356 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1358 0x00 0x00 NOPX + 1360 0xf8 0x1e 0x2a 0xe0 0x41 0xe4 MOVX r0, #-4; MOV lc, r0 + 1366 0x00 0x00 0x01 0xea 0xe0 0x44 MOVXM ls, #1392 + 1372 0x00 0x00 0x06 0xeb 0xa0 0x44 MOVXM le, #1488 + 1378 0x00 0x2c 0xf0 0x00 0x20 0x00 0x40 0x16 0x00 0x02 0x00 0x2b 0x60 0x7e NOPA; NOPB; NOPS; MOVXM p1, #524288 +.label ZLS_F_ZN3adf11block_writeEPKNS_7reg_valEj_48 +.loop_nesting 1 +.begin_of_loop + 1392 0x00 0x1c 0x56 0x98 LDA r2, [p0], #4 + 1396 0x00 0x1c 0x36 0x98 LDA r1, [p0], #4 + 1400 0x00 0x00 NOPX + 1402 0x00 0x2c 0xf0 0x00 0x20 0x3c NOPA; NOPB + 1408 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; NOPM; NOPV + 1424 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; NOPM; NOPV + 1440 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; NOPM; NOPV + 1456 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x04 0x30 0x24 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; AND r3, r2, r0; NOPM; NOPV + 1472 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x00 0x40 0xd0 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; MOV dj0, r3; NOPV +.label ZLE_F_ZN3adf11block_writeEPKNS_7reg_valEj_144 +.end_of_loop + 1488 0x00 0x2c 0xf0 0x00 0x21 0x00 0x3e 0x80 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; ST.TM r1, [p1, dj0]; NOPX; NOPM; NOPV +.label TGT_F_ZN3adf11block_writeEPKNS_7reg_valEj_160 +.loop_nesting 0 + 1504 0x10 0x28 0x00 0x18 RET lr +.delay_slot +.swstall delay_slot + 1508 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1510 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1512 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1514 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1516 0x00 0x00 NOPX +.label _ZN3adf11block_writeEPKNS_7reg_valEj__end +.label __ZN3adf11block_writeEPKNS_7reg_valEj___func_end0 + +.text_segment PM 1520 +.label _fini +.function_start + 1520 0x00 0x08 0x00 0x00 0x01 0xc4 PADDXM [sp], #64 + 1526 0xfe 0xba 0xb0 0x00 0x00 0x02 0x08 0x00 0x11 0x3a ST r14, [sp, #-12]; MOVXM r16, #0 + 1536 0xff 0x73 0xb0 0x00 0x00 0x01 0xc8 0x00 0x11 0x3a ST p7, [sp, #-8]; MOVXM r14, #0 + 1546 0x74 0x20 0xfe 0xd0 0x41 0xe4 EQ r16, r14, r16; MOV p7, r16 + 1552 0x80 0x03 0x38 0x40 0x01 0x84 JNZ r16, #1648 +.delay_slot + 1558 0x0f 0xfd 0xf5 0x98 ST r15, [sp, #-4] +.delay_slot + 1562 0x0f 0xf0 0x3d 0x98 ST lr, [sp, #-16] +.delay_slot +.swstall delay_slot + 1566 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1568 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1570 0x00 0x2c 0xf0 0x02 0xb6 0x00 0x00 0x40 0x1a 0x57 0x00 0x00 0x1c 0x2e NOPA; NOPS; NOPM; NOPV +.label TGT_F_fini_64 +.loop_nesting 1 + 1584 0x07 0x1c 0x1e 0x98 LDA p0, [p7], #4 + 1588 0x00 0x00 NOPX + 1590 0x00 0x00 NOPX + 1592 0x00 0x00 NOPX + 1594 0x00 0x00 NOPX + 1596 0x00 0x00 NOPX + 1598 0x00 0x00 NOPX +.no_stack_arguments + 1600 0x10 0x30 0x00 0x18 JL p0 +.delay_slot + 1604 0x1b 0xde 0xc0 0xf8 MOV r15, p7 +.delay_slot +.swstall delay_slot + 1608 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1610 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1612 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1614 0x00 0x00 NOPX +.return_address + 1616 0x13 0xe0 0xe8 0x98 NE r16, r15, r14 + 1620 0x80 0x03 0x18 0x40 0x01 0x84 JNZ r16, #1584 +.delay_slot +.swstall delay_slot + 1626 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1628 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1630 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1632 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1634 0x00 0x2c 0xf0 0x02 0xb6 0x00 0x00 0x40 0x1a 0x57 0x00 0x00 0x1c 0x2e NOPA; NOPS; NOPM; NOPV +.label TGT_F_fini_128 +.loop_nesting 0 + 1648 0x07 0xf0 0x39 0x18 LDA lr, [sp, #-16] + 1652 0x07 0xf5 0xd1 0x18 LDA r14, [sp, #-12] + 1656 0x07 0xfd 0xf1 0x18 LDA r15, [sp, #-4] + 1660 0x07 0xfb 0x99 0x18 LDA p7, [sp, #-8] + 1664 0x00 0x00 NOPX + 1666 0x00 0x00 NOPX + 1668 0x00 0x00 NOPX + 1670 0x10 0x28 0x00 0x18 RET lr +.delay_slot + 1674 0xff 0xf8 0x00 0x00 0x01 0xc4 PADDXM [sp], #-64 +.delay_slot +.swstall delay_slot + 1680 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1682 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1684 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1686 0x00 0x00 NOPX +.label _fini__end + +.text_segment PM 1696 +.label __cxa_finalize +.function_start + 1696 0xff 0xa1 0x00 0x00 0x02 0x00 0x00 0x00 0x70 0xba MOVA r1, #-3; PADDXM [sp], #64 + 1706 0xfb 0x07 0xb0 0x00 0x41 0x08 0x6d 0xe0 0x79 0x3a ST lr, [sp, #-40]; MOVX r4, #8; MOV r3, packSign1 + 1716 0xfa 0x8e 0xb0 0x00 0x69 0xe0 0x70 0x02 ST r3, [sp, #-44]; MOV r3, packSign0 + 1724 0xfc 0x0e 0xb0 0x00 0x6e 0x30 0x70 0x02 ST r3, [sp, #-32]; MOV r3, unpackSign1 + 1732 0xfc 0x8e 0xb0 0x00 0x6a 0x30 0x70 0x02 ST r3, [sp, #-28]; MOV r3, unpackSign0 + 1740 0xfd 0x8e 0xb0 0x00 0x01 0xf0 0x32 0x10 0x11 0x3a ST r3, [sp, #-20]; MOVXM p0, #508960 + 1750 0x00 0x8a 0xd1 0xb1 0xc1 0xd4 LDA r2, [p0]; MOV r3, crSRSMode + 1756 0xfe 0x0e 0xb0 0x00 0x6b 0xb0 0x70 0x02 ST r3, [sp, #-16]; MOV r3, crPackSize + 1764 0xfe 0x8e 0xb0 0x00 0x6a 0x70 0x70 0x02 ST r3, [sp, #-12]; MOV r3, crSat + 1772 0x00 0x92 0x30 0x01 0x2c 0x30 0x70 0x02 ST r4, [p0]; MOV r9, upsSign1 + 1780 0xff 0x0e 0xb0 0x01 0x08 0x30 0x70 0x02 ST r3, [sp, #-8]; MOV r8, upsSign0 + 1788 0x1a 0xda 0x60 0xf8 MOV r11, vaddSign1 + 1792 0x1a 0x92 0x60 0xf8 MOV r10, vaddSign0 + 1796 0x10 0xbc 0x66 0x3f 0x81 0xe4 ADD r2, r2, #-8; MOV r12, srsSign1 + 1802 0x13 0x43 0xd0 0xa5 0xc1 0xe4 ASHL r13, r2, r1; MOV r1, crUnpackSize + 1808 0xff 0x86 0xb0 0x00 0x2f 0xb0 0x70 0x02 ST r1, [sp, #-4]; MOV r1, crRnd + 1816 0xfd 0x06 0xb0 0x00 0x2e 0x70 0x70 0x02 ST r1, [sp, #-24]; MOV r1, crUPSMode + 1824 0xfb 0x86 0xb0 0x01 0xcb 0xe0 0x70 0x02 ST r1, [sp, #-36]; MOV r14, srsSign0 + 1832 0x00 0x07 0xc0 0xa8 0x00 0x44 MOVXM r1, #508928 + 1838 0x1e 0x60 0x89 0x58 ADD.NC p6, r1, r2 + 1842 0x00 0x2c 0xf0 0x00 0x20 0x00 0x00 0x76 0x74 0x02 0x00 0x2b 0x60 0x7e NOPA; NOPB; NOPS; MOVXM p7, #1856 +.label TGT_F__cxa_finalize_160 +.loop_nesting 1 + 1856 0x06 0x1c 0x9e 0x98 LDA p1, [p6], #4 + 1860 0x06 0xdc 0x1e 0x98 LDA p0, [p6], #-12 + 1864 0x00 0x00 NOPX + 1866 0x00 0x00 NOPX + 1868 0x00 0x00 NOPX + 1870 0x00 0x00 NOPX + 1872 0x00 0x00 NOPX +.no_stack_arguments + 1874 0x10 0x30 0x40 0x18 JL p1 +.delay_slot + 1878 0x1b 0xd0 0x20 0xf8 MOV r15, r0 +.delay_slot +.swstall delay_slot + 1882 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1884 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1886 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1888 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; NOPM; NOPV +.return_address + 1904 0x13 0x5b 0xe0 0x18 JNZD r13, r13, p7 +.delay_slot + 1908 0x18 0x17 0xa0 0xf8 MOV r0, r15 +.delay_slot +.swstall delay_slot + 1912 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1914 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1916 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1918 0x00 0x00 NOPX +.loop_nesting 0 + 1920 0xfb 0x07 0x20 0x13 0xea 0x02 0x9a 0xd0 0x78 0xba LDA lr, [sp, #-40]; MOVX upsSign1, r9; MOV vaddSign1, r11 + 1930 0xfa 0x86 0x20 0x11 0xca 0x00 0x9a 0x90 0x78 0xba LDA r1, [sp, #-44]; MOVX upsSign0, r8; MOV vaddSign0, r10 + 1940 0xfc 0x0a 0x26 0x79 0x80 0x2c LDA r2, [sp, #-32]; MOVX srsSign1, r12 + 1946 0xfc 0x8e 0x27 0x71 0x80 0x2c LDA r3, [sp, #-28]; MOVX srsSign0, r14 + 1952 0x07 0xec 0xf1 0x18 LDA r7, [sp, #-20] + 1956 0x07 0xf0 0x91 0x18 LDA r4, [sp, #-16] + 1960 0x07 0xf4 0xb1 0x18 LDA r5, [sp, #-12] + 1964 0x07 0xf8 0xd1 0x18 LDA r6, [sp, #-8] + 1968 0xff 0x86 0x20 0xf8 0x80 0x2c LDA r1, [sp, #-4]; MOVX packSign1, r1 + 1974 0xfd 0x0a 0x21 0x70 0x80 0x2c LDA r2, [sp, #-24]; MOVX packSign0, r2 + 1980 0xfb 0x8e 0x21 0xfb 0x80 0x2c LDA r3, [sp, #-36]; MOVX unpackSign1, r3 + 1986 0xff 0xf8 0x00 0x00 0x01 0xc4 PADDXM [sp], #-64 + 1992 0x05 0x00 0x04 0x67 0x41 0xe4 RET lr; MOV unpackSign0, r7 +.delay_slot + 1998 0x27 0x60 0x07 0x65 0x41 0xe4 MOVX crSRSMode, r4; MOV crPackSize, r5 +.delay_slot + 2004 0x11 0xbb 0x80 0x18 MOVX crSat, r6 +.delay_slot + 2008 0x10 0x7c 0x80 0x18 MOVX crUnpackSize, r1 +.delay_slot + 2012 0x10 0xba 0x80 0x18 MOVX crRnd, r2 +.delay_slot + 2016 0x10 0xfc 0x00 0x18 MOVX crUPSMode, r3 +.label __cxa_finalize__end + +.bss_segment DMb 504448 2048 + +.data_segment DMb 508928 +.label _ZL7atexits + 0xf0 + 0x5 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 + 0x0 +.label _ZL10atexit_cnt + 0x8 + 0x0 + 0x0 + 0x0 + +.stack DM_stack 506560 508928 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.map b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.map new file mode 100644 index 0000000000000000000000000000000000000000..ef46d1312460834f68715f0d9820c058bd977f34 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.map @@ -0,0 +1,143 @@ + +// File generated by bridge version V-2024.06#84922c0d9f#241219, Fri Mar 21 03:42:51 2025 +// Copyright 2014-2024 Synopsys, Inc. All rights reserved. +// bridge -o../Release/0_0 ../Release/0_0.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork848 -pme + +// Release: ipp V-2024.06-TGT-241219 + +Memory map for memory 'DM_bankA': + + Size = 1048576 + Width = 8 bits + Offset = 0 + Used = 36 + + 0x0007c400..0x0007c41f ( 32 items) : atexit.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release/libc.a)::_ZL7atexits (Data, Local, .data.DM_bankA.4) + + Called functions : _fini + + 0x0007c420..0x0007c423 ( 4 items) : atexit.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release/libc.a)::_ZL10atexit_cnt (Data, Local, .data.DM_bankA.4) + +Memory map for memory 'DM_stack': + + Size = 1048576 + Width = 8 bits + Offset = 0 + Used = 2368 + + 0x0007bac0..0x0007c3ff ( 2368 items) : Stack + +Memory map for memory 'DMb': + + Size = 1048576 + Width = 8 bits + Offset = 0 + Used = 4452 + + 0x00000000..0x0007b27f ( 504448 items) : Reserved + 0x0007b280..0x0007b67f ( 1024 items) : ../Release/0_0.o::lcpPing (Data, Global, .bss.DMb.4) + 0x0007b680..0x0007ba7f ( 1024 items) : ../Release/0_0.o::lcpPong (Data, Global, .bss.DMb.4) + 0x0007ba80..0x0007babf ( 64 items) : Reserved + 0x0007bac0..0x0007c3ff ( 2368 items) : Stack + 0x0007c400..0x0007c41f : Occupied in alias or record memory 'DM_bankA' by symbol '_ZL7atexits' + 0x0007c420..0x0007c423 : Occupied in alias or record memory 'DM_bankA' by symbol '_ZL10atexit_cnt' + 0x0007c440..0x000fffff ( 539584 items) : Reserved + +Memory map for memory 'PM': + + Size = 1048576 + Width = 8 bits + Offset = 0 + Used = 1998 + + 0x00000000..0x000000df ( 224 items) : me_basic.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release/libme.a)::_main_init (Function, Global, .text) (stack frame size = 0) + + Called functions : _main + __cxa_finalize + + Referenced symbols: _sp_start_value_DM_stack + _ctors_start + _ctors_end + + 0x000000e0..0x00000533 ( 1108 items) : ../Release/0_0.o::_main (Function, Global, .text) (stack frame size = 192) + + Called functions : _ZN3adf11block_writeEPKNS_7reg_valEj + _Z13kernelWrapperPPvjjjj + + Referenced symbols: lcpPing + lcpPong + + 0x00000540..0x000005ed ( 174 items) : ../Release/0_0.o::_ZN3adf11block_writeEPKNS_7reg_valEj (Function, Weak, .text) (stack frame size = 0) + 0x000005f0..0x00000697 ( 168 items) : me_basic.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release/libme.a)::_fini (Function, Global, .text) (stack frame size = 64) + + Referenced symbols: _dtors_start + _dtors_end + + 0x000006a0..0x000007e3 ( 324 items) : atexit.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release/libc.a)::__cxa_finalize (Function, Global, .text) (stack frame size = 64) + + Referenced symbols: _ZL10atexit_cnt + _ZL7atexits + + 0x00000930..0x00003fff ( 14032 items) : Reserved + +External symbols: + + _Z13kernelWrapperPPvjjjj = 0x930 + __dso_handle = 0x0 + _ctors_end = 0x0 + _ctors_start = 0x0 + _dtors_end = 0x0 + _dtors_start = 0x0 + _pc_end = 0x7e4 + _pc_start = 0x0 + _sp_end_DM_stack = 0x7c400 + _sp_start_DM_stack = 0x7bac0 + +Section summary for memory 'DM_bankA': + + .data File + ---------- ---------- + 36 atexit.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release/libc.a) + ---------- ---------- + 36 Total + +Section summary for memory 'DM_stack': + + .stack File + ---------- ---------- + 2368 + ---------- ---------- + 2368 Total + +Section summary for memory 'DMb': + + .bss .data File + ---------- ---------- ---------- + 2048 0 ../Release/0_0.o + 0 36 atexit.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release/libc.a) (in DM_bankA) + ---------- ---------- ---------- + 2048 36 Total + +Section summary for memory 'PM': + + .text File + ---------- ---------- + 1282 ../Release/0_0.o + 392 me_basic.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release/libme.a) + 324 atexit.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release/libc.a) + ---------- ---------- + 1998 Total + +File summary: + +atexit.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release/libc.a) + DM_bankA 36 + PM 324 + +../Release/0_0.o + DMb 2048 + PM 1282 + +me_basic.o(/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release/libme.a) + PM 392 + diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.o b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.o new file mode 100644 index 0000000000000000000000000000000000000000..7dd7ec15d6bdaf5c3a039e0cf9ff171994e85272 Binary files /dev/null and b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.o differ diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.o.lst b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.o.lst new file mode 100644 index 0000000000000000000000000000000000000000..bf97e4034ea94a75321e185530dc7397ba532efb --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.o.lst @@ -0,0 +1,471 @@ + +// File generated by darts version V-2024.06#84922c0d9f#241219, Fri Mar 21 03:42:50 2025 +// Copyright 2014-2024 Synopsys, Inc. All rights reserved. +// darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0.o me + +// Release: ipp V-2024.06-TGT-241219 +.data_segment_name +.bss local .at 524288 _ZL22TM_Address_Space_Start TM 4 + + +.data_segment_name +.bss global 4 lcpPing DMb 1024 + +.data_segment_name +.bss global 4 lcpPong DMb 1024 + +.undef global data _ZN12me_primitive11control_satE + +.undef global data _ZN12me_primitive11control_rndE + +.text_segment_name +.text weak 16 _ZN3adf11block_writeEPKNS_7reg_valEj +.label __ZN3adf11block_writeEPKNS_7reg_valEj___func_begin0 +.function_start + 0 0x00 0x00 0x50 0x00 0x01 0x84 JZ r0, #TGT_F_ZN3adf11block_writeEPKNS_7reg_valEj_160 +.delay_slot +.swstall delay_slot + 6 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 8 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 10 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 12 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 14 0x00 0x00 NOPX + 16 0xf8 0x1e 0x2a 0xe0 0x41 0xe4 MOVX r0, #-4; MOV lc, r0 + 22 0x00 0x00 0x01 0xe0 0x60 0x44 MOVXM ls, #(ZLS_F_ZN3adf11block_writeEPKNS_7reg_valEj_48 + 0) + 28 0x00 0x00 0x06 0xe1 0x20 0x44 MOVXM le, #(ZLE_F_ZN3adf11block_writeEPKNS_7reg_valEj_144 + 0) + 34 0x00 0x2c 0xf0 0x00 0x20 0x00 0x40 0x16 0x00 0x02 0x00 0x2b 0x60 0x7e NOPA; NOPB; NOPS; MOVXM p1, #524288 +.label ZLS_F_ZN3adf11block_writeEPKNS_7reg_valEj_48 +.loop_nesting 1 +.begin_of_loop + 48 0x00 0x1c 0x56 0x98 LDA r2, [p0], #4 + 52 0x00 0x1c 0x36 0x98 LDA r1, [p0], #4 + 56 0x00 0x00 NOPX + 58 0x00 0x2c 0xf0 0x00 0x20 0x3c NOPA; NOPB + 64 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; NOPM; NOPV + 80 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; NOPM; NOPV + 96 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; NOPM; NOPV + 112 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x04 0x30 0x24 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; AND r3, r2, r0; NOPM; NOPV + 128 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x00 0x40 0xd0 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; MOV dj0, r3; NOPV +.label ZLE_F_ZN3adf11block_writeEPKNS_7reg_valEj_144 +.end_of_loop + 144 0x00 0x2c 0xf0 0x00 0x21 0x00 0x3e 0x80 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; ST.TM r1, [p1, dj0]; NOPX; NOPM; NOPV +.label TGT_F_ZN3adf11block_writeEPKNS_7reg_valEj_160 +.loop_nesting 0 + 160 0x10 0x28 0x00 0x18 RET lr +.delay_slot +.swstall delay_slot + 164 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 166 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 168 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 170 0x00 0x00 NOPX +.label _ZN3adf11block_writeEPKNS_7reg_valEj__end last +.label __ZN3adf11block_writeEPKNS_7reg_valEj___func_end0 last +.delay_slot +.swstall delay_slot + 172 0x00 0x00 NOPX + +.text_segment_name +.text global 10 _main +.label _main___func_begin0 +.function_start + 0 0xec 0x00 0x80 0x00 0x06 0x00 0x00 0x00 0x70 0xba MOVA m0, #-160; PADDXM [sp], #192 + 10 0xe9 0x04 0x80 0x00 0xc0 0x48 0xb2 0xf0 0x78 0xba MOVA m1, #-184; MOVX r12, #2; MOV p1, sp + 20 0x00 0x19 0x02 0x17 0x20 0x00 0x4f 0x86 0x0e 0x02 0xd0 0x91 0x60 0x7e MOVA r25, #0; PADDB [p1], m0; MOVS p6, p1; MOVXM p0, #651488 + 34 0x00 0x2c 0xfc 0x57 0x20 0x20 0x05 0x61 0x00 0x0b 0xff 0x93 0xb0 0x7e NOPA; PADDB [p6], m1; ST p1, [sp, #-4]; MOVX r16, #1; MOV r24, #0 +.label TGT_F_main_48 +.loop_nesting 1 + 48 0x08 0x4f 0x3e 0x98 ST.TM r25, [p0], #16 + 52 0x08 0xcf 0x3e 0x98 ST.TM r25, [p0], #-16 + 56 0x00 0x00 NOPX + 58 0x00 0x00 NOPX + 60 0x00 0x00 NOPX + 62 0x17 0xc3 0x08 0x18 ACQ #62, r16 + 66 0xf4 0x9c 0x80 0x00 0x00 0x03 0xb0 0x00 0x10 0xba MOVA m7, #-92; MOVXM p7, #lcpPing + 76 0x00 0x00 NOPX + 78 0x00 0x00 NOPX + 80 0x0f 0xf8 0x1d 0x98 ST p0, [sp, #-8] + 84 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 88 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 92 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 96 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 100 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 104 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 108 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 112 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 116 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 120 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 124 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 128 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 132 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 136 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 140 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 144 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 148 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 152 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 156 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 160 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 164 0x09 0x1f 0x11 0x98 ST r24, [p1], #4 + 168 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 172 0x09 0x1f 0x31 0x98 ST r25, [p1], #4 + 176 0x09 0xeb 0x31 0x98 ST r25, [p1], m7 + 180 0x00 0x2c 0xf0 0x00 0x20 0x00 0x00 0x03 0xfa 0x4e 0xc1 0x36 NOPA; NOPB; ST p1, [sp, #-12]; NOPX +.label TGT_F_main_192 +.loop_nesting 2 + 192 0xe3 0xc6 0xd0 0x00 0x00 0x00 0x78 0x88 0x10 0xba LDA r17, [p7], #4; MOVXM ls, #(ZLS_F_main_272 + 0) + 202 0xe3 0xaa 0xd0 0x00 0x00 0x01 0xb8 0xb8 0x10 0xba LDA r10, [p7], #4; MOVXM le, #(ZLE_F_main_368 + 0) + 212 0x07 0x1d 0x36 0x98 LDA r9, [p7], #4 + 216 0x07 0x1d 0x16 0x98 LDA r8, [p7], #4 + 220 0x07 0x2e 0x76 0x98 LDA r19, [p7], #8 + 224 0xff 0x93 0x24 0xdd 0x81 0xd4 LDA p1, [sp, #-4]; MOV p2, p7 + 230 0x00 0x00 NOPX + 232 0x04 0x00 0x27 0x31 0x39 0xe4 MOVX r16, #0; MOV el7, r24 + 238 0x1c 0x90 0x9c 0xf8 MOV el9, r16 + 242 0x1d 0x14 0xa9 0x58 ADD.NC r20, r9, r10 + 246 0x00 0x2c 0xf0 0x98 0x8b 0x02 0x8d 0x10 0xa2 0xba NOPA; MOVS p0, p6; ADD.NC r20, r20, r8 + 256 0x00 0xd2 0x00 0x00 0x20 0x01 0x5b 0x01 0x30 0x0a 0xbc 0xe8 0xa8 0x00 0x00 0xe1 MOVA r18, #6; NOPB; NOPS; MOVX r19, #0; ADD.NC lc, r19, r20; NOPV +.label ZLS_F_main_272 +.loop_nesting 3 +.begin_of_loop + 272 0x43 0xe4 0xd1 0x1d 0xe9 0x82 0x6c 0xc0 0x42 0xba LDA dn6, [p2], #4; ST el7, [p1], #4; ADD.NC r19, r19, #1 + 282 0x43 0xd2 0xd9 0x32 0x04 0x14 LDA r20, [p2], #4; ADD.NC r18, r18, #4 + 288 0x02 0x1f 0xe6 0x98 LDA dc7, [p2], #4 + 292 0x02 0x1e 0xee 0x98 LDA el11, [p2], #4 + 296 0x00 0x00 NOPX + 298 0x00 0x00 NOPX + 300 0x00 0x00 NOPX + 302 0x09 0x1f 0x21 0x98 ST dn6, [p1], #4 + 306 0x00 0x00 NOPX + 308 0x09 0x1f 0xe1 0x98 ST dc7, [p1], #4 + 312 0x23 0xcd 0x30 0x00 0x01 0xa5 0x70 0x02 ST el9, [p1], #4; NOPM + 320 0x00 0x2c 0xf0 0x00 0x21 0x1e 0x91 0x80 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; ST r20, [p1], #4; NOPX; NOPM; NOPV + 336 0x00 0x2c 0xf0 0x00 0x21 0xbe 0xe9 0x80 0x00 0x00 0x01 0xa5 0x78 0x00 0x00 0xe1 NOPA; NOPB; ST el11, [p1], #-20; NOPX; NOPM; NOPV + 352 0x00 0x2c 0xf0 0x00 0x20 0x01 0x5b 0x00 0x00 0x02 0xa9 0x60 0x78 0x00 0x00 0xe1 NOPA; NOPB; NOPS; NOPX; MOV r21, p1; NOPV +.label ZLE_F_main_368 +.end_of_loop + 368 0x00 0x2c 0xf0 0x00 0x20 0x1e 0xb1 0x80 0x00 0x00 0xb5 0x46 0x08 0x00 0x00 0xe1 NOPA; NOPB; ST r21, [p0], #4; NOPX; ADD.NC p1, r21, #24; NOPV +.loop_nesting 2 + 384 0x00 0x0b 0x00 0x27 0x46 0x6e 0x6f 0x60 0x78 0xba MOVA r11, #0; LSHL r20, r19, r12; MOV r19, p7 + 394 0x9b 0x74 0x69 0xb2 0x01 0x24 ADD r13, r19, #-24; ADD.NC r19, r18, #1 + 400 0xf1 0xa1 0x60 0x25 0x26 0x6f 0x45 0x10 0x79 0x3a MOVS p7, r13; LSHL r18, r18, r12; MOV dj6, r20 + 410 0x9c 0x99 0xbb 0x12 0x41 0xe4 LSHL r18, r19, r12; MOV dj5, r18 + 416 0xf4 0x4a 0xde 0x8d 0x92 0x94 LDA r18, [p7, dj5]; ADD.NC dn7, r13, r18 + 422 0xd8 0x74 0x30 0x02 0x2c 0x7f 0xc0 0x02 ST dn7, [p6, dj6]; ADD.NC r17, r17, #-1 + 430 0x00 0x00 NOPX + 432 0x00 0x00 NOPX + 434 0x00 0x00 NOPX + 436 0x00 0x00 NOPX + 438 0x00 0x00 NOPX + 440 0x00 0x2b 0x60 0x02 0x4c 0xe4 0xa0 0x02 NOPS; ADD.NC r18, r19, r18 +.label TGT_F_main_448 +.loop_nesting 3 + 448 0x14 0xa1 0x01 0x98 SUB r16, r18, r16 + 452 0x14 0x1e 0xcd 0x98 LSHL r15, r16, r12 + 456 0xf1 0xa1 0x60 0x03 0xc3 0xd0 0x70 0x02 MOVS p7, r13; MOV dj7, r15 + 464 0x07 0xe0 0x16 0x98 LDA r0, [p7, dj7] +.no_stack_arguments + 468 0x00 0x00 0x00 0x00 0x01 0x04 JL #_ZN3adf11block_writeEPKNS_7reg_valEj +.delay_slot + 474 0x1b 0x98 0x00 0x98 ADD.NC r14, r16, #1 +.delay_slot +.swstall delay_slot + 478 0x00 0x00 NOPX +.delay_slot + 480 0x13 0xa0 0xcd 0x98 LSHL r16, r14, r12 +.delay_slot + 484 0x18 0x66 0xc1 0x58 ADD.NC p0, r13, r16 +.delay_slot + 488 0xfe 0x46 0xb0 0x00 0x01 0xa5 0x70 0x02 ST r17, [sp, #-16]; NOPM +.return_address + 496 0x04 0x40 0xa1 0x0f 0x41 0xe4 MOVX r17, #1; MOV dj0, r15 + 502 0xe0 0x4a 0xd0 0x00 0x00 0x1e 0x8b 0xe0 0x10 0xba LDA r18, [p7, dj0]; MOVXM r20, #30656 + 512 0x00 0x00 0x7a 0xaf 0x84 0x44 MOVXM r21, #30658 + 518 0x00 0x08 0x00 0xc0 0x00 0x44 MOVXM p0, #524288 + 524 0x00 0x70 0x08 0x20 0x06 0x44 MOVXM r16, #7340035 + 530 0x00 0x00 NOPX + 532 0x00 0x00 NOPX + 534 0x00 0x00 NOPX + 536 0x14 0xa3 0x1d 0x98 LSHL r17, r18, r17 + 540 0x1c 0x98 0xb9 0x58 ADD.NC r18, r17, r14 + 544 0x14 0xa6 0xcd 0x98 LSHL r19, r18, r12 + 548 0x18 0x89 0xa0 0xf8 MOV dj0, r19 + 552 0x07 0x02 0xd6 0x98 LDA r22, [p7, dj0] + 556 0x00 0x00 NOPX + 558 0x00 0x00 NOPX + 560 0x00 0x00 NOPX + 562 0x00 0x00 NOPX + 564 0x00 0x00 NOPX + 566 0x00 0x00 NOPX + 568 0x15 0xb6 0xcc 0x98 LTU r27, r22, r12 + 572 0x15 0x69 0x42 0x18 SEL.EQZ r20, r21, r20, r27 + 576 0x1d 0x1b 0x51 0x58 ADD.NC r20, r22, r20 + 580 0x15 0x28 0xcd 0x98 LSHL r20, r20, r12 + 584 0x00 0x2b 0x60 0x00 0x45 0x10 0x70 0x02 NOPS; MOV dj0, r20 +.label TGT_F_main_592 +.loop_nesting 4 + 592 0x00 0x02 0x93 0x98 LDA.TM r20, [p0, dj0] + 596 0x00 0x00 NOPX + 598 0x00 0x00 NOPX + 600 0x00 0x00 NOPX + 602 0x00 0x00 NOPX + 604 0x00 0x00 NOPX + 606 0x00 0x00 NOPX + 608 0x15 0x2b 0x04 0x98 AND r21, r20, r16 + 612 0xa8 0x01 0x28 0x40 0x01 0x84 JNZ r21, #TGT_F_main_592 +.delay_slot +.swstall delay_slot + 618 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 620 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 622 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 624 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 626 0x00 0x00 NOPX +.loop_nesting 3 + 628 0x1b 0xd9 0x00 0x98 ADD.NC r15, r18, #1 + 632 0x13 0xdc 0xcd 0x98 LSHL r14, r15, r12 + 636 0x19 0x87 0x20 0xf8 MOV dj1, r14 + 640 0x07 0x20 0x16 0x98 LDA r0, [p7, dj1] +.no_stack_arguments + 644 0x00 0x00 0x00 0x00 0x01 0x04 JL #_ZN3adf11block_writeEPKNS_7reg_valEj +.delay_slot +.swstall delay_slot + 650 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 652 0x00 0x00 NOPX +.delay_slot + 654 0x13 0x5a 0x23 0x18 ADD r13, r13, #8 +.delay_slot + 658 0x18 0x69 0xb5 0x58 ADD.NC p0, r19, r13 +.delay_slot + 662 0x00 0x2c 0xf7 0xea 0x35 0x80 0x00 0x00 0x00 0x7a NOPA; ST r17, [sp, #-24]; NOPX +.return_address + 672 0x00 0x0e 0x00 0x01 0x00 0x28 0x43 0x90 0x78 0xba MOVA r14, #0; MOVX r16, #1; MOV dj0, r14 + 682 0xe0 0x4a 0xd8 0xad 0xfc 0x14 LDA r18, [p7, dj0]; ADD.NC r17, r13, #-4 + 688 0x0f 0xee 0x35 0x98 ST r17, [sp, #-20] + 692 0x00 0x00 NOPX + 694 0x00 0x00 NOPX + 696 0x00 0x00 NOPX + 698 0x00 0x00 NOPX + 700 0x00 0x00 NOPX + 702 0x14 0xa5 0x0d 0x98 LSHL r18, r18, r16 + 706 0x14 0xa1 0x05 0x98 OR r16, r18, r16 + 710 0x1c 0x97 0xc1 0x58 ADD.NC r18, r15, r16 + 714 0xfc 0x42 0xb0 0x24 0xc6 0x6c 0x37 0x60 0x79 0x3a ST r16, [sp, #-32]; LSHL r12, r18, r12; MOV p0, p7 + 724 0xfc 0xca 0xb0 0x23 0x06 0x04 0x43 0x10 0x79 0x3a ST r18, [sp, #-28]; ADD r16, r17, r12; MOV dj0, r12 + 734 0x00 0x42 0xd7 0xde 0x15 0x80 0x00 0x03 0xb1 0x80 0x10 0x76 LDA r16, [p0, dj0]; ST r16, [sp, #-36]; MOVXM p7, #(TGT_F_main_768 + 0) + 746 0x00 0x00 NOPX + 748 0x00 0x00 NOPX + 750 0x00 0x00 NOPX + 752 0x00 0x00 NOPX + 754 0x00 0x00 NOPX + 756 0x00 0x00 NOPX + 758 0x00 0x2c 0xf0 0x00 0x10 0x01 0xec 0x3f 0xce 0xba NOPA; NOPB; ADD.NC r15, r16, #-1 +.label TGT_F_main_768 +.loop_nesting 4 + 768 0x12 0xe0 0xe5 0x98 OR r16, r11, r14 + 772 0x80 0x01 0xb8 0x40 0x01 0x84 JNZ r16, #TGT_F_main_880 +.delay_slot +.swstall delay_slot + 778 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 780 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 782 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 784 0x00 0x00 NOPX +.delay_slot + 786 0x11 0xa1 0x60 0x00 0xc3 0x10 0x70 0x02 MOVS p0, r13; MOV dj1, r12 + 794 0x07 0xdc 0x99 0x18 LDA p1, [sp, #-36] + 798 0x00 0x00 NOPX + 800 0x00 0x00 NOPX + 802 0x00 0x00 NOPX + 804 0x00 0x00 NOPX + 806 0x00 0x00 NOPX + 808 0x00 0x00 NOPX + 810 0x01 0x06 0x16 0x98 LDA r16, [p1] + 814 0x00 0x00 NOPX + 816 0x00 0x00 NOPX + 818 0x00 0x00 NOPX + 820 0x00 0x00 NOPX + 822 0x00 0x00 NOPX + 824 0x00 0x00 NOPX + 826 0x80 0x01 0xb8 0x00 0x01 0x84 JZ r16, #TGT_F_main_880 +.delay_slot +.swstall delay_slot + 832 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 834 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 836 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 838 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 840 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 842 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 844 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 846 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 848 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 850 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 852 0x00 0x00 NOPX + 854 0x10 0x08 0x00 0x18 DONE +.swstall chess_separator_scheduler + 858 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 860 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 862 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 864 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 866 0x00 0x00 NOPX +.swstall chess_separator_scheduler + 868 0x00 0x2c 0xf0 0x00 0x20 0x00 0x00 0x00 0x00 0xad 0x81 0x36 NOPA; NOPB; NOPS; NOPX +.label TGT_F_main_880 + 880 0x04 0x02 0xd0 0xd9 0x81 0xd4 LDA r0, [p0, dj1]; MOV p0, p6 +.no_stack_arguments + 886 0x00 0x00 0x00 0x00 0x01 0x04 JL #_Z13kernelWrapperPPvjjjj +.delay_slot +.swstall delay_slot + 892 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 894 0x00 0x00 NOPX +.delay_slot + 896 0x18 0x55 0x20 0xf8 MOV r1, r10 +.delay_slot + 900 0x18 0x94 0xa0 0xf8 MOV r2, r9 +.delay_slot + 904 0x00 0x2b 0x60 0x00 0x6a 0x10 0x70 0x02 NOPS; MOV r3, r8 +.return_address + 912 0x13 0xdf 0xe0 0x18 JNZD r15, r15, p7 +.delay_slot +.swstall delay_slot + 916 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 918 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 920 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 922 0x00 0x00 NOPX +.delay_slot + 924 0x1b 0x97 0x00 0x98 ADD.NC r14, r14, #1 +.loop_nesting 3 + 928 0xfe 0x46 0x20 0x00 0xc0 0x49 0x6a 0xc0 0x48 0xba LDA r17, [sp, #-16]; MOVX r12, #2; ADD.NC r11, r11, #1 + 938 0xfd 0xca 0x20 0x00 0x00 0x03 0xb0 0xe0 0x10 0xba LDA r18, [sp, #-20]; MOVXM p7, #(TGT_F_main_448 + 0) + 948 0x07 0xe6 0x71 0x18 LDA r19, [sp, #-28] + 952 0x07 0xe2 0x11 0x18 LDA r16, [sp, #-32] + 956 0x07 0xea 0x91 0x18 LDA r20, [sp, #-24] + 960 0x00 0x00 NOPX + 962 0x00 0x00 NOPX + 964 0x14 0x63 0xe0 0x18 JNZD r17, r17, p7 +.delay_slot + 968 0x14 0x9b 0xf3 0x18 ADD r13, r18, #-4 +.delay_slot + 972 0x1c 0x99 0x81 0x98 ADD.NC r18, r19, #3 +.delay_slot +.swstall delay_slot + 976 0x00 0x00 NOPX +.delay_slot + 978 0x1c 0x1a 0x41 0x58 ADD.NC r16, r20, r16 +.delay_slot + 982 0x1c 0x18 0x02 0x98 ADD.NC r16, r16, #5 +.loop_nesting 2 + 986 0x07 0xd2 0x07 0x8d 0x0b 0x25 0x06 0x6e 0x88 0x3f 0x58 0x76 MOVA r18, #62; MOVS p7, r13; LSHL r16, r18, r12; MOV r20, #63 + 998 0x00 0x30 0x00 0x3f 0x67 0xe8 0x44 0x10 0x78 0xba MOVA r16, #1; MOVX r22, #-1; MOV dj0, r16 + 1008 0xe0 0x46 0xd0 0x00 0x00 0x02 0x68 0x00 0x10 0xba LDA r17, [p7, dj0]; MOVXM r19, #lcpPing + 1018 0x9e 0xda 0xfc 0x20 0x01 0x64 EQ r27, r19, r13; MOV r24, #0 + 1024 0x15 0x2f 0x22 0x18 SEL.EQZ r23, r20, r18, r27 + 1028 0x00 0x00 0x0a 0xa0 0x00 0x44 MOVXM r21, #lcpPong + 1034 0x00 0x00 NOPX + 1036 0x00 0x00 NOPX + 1038 0x00 0x00 NOPX + 1040 0x14 0x74 0xd0 0x18 EQZ r26, r17 + 1044 0x15 0xd1 0x68 0x18 REL r23, r22 + 1048 0x14 0xa5 0x42 0x18 SEL.EQZ r18, r18, r20, r27 + 1052 0x00 0x00 NOPX + 1054 0x00 0x00 NOPX + 1056 0x14 0x97 0x08 0x18 ACQ.COND r18, r16, r26 + 1060 0x88 0x00 0x60 0x00 0x01 0x84 JZ r17, #TGT_F_main_192 +.delay_slot + 1066 0x14 0xe1 0x52 0x18 SEL.EQZ r16, r19, r21, r27 +.delay_slot + 1070 0x1f 0x68 0x20 0xf8 MOV p7, r16 +.delay_slot +.swstall delay_slot + 1074 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1076 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1078 0x00 0x00 NOPX +.loop_nesting 1 + 1080 0xff 0x03 0x20 0x01 0x90 0x0a 0x08 0x01 0x58 0xba LDA p0, [sp, #-8]; MOVX r25, #0; MOV r16, #1 + 1090 0x00 0x00 0x18 0x00 0x00 0x84 J #TGT_F_main_48 +.delay_slot +.swstall delay_slot + 1096 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1098 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1100 0x00 0x00 NOPX +.delay_slot +.swstall delay_slot + 1102 0x00 0x00 NOPX +.label _main__end last +.label _main___func_end0 last +.delay_slot + 1104 0x07 0xf4 0x99 0x18 LDA p1, [sp, #-12] + +.undef global data main + +.undef global data lcpPing + +.undef global data lcpPong + +.undef global text _ZN3adf11block_writeEPKNS_7reg_valEj + +.undef global text _Z13kernelWrapperPPvjjjj + + + +.direct_eval +,-,(,) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.sdr b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.sdr new file mode 100644 index 0000000000000000000000000000000000000000..d2b4c2317a9a0aeb9f20fc4389bef9abee6493bf --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/Release/0_0.sdr @@ -0,0 +1,90 @@ + +// File generated by bridge version V-2024.06#84922c0d9f#241219, Fri Mar 21 03:42:51 2025 +// Copyright 2014-2024 Synopsys, Inc. All rights reserved. +// bridge -o../Release/0_0 ../Release/0_0.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork848 -pme + +// Release: ipp V-2024.06-TGT-241219 + +// Symbols in memory 'DM_bankA': +// Symbols in memory 'DM_bankAB': +// Symbols in memory 'DM_bankAC': +// Symbols in memory 'DM_bankAD': +// Symbols in memory 'DM_bankB': +// Symbols in memory 'DM_bankBC': +// Symbols in memory 'DM_bankBD': +// Symbols in memory 'DM_bankC': +// Symbols in memory 'DM_bankCD': +// Symbols in memory 'DM_bankD': +// Symbols in memory 'DM_stack': +// Symbols in memory 'DM_test': +// Symbols in memory 'DMb': +_symbol lcpPing 0x0007b280 +_symbol lcpPong 0x0007b680 +// Symbols in memory 'DMh': +// Symbols in memory 'DMh_bankA': +// Symbols in memory 'DMh_bankAB': +// Symbols in memory 'DMh_bankAC': +// Symbols in memory 'DMh_bankAD': +// Symbols in memory 'DMh_bankB': +// Symbols in memory 'DMh_bankBC': +// Symbols in memory 'DMh_bankBD': +// Symbols in memory 'DMh_bankC': +// Symbols in memory 'DMh_bankCD': +// Symbols in memory 'DMh_bankD': +// Symbols in memory 'DMh_stack': +// Symbols in memory 'DMs': +// Symbols in memory 'DMs_bankA': +// Symbols in memory 'DMs_bankAB': +// Symbols in memory 'DMs_bankAC': +// Symbols in memory 'DMs_bankAD': +// Symbols in memory 'DMs_bankB': +// Symbols in memory 'DMs_bankBC': +// Symbols in memory 'DMs_bankBD': +// Symbols in memory 'DMs_bankC': +// Symbols in memory 'DMs_bankCD': +// Symbols in memory 'DMs_bankD': +// Symbols in memory 'DMs_stack': +// Symbols in memory 'DMv': +// Symbols in memory 'DMv_bankA': +// Symbols in memory 'DMv_bankAB': +// Symbols in memory 'DMv_bankAC': +// Symbols in memory 'DMv_bankAD': +// Symbols in memory 'DMv_bankB': +// Symbols in memory 'DMv_bankBC': +// Symbols in memory 'DMv_bankBD': +// Symbols in memory 'DMv_bankC': +// Symbols in memory 'DMv_bankCD': +// Symbols in memory 'DMv_bankD': +// Symbols in memory 'DMv_stack': +// Symbols in memory 'DMw': +// Symbols in memory 'DMw_bankA': +// Symbols in memory 'DMw_bankAB': +// Symbols in memory 'DMw_bankAC': +// Symbols in memory 'DMw_bankAD': +// Symbols in memory 'DMw_bankB': +// Symbols in memory 'DMw_bankBC': +// Symbols in memory 'DMw_bankBD': +// Symbols in memory 'DMw_bankC': +// Symbols in memory 'DMw_bankCD': +// Symbols in memory 'DMw_bankD': +// Symbols in memory 'DMw_stack': +// Symbols in memory 'DMx': +// Symbols in memory 'DMx_bankA': +// Symbols in memory 'DMx_bankAB': +// Symbols in memory 'DMx_bankAC': +// Symbols in memory 'DMx_bankAD': +// Symbols in memory 'DMx_bankB': +// Symbols in memory 'DMx_bankBC': +// Symbols in memory 'DMx_bankBD': +// Symbols in memory 'DMx_bankC': +// Symbols in memory 'DMx_bankCD': +// Symbols in memory 'DMx_bankD': +// Symbols in memory 'DMx_stack': +// Symbols in memory 'PM': +_symbol _main_init 0x00000000 +_symbol _main 0x000000e0 +_symbol _ZN3adf11block_writeEPKNS_7reg_valEj 0x00000540 +_symbol _fini 0x000005f0 +_symbol __cxa_finalize 0x000006a0 +// Symbols in memory 'PMw': +// Symbols in memory 'TM4': diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/elf_ctrl_pkt.bin b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/elf_ctrl_pkt.bin new file mode 100644 index 0000000000000000000000000000000000000000..b974eb24301d1e1cc110fea1b3fe82ac0bd7810c --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/elf_ctrl_pkt.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7bae6d17370fa6ff9dce1f15ddab525892dba32c85b49d7303d98c7938009da9 +size 6148 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_0.bin b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_0.bin new file mode 100644 index 0000000000000000000000000000000000000000..647f2d2685cd52308e5bd7e87d058689c1a974a0 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_0.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:981b0e1230eb5dbd26bf35b2717875f8d4f352bc7974fd48c2ee0d81ca343686 +size 296 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_1.bin b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_1.bin new file mode 100644 index 0000000000000000000000000000000000000000..d08cc647f5c28116fded09c7ef45a297c06aa528 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_1.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e06471c10dc9f1b825c878a76d2d029bada052ace379236c9916a0d302d6e270 +size 300 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_100.bin b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_100.bin new file mode 100644 index 0000000000000000000000000000000000000000..c28a10e4854f0a442ab6aa8d2e165576197153f6 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_100.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ae43a8dcc0cd99130f42249d4f2a9f1fc91c3c99699db65bb98ad360a326ef6 +size 376 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_11.bin b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_11.bin new file mode 100644 index 0000000000000000000000000000000000000000..31f4dbc4940baf72ec6234a911e645e141887992 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/lcp/0_0_11.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05761cae01b301200d2d2ef4aa2f3f81493591bd9ea0b012b956cd6fef12f5dd +size 384 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/xlopt.log b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/xlopt.log new file mode 100644 index 0000000000000000000000000000000000000000..a674b5c9d3752966e4bd5e92dd84abe293103957 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/xlopt.log @@ -0,0 +1,81 @@ +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. + + +--------------- FILTER ANALYSIS INFO LOG --------------- + +Reading Header IR from ir/_header.ll + +-------------------------------------------------------- + +Add module pass 363{anonymous}::GuidancePass +Add module pass 363{anonymous}::ChessOptionsPass +Add module pass 363{anonymous}::DisableInliningInMainPass +Add module pass 363cdno::xlopt::AIEMergeSubWordStoresOpt +Add module pass 363{anonymous}::XLModuleAdaptor +Add module pass 363{anonymous}::IpConstPropPass +Add module pass 363{anonymous}::XLModuleAdaptor +Add module pass 363{anonymous}::XLModuleAdaptor +Add module pass 363{anonymous}::XLModuleAdaptor +Add module pass 363{anonymous}::XLModuleAdaptor<{anonymous}::AIELoopInfoPass> +Add module pass 363cdno::xlopt::AIEAnnotatePragmaPass +Add module pass 363{anonymous}::XLModuleAdaptor<{anonymous}::AIELoopPeelPass> +Add module pass 363{anonymous}::AIEAliasAnalysisPass + + +--------------- MEMORY MANAGEMENT GUIDANCE LOG --------------- + +SIZE_HINT: Global array lcpPing is 1024 bytes. Consider making it mapper-managed LUT or memory buffer. +SIZE_HINT: Global array lcpPong is 1024 bytes. Consider making it mapper-managed LUT or memory buffer. +GLOBAL_RBW_HINT: Global variable 'lcpPong' reads its default/external initialization at 0_0/src/0_0.cc:81 in kernel 'main', and is not explicitly written before this read +GLOBAL_CONFLICT_HINT: Kernels 'block_write' and 'initialize_lock' have no execution dependence, but access the same global variable 'TM_Address_Space_Start', and one of those accesses is a write +GLOBAL_CONFLICT_HINT: Kernels 'block_write' and 'wait_dma_channel_done' have no execution dependence, but access the same global variable 'TM_Address_Space_Start', and one of those accesses is a write +GLOBAL_CONFLICT_HINT: Kernels 'initialize_lock' and 'wait_dma_channel_done' have no execution dependence, but access the same global variable 'TM_Address_Space_Start', and one of those accesses is a write + +-------------------------------------------------------------- + + + +--------------- MERGING SUBWORD STORES OPT LOG --------------- + + + +--------------- LOOP STATISTICS : main --------------- + +Total loops = 7 +Loops with prepare for pipelining pragma = 0 +Loops with unroll pragma = 0 +Loops with flatten pragma = 0 +Loops with min range pragma = 4 +Loops with max range pragma = 2 +Loops with known trip count = 0 + +------------------------------------------------------ + + + +--------------- LOOP STATISTICS : _ZN3adf11block_writeEPKNS_7reg_valEj --------------- + +Total loops = 1 +Loops with prepare for pipelining pragma = 0 +Loops with unroll pragma = 0 +Loops with flatten pragma = 0 +Loops with min range pragma = 0 +Loops with max range pragma = 0 +Loops with known trip count = 0 + +-------------------------------------------------------------------------------------- + + + +--------------- LOOP STATISTICS : _ZN3adf21wait_dma_channel_doneEj --------------- + +Total loops = 1 +Loops with prepare for pipelining pragma = 0 +Loops with unroll pragma = 0 +Loops with flatten pragma = 0 +Loops with min range pragma = 0 +Loops with max range pragma = 0 +Loops with known trip count = 0 + +---------------------------------------------------------------------------------- + diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/AddressSpace.txt b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/AddressSpace.txt new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/AliasAnalysisReport.txt b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/AliasAnalysisReport.txt new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/Makefile b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/Makefile new file mode 100644 index 0000000000000000000000000000000000000000..c50c1f408bbca68d6525c084ca16001e21df4bd3 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/Makefile @@ -0,0 +1,3104 @@ +SHELL=/bin/bash +.PHONY: all 0_0 0_0_reloadable0 0_0_reloadable1 0_0_reloadable2 0_0_reloadable3 0_0_reloadable4 0_0_reloadable5 0_0_reloadable6 0_0_reloadable7 0_0_reloadable8 0_0_reloadable9 0_0_reloadable10 0_0_reloadable11 0_0_reloadable12 0_0_reloadable13 0_0_reloadable14 0_0_reloadable15 0_0_reloadable16 0_0_reloadable17 0_0_reloadable18 0_0_reloadable19 0_0_reloadable20 0_0_reloadable21 0_0_reloadable22 0_0_reloadable23 0_0_reloadable24 0_0_reloadable25 0_1 0_1_reloadable0 0_1_reloadable1 0_1_reloadable2 0_1_reloadable3 0_1_reloadable4 0_1_reloadable5 0_1_reloadable6 0_1_reloadable7 0_1_reloadable8 0_1_reloadable9 0_1_reloadable10 0_1_reloadable11 0_1_reloadable12 0_1_reloadable13 0_1_reloadable14 0_1_reloadable15 0_1_reloadable16 0_1_reloadable17 0_1_reloadable18 0_1_reloadable19 0_1_reloadable20 0_1_reloadable21 0_1_reloadable22 0_1_reloadable23 0_1_reloadable24 0_1_reloadable25 0_2 0_2_reloadable0 0_2_reloadable1 0_2_reloadable2 0_2_reloadable3 0_2_reloadable4 0_2_reloadable5 0_2_reloadable6 0_2_reloadable7 0_2_reloadable8 0_2_reloadable9 0_2_reloadable10 0_2_reloadable11 0_2_reloadable12 0_2_reloadable13 0_2_reloadable14 0_2_reloadable15 0_2_reloadable16 0_2_reloadable17 0_2_reloadable18 0_2_reloadable19 0_2_reloadable20 0_2_reloadable21 0_2_reloadable22 0_2_reloadable23 0_2_reloadable24 0_2_reloadable25 0_3 0_3_reloadable0 0_3_reloadable1 0_3_reloadable2 0_3_reloadable3 0_3_reloadable4 0_3_reloadable5 0_3_reloadable6 0_3_reloadable7 0_3_reloadable8 0_3_reloadable9 0_3_reloadable10 0_3_reloadable11 0_3_reloadable12 0_3_reloadable13 0_3_reloadable14 0_3_reloadable15 0_3_reloadable16 0_3_reloadable17 0_3_reloadable18 0_3_reloadable19 0_3_reloadable20 0_3_reloadable21 0_3_reloadable22 0_3_reloadable23 0_3_reloadable24 0_3_reloadable25 1_0 1_0_reloadable0 1_0_reloadable1 1_0_reloadable2 1_0_reloadable3 1_0_reloadable4 1_0_reloadable5 1_0_reloadable6 1_0_reloadable7 1_0_reloadable8 1_0_reloadable9 1_0_reloadable10 1_0_reloadable11 1_0_reloadable12 1_0_reloadable13 1_0_reloadable14 1_0_reloadable15 1_0_reloadable16 1_0_reloadable17 1_0_reloadable18 1_0_reloadable19 1_0_reloadable20 1_0_reloadable21 1_0_reloadable22 1_0_reloadable23 1_0_reloadable24 1_0_reloadable25 1_1 1_1_reloadable0 1_1_reloadable1 1_1_reloadable2 1_1_reloadable3 1_1_reloadable4 1_1_reloadable5 1_1_reloadable6 1_1_reloadable7 1_1_reloadable8 1_1_reloadable9 1_1_reloadable10 1_1_reloadable11 1_1_reloadable12 1_1_reloadable13 1_1_reloadable14 1_1_reloadable15 1_1_reloadable16 1_1_reloadable17 1_1_reloadable18 1_1_reloadable19 1_1_reloadable20 1_1_reloadable21 1_1_reloadable22 1_1_reloadable23 1_1_reloadable24 1_1_reloadable25 1_2 1_2_reloadable0 1_2_reloadable1 1_2_reloadable2 1_2_reloadable3 1_2_reloadable4 1_2_reloadable5 1_2_reloadable6 1_2_reloadable7 1_2_reloadable8 1_2_reloadable9 1_2_reloadable10 1_2_reloadable11 1_2_reloadable12 1_2_reloadable13 1_2_reloadable14 1_2_reloadable15 1_2_reloadable16 1_2_reloadable17 1_2_reloadable18 1_2_reloadable19 1_2_reloadable20 1_2_reloadable21 1_2_reloadable22 1_2_reloadable23 1_2_reloadable24 1_2_reloadable25 1_3 1_3_reloadable0 1_3_reloadable1 1_3_reloadable2 1_3_reloadable3 1_3_reloadable4 1_3_reloadable5 1_3_reloadable6 1_3_reloadable7 1_3_reloadable8 1_3_reloadable9 1_3_reloadable10 1_3_reloadable11 1_3_reloadable12 1_3_reloadable13 1_3_reloadable14 1_3_reloadable15 1_3_reloadable16 1_3_reloadable17 1_3_reloadable18 1_3_reloadable19 1_3_reloadable20 1_3_reloadable21 1_3_reloadable22 1_3_reloadable23 1_3_reloadable24 1_3_reloadable25 2_0 2_0_reloadable0 2_0_reloadable1 2_0_reloadable2 2_0_reloadable3 2_0_reloadable4 2_0_reloadable5 2_0_reloadable6 2_0_reloadable7 2_0_reloadable8 2_0_reloadable9 2_0_reloadable10 2_0_reloadable11 2_0_reloadable12 2_0_reloadable13 2_0_reloadable14 2_0_reloadable15 2_0_reloadable16 2_0_reloadable17 2_0_reloadable18 2_0_reloadable19 2_0_reloadable20 2_0_reloadable21 2_0_reloadable22 2_0_reloadable23 2_0_reloadable24 2_0_reloadable25 2_1 2_1_reloadable0 2_1_reloadable1 2_1_reloadable2 2_1_reloadable3 2_1_reloadable4 2_1_reloadable5 2_1_reloadable6 2_1_reloadable7 2_1_reloadable8 2_1_reloadable9 2_1_reloadable10 2_1_reloadable11 2_1_reloadable12 2_1_reloadable13 2_1_reloadable14 2_1_reloadable15 2_1_reloadable16 2_1_reloadable17 2_1_reloadable18 2_1_reloadable19 2_1_reloadable20 2_1_reloadable21 2_1_reloadable22 2_1_reloadable23 2_1_reloadable24 2_1_reloadable25 2_2 2_2_reloadable0 2_2_reloadable1 2_2_reloadable2 2_2_reloadable3 2_2_reloadable4 2_2_reloadable5 2_2_reloadable6 2_2_reloadable7 2_2_reloadable8 2_2_reloadable9 2_2_reloadable10 2_2_reloadable11 2_2_reloadable12 2_2_reloadable13 2_2_reloadable14 2_2_reloadable15 2_2_reloadable16 2_2_reloadable17 2_2_reloadable18 2_2_reloadable19 2_2_reloadable20 2_2_reloadable21 2_2_reloadable22 2_2_reloadable23 2_2_reloadable24 2_2_reloadable25 2_3 2_3_reloadable0 2_3_reloadable1 2_3_reloadable2 2_3_reloadable3 2_3_reloadable4 2_3_reloadable5 2_3_reloadable6 2_3_reloadable7 2_3_reloadable8 2_3_reloadable9 2_3_reloadable10 2_3_reloadable11 2_3_reloadable12 2_3_reloadable13 2_3_reloadable14 2_3_reloadable15 2_3_reloadable16 2_3_reloadable17 2_3_reloadable18 2_3_reloadable19 2_3_reloadable20 2_3_reloadable21 2_3_reloadable22 2_3_reloadable23 2_3_reloadable24 2_3_reloadable25 3_0 3_0_reloadable0 3_0_reloadable1 3_0_reloadable2 3_0_reloadable3 3_0_reloadable4 3_0_reloadable5 3_0_reloadable6 3_0_reloadable7 3_0_reloadable8 3_0_reloadable9 3_0_reloadable10 3_0_reloadable11 3_0_reloadable12 3_0_reloadable13 3_0_reloadable14 3_0_reloadable15 3_0_reloadable16 3_0_reloadable17 3_0_reloadable18 3_0_reloadable19 3_0_reloadable20 3_0_reloadable21 3_0_reloadable22 3_0_reloadable23 3_0_reloadable24 3_0_reloadable25 3_1 3_1_reloadable0 3_1_reloadable1 3_1_reloadable2 3_1_reloadable3 3_1_reloadable4 3_1_reloadable5 3_1_reloadable6 3_1_reloadable7 3_1_reloadable8 3_1_reloadable9 3_1_reloadable10 3_1_reloadable11 3_1_reloadable12 3_1_reloadable13 3_1_reloadable14 3_1_reloadable15 3_1_reloadable16 3_1_reloadable17 3_1_reloadable18 3_1_reloadable19 3_1_reloadable20 3_1_reloadable21 3_1_reloadable22 3_1_reloadable23 3_1_reloadable24 3_1_reloadable25 3_2 3_2_reloadable0 3_2_reloadable1 3_2_reloadable2 3_2_reloadable3 3_2_reloadable4 3_2_reloadable5 3_2_reloadable6 3_2_reloadable7 3_2_reloadable8 3_2_reloadable9 3_2_reloadable10 3_2_reloadable11 3_2_reloadable12 3_2_reloadable13 3_2_reloadable14 3_2_reloadable15 3_2_reloadable16 3_2_reloadable17 3_2_reloadable18 3_2_reloadable19 3_2_reloadable20 3_2_reloadable21 3_2_reloadable22 3_2_reloadable23 3_2_reloadable24 3_2_reloadable25 3_3 3_3_reloadable0 3_3_reloadable1 3_3_reloadable2 3_3_reloadable3 3_3_reloadable4 3_3_reloadable5 3_3_reloadable6 3_3_reloadable7 3_3_reloadable8 3_3_reloadable9 3_3_reloadable10 3_3_reloadable11 3_3_reloadable12 3_3_reloadable13 3_3_reloadable14 3_3_reloadable15 3_3_reloadable16 3_3_reloadable17 3_3_reloadable18 3_3_reloadable19 3_3_reloadable20 3_3_reloadable21 3_3_reloadable22 3_3_reloadable23 3_3_reloadable24 3_3_reloadable25 clean + +ifeq ($(XILINX_VITIS_AIETOOLS),) +XILINX_VITIS_AIETOOLS := ${XILINX_VITIS}/aietools +endif + +ifeq ($(CARDANO_AIE_ARCH_MODEL_DIR),) +CARDANO_AIE_ARCH_MODEL_DIR := ${XILINX_VITIS_AIETOOLS}/data/aie2p/lib +endif + +ts := $(shell date "+%Y-%m-%d-%H-%M-%S") + +XCHESSMK := xchessmk -aiearch aie2p +XCHESSCC := xchesscc -aiearch aie2p +CHESSLLVMLink := ${XILINX_VITIS_AIETOOLS}/tps/${RDI_PLATFORM}/target_aie2p/bin/LNa64bin/chess-llvm-link +INCLUDE_PATH := -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I ${XILINX_VITIS_AIETOOLS}/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I ${CARDANO_AIE_ARCH_MODEL_DIR}/runtime_cxx/libcxx-lite/include -I ${CARDANO_AIE_ARCH_MODEL_DIR}/runtime_cxx/libs/libcxx-9.0.0/include-lite -I ${CARDANO_AIE_ARCH_MODEL_DIR}/runtime/include + +.EXPORT_ALL_VARIABLES: +export XILINX_CARDANO_XLOPT_OPTIONS=-xlopt=1 -mapped-json=top_mapped.json -max-required-vector-alignment=64 + +all: 0_0 0_0_reloadable0 0_0_reloadable1 0_0_reloadable2 0_0_reloadable3 0_0_reloadable4 0_0_reloadable5 0_0_reloadable6 0_0_reloadable7 0_0_reloadable8 0_0_reloadable9 0_0_reloadable10 0_0_reloadable11 0_0_reloadable12 0_0_reloadable13 0_0_reloadable14 0_0_reloadable15 0_0_reloadable16 0_0_reloadable17 0_0_reloadable18 0_0_reloadable19 0_0_reloadable20 0_0_reloadable21 0_0_reloadable22 0_0_reloadable23 0_0_reloadable24 0_0_reloadable25 0_1 0_1_reloadable0 0_1_reloadable1 0_1_reloadable2 0_1_reloadable3 0_1_reloadable4 0_1_reloadable5 0_1_reloadable6 0_1_reloadable7 0_1_reloadable8 0_1_reloadable9 0_1_reloadable10 0_1_reloadable11 0_1_reloadable12 0_1_reloadable13 0_1_reloadable14 0_1_reloadable15 0_1_reloadable16 0_1_reloadable17 0_1_reloadable18 0_1_reloadable19 0_1_reloadable20 0_1_reloadable21 0_1_reloadable22 0_1_reloadable23 0_1_reloadable24 0_1_reloadable25 0_2 0_2_reloadable0 0_2_reloadable1 0_2_reloadable2 0_2_reloadable3 0_2_reloadable4 0_2_reloadable5 0_2_reloadable6 0_2_reloadable7 0_2_reloadable8 0_2_reloadable9 0_2_reloadable10 0_2_reloadable11 0_2_reloadable12 0_2_reloadable13 0_2_reloadable14 0_2_reloadable15 0_2_reloadable16 0_2_reloadable17 0_2_reloadable18 0_2_reloadable19 0_2_reloadable20 0_2_reloadable21 0_2_reloadable22 0_2_reloadable23 0_2_reloadable24 0_2_reloadable25 0_3 0_3_reloadable0 0_3_reloadable1 0_3_reloadable2 0_3_reloadable3 0_3_reloadable4 0_3_reloadable5 0_3_reloadable6 0_3_reloadable7 0_3_reloadable8 0_3_reloadable9 0_3_reloadable10 0_3_reloadable11 0_3_reloadable12 0_3_reloadable13 0_3_reloadable14 0_3_reloadable15 0_3_reloadable16 0_3_reloadable17 0_3_reloadable18 0_3_reloadable19 0_3_reloadable20 0_3_reloadable21 0_3_reloadable22 0_3_reloadable23 0_3_reloadable24 0_3_reloadable25 1_0 1_0_reloadable0 1_0_reloadable1 1_0_reloadable2 1_0_reloadable3 1_0_reloadable4 1_0_reloadable5 1_0_reloadable6 1_0_reloadable7 1_0_reloadable8 1_0_reloadable9 1_0_reloadable10 1_0_reloadable11 1_0_reloadable12 1_0_reloadable13 1_0_reloadable14 1_0_reloadable15 1_0_reloadable16 1_0_reloadable17 1_0_reloadable18 1_0_reloadable19 1_0_reloadable20 1_0_reloadable21 1_0_reloadable22 1_0_reloadable23 1_0_reloadable24 1_0_reloadable25 1_1 1_1_reloadable0 1_1_reloadable1 1_1_reloadable2 1_1_reloadable3 1_1_reloadable4 1_1_reloadable5 1_1_reloadable6 1_1_reloadable7 1_1_reloadable8 1_1_reloadable9 1_1_reloadable10 1_1_reloadable11 1_1_reloadable12 1_1_reloadable13 1_1_reloadable14 1_1_reloadable15 1_1_reloadable16 1_1_reloadable17 1_1_reloadable18 1_1_reloadable19 1_1_reloadable20 1_1_reloadable21 1_1_reloadable22 1_1_reloadable23 1_1_reloadable24 1_1_reloadable25 1_2 1_2_reloadable0 1_2_reloadable1 1_2_reloadable2 1_2_reloadable3 1_2_reloadable4 1_2_reloadable5 1_2_reloadable6 1_2_reloadable7 1_2_reloadable8 1_2_reloadable9 1_2_reloadable10 1_2_reloadable11 1_2_reloadable12 1_2_reloadable13 1_2_reloadable14 1_2_reloadable15 1_2_reloadable16 1_2_reloadable17 1_2_reloadable18 1_2_reloadable19 1_2_reloadable20 1_2_reloadable21 1_2_reloadable22 1_2_reloadable23 1_2_reloadable24 1_2_reloadable25 1_3 1_3_reloadable0 1_3_reloadable1 1_3_reloadable2 1_3_reloadable3 1_3_reloadable4 1_3_reloadable5 1_3_reloadable6 1_3_reloadable7 1_3_reloadable8 1_3_reloadable9 1_3_reloadable10 1_3_reloadable11 1_3_reloadable12 1_3_reloadable13 1_3_reloadable14 1_3_reloadable15 1_3_reloadable16 1_3_reloadable17 1_3_reloadable18 1_3_reloadable19 1_3_reloadable20 1_3_reloadable21 1_3_reloadable22 1_3_reloadable23 1_3_reloadable24 1_3_reloadable25 2_0 2_0_reloadable0 2_0_reloadable1 2_0_reloadable2 2_0_reloadable3 2_0_reloadable4 2_0_reloadable5 2_0_reloadable6 2_0_reloadable7 2_0_reloadable8 2_0_reloadable9 2_0_reloadable10 2_0_reloadable11 2_0_reloadable12 2_0_reloadable13 2_0_reloadable14 2_0_reloadable15 2_0_reloadable16 2_0_reloadable17 2_0_reloadable18 2_0_reloadable19 2_0_reloadable20 2_0_reloadable21 2_0_reloadable22 2_0_reloadable23 2_0_reloadable24 2_0_reloadable25 2_1 2_1_reloadable0 2_1_reloadable1 2_1_reloadable2 2_1_reloadable3 2_1_reloadable4 2_1_reloadable5 2_1_reloadable6 2_1_reloadable7 2_1_reloadable8 2_1_reloadable9 2_1_reloadable10 2_1_reloadable11 2_1_reloadable12 2_1_reloadable13 2_1_reloadable14 2_1_reloadable15 2_1_reloadable16 2_1_reloadable17 2_1_reloadable18 2_1_reloadable19 2_1_reloadable20 2_1_reloadable21 2_1_reloadable22 2_1_reloadable23 2_1_reloadable24 2_1_reloadable25 2_2 2_2_reloadable0 2_2_reloadable1 2_2_reloadable2 2_2_reloadable3 2_2_reloadable4 2_2_reloadable5 2_2_reloadable6 2_2_reloadable7 2_2_reloadable8 2_2_reloadable9 2_2_reloadable10 2_2_reloadable11 2_2_reloadable12 2_2_reloadable13 2_2_reloadable14 2_2_reloadable15 2_2_reloadable16 2_2_reloadable17 2_2_reloadable18 2_2_reloadable19 2_2_reloadable20 2_2_reloadable21 2_2_reloadable22 2_2_reloadable23 2_2_reloadable24 2_2_reloadable25 2_3 2_3_reloadable0 2_3_reloadable1 2_3_reloadable2 2_3_reloadable3 2_3_reloadable4 2_3_reloadable5 2_3_reloadable6 2_3_reloadable7 2_3_reloadable8 2_3_reloadable9 2_3_reloadable10 2_3_reloadable11 2_3_reloadable12 2_3_reloadable13 2_3_reloadable14 2_3_reloadable15 2_3_reloadable16 2_3_reloadable17 2_3_reloadable18 2_3_reloadable19 2_3_reloadable20 2_3_reloadable21 2_3_reloadable22 2_3_reloadable23 2_3_reloadable24 2_3_reloadable25 3_0 3_0_reloadable0 3_0_reloadable1 3_0_reloadable2 3_0_reloadable3 3_0_reloadable4 3_0_reloadable5 3_0_reloadable6 3_0_reloadable7 3_0_reloadable8 3_0_reloadable9 3_0_reloadable10 3_0_reloadable11 3_0_reloadable12 3_0_reloadable13 3_0_reloadable14 3_0_reloadable15 3_0_reloadable16 3_0_reloadable17 3_0_reloadable18 3_0_reloadable19 3_0_reloadable20 3_0_reloadable21 3_0_reloadable22 3_0_reloadable23 3_0_reloadable24 3_0_reloadable25 3_1 3_1_reloadable0 3_1_reloadable1 3_1_reloadable2 3_1_reloadable3 3_1_reloadable4 3_1_reloadable5 3_1_reloadable6 3_1_reloadable7 3_1_reloadable8 3_1_reloadable9 3_1_reloadable10 3_1_reloadable11 3_1_reloadable12 3_1_reloadable13 3_1_reloadable14 3_1_reloadable15 3_1_reloadable16 3_1_reloadable17 3_1_reloadable18 3_1_reloadable19 3_1_reloadable20 3_1_reloadable21 3_1_reloadable22 3_1_reloadable23 3_1_reloadable24 3_1_reloadable25 3_2 3_2_reloadable0 3_2_reloadable1 3_2_reloadable2 3_2_reloadable3 3_2_reloadable4 3_2_reloadable5 3_2_reloadable6 3_2_reloadable7 3_2_reloadable8 3_2_reloadable9 3_2_reloadable10 3_2_reloadable11 3_2_reloadable12 3_2_reloadable13 3_2_reloadable14 3_2_reloadable15 3_2_reloadable16 3_2_reloadable17 3_2_reloadable18 3_2_reloadable19 3_2_reloadable20 3_2_reloadable21 3_2_reloadable22 3_2_reloadable23 3_2_reloadable24 3_2_reloadable25 3_3 3_3_reloadable0 3_3_reloadable1 3_3_reloadable2 3_3_reloadable3 3_3_reloadable4 3_3_reloadable5 3_3_reloadable6 3_3_reloadable7 3_3_reloadable8 3_3_reloadable9 3_3_reloadable10 3_3_reloadable11 3_3_reloadable12 3_3_reloadable13 3_3_reloadable14 3_3_reloadable15 3_3_reloadable16 3_3_reloadable17 3_3_reloadable18 3_3_reloadable19 3_3_reloadable20 3_3_reloadable21 3_3_reloadable22 3_3_reloadable23 3_3_reloadable24 3_3_reloadable25 + +0_0_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/src/0_0.cc -o ir/0_0_orig.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_orig.ll -o ir/0_0.ll 2> 0_0/xlopt.log +0_0_llopt: 0_0_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0.ll -o ir/0_0.ll 2>/dev/null + +0_0_reloadable0_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable0/src/0_0_reloadable0.cc -o ir/0_0_reloadable0_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable0_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable0_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable0_orig.ll -o ir/0_0_reloadable0.ll 2> 0_0_reloadable0/xlopt.log +0_0_reloadable0_llopt: 0_0_reloadable0_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable0.ll -o ir/0_0_reloadable0.ll 2>/dev/null + +0_0_reloadable1_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable1/src/0_0_reloadable1.cc -o ir/0_0_reloadable1_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable1_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable1_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable1_orig.ll -o ir/0_0_reloadable1.ll 2> 0_0_reloadable1/xlopt.log +0_0_reloadable1_llopt: 0_0_reloadable1_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable1.ll -o ir/0_0_reloadable1.ll 2>/dev/null + +0_0_reloadable2_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable2/src/0_0_reloadable2.cc -o ir/0_0_reloadable2_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable2_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable2_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable2_orig.ll -o ir/0_0_reloadable2.ll 2> 0_0_reloadable2/xlopt.log +0_0_reloadable2_llopt: 0_0_reloadable2_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable2.ll -o ir/0_0_reloadable2.ll 2>/dev/null + +0_0_reloadable3_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable3/src/0_0_reloadable3.cc -o ir/0_0_reloadable3_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable3_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable3_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable3_orig.ll -o ir/0_0_reloadable3.ll 2> 0_0_reloadable3/xlopt.log +0_0_reloadable3_llopt: 0_0_reloadable3_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable3.ll -o ir/0_0_reloadable3.ll 2>/dev/null + +0_0_reloadable5_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable5/src/0_0_reloadable5.cc -o ir/0_0_reloadable5_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable5_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable5_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable5_orig.ll -o ir/0_0_reloadable5.ll 2> 0_0_reloadable5/xlopt.log +0_0_reloadable5_llopt: 0_0_reloadable5_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable5.ll -o ir/0_0_reloadable5.ll 2>/dev/null + +0_0_reloadable17_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable17/src/0_0_reloadable17.cc -o ir/0_0_reloadable17_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable17_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable17_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable17_orig.ll -o ir/0_0_reloadable17.ll 2> 0_0_reloadable17/xlopt.log +0_0_reloadable17_llopt: 0_0_reloadable17_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable17.ll -o ir/0_0_reloadable17.ll 2>/dev/null + +0_0_reloadable19_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable19/src/0_0_reloadable19.cc -o ir/0_0_reloadable19_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable19_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable19_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable19_orig.ll -o ir/0_0_reloadable19.ll 2> 0_0_reloadable19/xlopt.log +0_0_reloadable19_llopt: 0_0_reloadable19_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable19.ll -o ir/0_0_reloadable19.ll 2>/dev/null + +0_0_reloadable20_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable20/src/0_0_reloadable20.cc -o ir/0_0_reloadable20_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable20_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable20_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable20_orig.ll -o ir/0_0_reloadable20.ll 2> 0_0_reloadable20/xlopt.log +0_0_reloadable20_llopt: 0_0_reloadable20_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable20.ll -o ir/0_0_reloadable20.ll 2>/dev/null + +0_0_reloadable21_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable21/src/0_0_reloadable21.cc -o ir/0_0_reloadable21_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable21_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable21_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable21_orig.ll -o ir/0_0_reloadable21.ll 2> 0_0_reloadable21/xlopt.log +0_0_reloadable21_llopt: 0_0_reloadable21_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable21.ll -o ir/0_0_reloadable21.ll 2>/dev/null + +0_0_reloadable22_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc -o ir/0_0_reloadable22_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable22_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable22_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable22_orig.ll -o ir/0_0_reloadable22.ll 2> 0_0_reloadable22/xlopt.log +0_0_reloadable22_llopt: 0_0_reloadable22_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable22.ll -o ir/0_0_reloadable22.ll 2>/dev/null + +0_0_reloadable23_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc -o ir/0_0_reloadable23_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable23_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable23_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable23_orig.ll -o ir/0_0_reloadable23.ll 2> 0_0_reloadable23/xlopt.log +0_0_reloadable23_llopt: 0_0_reloadable23_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable23.ll -o ir/0_0_reloadable23.ll 2>/dev/null + +0_0_reloadable24_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable24/src/0_0_reloadable24.cc -o ir/0_0_reloadable24_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable24_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable24_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable24_orig.ll -o ir/0_0_reloadable24.ll 2> 0_0_reloadable24/xlopt.log +0_0_reloadable24_llopt: 0_0_reloadable24_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable24.ll -o ir/0_0_reloadable24.ll 2>/dev/null + +0_0_reloadable25_xlopt: + ${XCHESSCC} +f +s -p me -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR ${INCLUDE_PATH} /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable25/src/0_0_reloadable25.cc -o ir/0_0_reloadable25_main.ll + ${CHESSLLVMLink} --chess-config ${CARDANO_AIE_ARCH_MODEL_DIR}/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable25_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable25_main.ll + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable25_orig.ll -o ir/0_0_reloadable25.ll 2> 0_0_reloadable25/xlopt.log +0_0_reloadable25_llopt: 0_0_reloadable25_xlopt + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable25.ll -o ir/0_0_reloadable25.ll 2>/dev/null + +0_0: 0_0_xlopt 0_0_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0/scripts/0_0.prx) 2>&1 |& tee -a 0_0/0_0.log 0_0/timestamped_log/0_0-${ts}.log + (readelf --debug-dump=decodedline 0_0/Release/0_0 >> 0_0/Release/0_0.txt) +0_0_reloadable0: 0_0_reloadable0_xlopt 0_0_reloadable0_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable0/scripts/0_0_reloadable0.prx) 2>&1 |& tee -a 0_0_reloadable0/0_0_reloadable0.log 0_0_reloadable0/timestamped_log/0_0_reloadable0-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable0/Release/0_0_reloadable0 >> 0_0_reloadable0/Release/0_0_reloadable0.txt) +0_0_reloadable1: 0_0_reloadable1_xlopt 0_0_reloadable1_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable1/scripts/0_0_reloadable1.prx) 2>&1 |& tee -a 0_0_reloadable1/0_0_reloadable1.log 0_0_reloadable1/timestamped_log/0_0_reloadable1-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable1/Release/0_0_reloadable1 >> 0_0_reloadable1/Release/0_0_reloadable1.txt) +0_0_reloadable2: 0_0_reloadable2_xlopt 0_0_reloadable2_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable2/scripts/0_0_reloadable2.prx) 2>&1 |& tee -a 0_0_reloadable2/0_0_reloadable2.log 0_0_reloadable2/timestamped_log/0_0_reloadable2-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable2/Release/0_0_reloadable2 >> 0_0_reloadable2/Release/0_0_reloadable2.txt) +0_0_reloadable3: 0_0_reloadable3_xlopt 0_0_reloadable3_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable3/scripts/0_0_reloadable3.prx) 2>&1 |& tee -a 0_0_reloadable3/0_0_reloadable3.log 0_0_reloadable3/timestamped_log/0_0_reloadable3-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable3/Release/0_0_reloadable3 >> 0_0_reloadable3/Release/0_0_reloadable3.txt) +0_0_reloadable5: 0_0_reloadable5_xlopt 0_0_reloadable5_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable5/scripts/0_0_reloadable5.prx) 2>&1 |& tee -a 0_0_reloadable5/0_0_reloadable5.log 0_0_reloadable5/timestamped_log/0_0_reloadable5-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable5/Release/0_0_reloadable5 >> 0_0_reloadable5/Release/0_0_reloadable5.txt) +0_0_reloadable17: 0_0_reloadable17_xlopt 0_0_reloadable17_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable17/scripts/0_0_reloadable17.prx) 2>&1 |& tee -a 0_0_reloadable17/0_0_reloadable17.log 0_0_reloadable17/timestamped_log/0_0_reloadable17-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable17/Release/0_0_reloadable17 >> 0_0_reloadable17/Release/0_0_reloadable17.txt) +0_0_reloadable19: 0_0_reloadable19_xlopt 0_0_reloadable19_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable19/scripts/0_0_reloadable19.prx) 2>&1 |& tee -a 0_0_reloadable19/0_0_reloadable19.log 0_0_reloadable19/timestamped_log/0_0_reloadable19-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable19/Release/0_0_reloadable19 >> 0_0_reloadable19/Release/0_0_reloadable19.txt) +0_0_reloadable20: 0_0_reloadable20_xlopt 0_0_reloadable20_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable20/scripts/0_0_reloadable20.prx) 2>&1 |& tee -a 0_0_reloadable20/0_0_reloadable20.log 0_0_reloadable20/timestamped_log/0_0_reloadable20-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable20/Release/0_0_reloadable20 >> 0_0_reloadable20/Release/0_0_reloadable20.txt) +0_0_reloadable21: 0_0_reloadable21_xlopt 0_0_reloadable21_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable21/scripts/0_0_reloadable21.prx) 2>&1 |& tee -a 0_0_reloadable21/0_0_reloadable21.log 0_0_reloadable21/timestamped_log/0_0_reloadable21-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable21/Release/0_0_reloadable21 >> 0_0_reloadable21/Release/0_0_reloadable21.txt) +0_0_reloadable22: 0_0_reloadable22_xlopt 0_0_reloadable22_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable22/scripts/0_0_reloadable22.prx) 2>&1 |& tee -a 0_0_reloadable22/0_0_reloadable22.log 0_0_reloadable22/timestamped_log/0_0_reloadable22-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable22/Release/0_0_reloadable22 >> 0_0_reloadable22/Release/0_0_reloadable22.txt) +0_0_reloadable23: 0_0_reloadable23_xlopt 0_0_reloadable23_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable23/scripts/0_0_reloadable23.prx) 2>&1 |& tee -a 0_0_reloadable23/0_0_reloadable23.log 0_0_reloadable23/timestamped_log/0_0_reloadable23-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable23/Release/0_0_reloadable23 >> 0_0_reloadable23/Release/0_0_reloadable23.txt) +0_0_reloadable24: 0_0_reloadable24_xlopt 0_0_reloadable24_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable24/scripts/0_0_reloadable24.prx) 2>&1 |& tee -a 0_0_reloadable24/0_0_reloadable24.log 0_0_reloadable24/timestamped_log/0_0_reloadable24-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable24/Release/0_0_reloadable24 >> 0_0_reloadable24/Release/0_0_reloadable24.txt) +0_0_reloadable25: 0_0_reloadable25_xlopt 0_0_reloadable25_llopt + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable25/scripts/0_0_reloadable25.prx) 2>&1 |& tee -a 0_0_reloadable25/0_0_reloadable25.log 0_0_reloadable25/timestamped_log/0_0_reloadable25-${ts}.log + (readelf --debug-dump=decodedline 0_0_reloadable25/Release/0_0_reloadable25 >> 0_0_reloadable25/Release/0_0_reloadable25.txt) + +0_1: 0_0 + mkdir -p 0_1/Release + cp 0_0/Release/0_0.o 0_1/Release/0_1.o + set -o pipefail; (rm -f "0_1/Release/0_1.##") 2>&1 |& tee -a 0_1/0_1.log 0_1/timestamped_log/0_1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1/scripts/0_1.prx) 2>&1 |& tee -a 0_1/0_1.log 0_1/timestamped_log/0_1-${ts}.log + +0_2: 0_0 + mkdir -p 0_2/Release + cp 0_0/Release/0_0.o 0_2/Release/0_2.o + set -o pipefail; (rm -f "0_2/Release/0_2.##") 2>&1 |& tee -a 0_2/0_2.log 0_2/timestamped_log/0_2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2/scripts/0_2.prx) 2>&1 |& tee -a 0_2/0_2.log 0_2/timestamped_log/0_2-${ts}.log + +0_3: 0_0 + mkdir -p 0_3/Release + cp 0_0/Release/0_0.o 0_3/Release/0_3.o + set -o pipefail; (rm -f "0_3/Release/0_3.##") 2>&1 |& tee -a 0_3/0_3.log 0_3/timestamped_log/0_3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3/scripts/0_3.prx) 2>&1 |& tee -a 0_3/0_3.log 0_3/timestamped_log/0_3-${ts}.log + +1_0: 0_0 + mkdir -p 1_0/Release + cp 0_0/Release/0_0.o 1_0/Release/1_0.o + set -o pipefail; (rm -f "1_0/Release/1_0.##") 2>&1 |& tee -a 1_0/1_0.log 1_0/timestamped_log/1_0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0/scripts/1_0.prx) 2>&1 |& tee -a 1_0/1_0.log 1_0/timestamped_log/1_0-${ts}.log + +1_1: 0_0 + mkdir -p 1_1/Release + cp 0_0/Release/0_0.o 1_1/Release/1_1.o + set -o pipefail; (rm -f "1_1/Release/1_1.##") 2>&1 |& tee -a 1_1/1_1.log 1_1/timestamped_log/1_1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1/scripts/1_1.prx) 2>&1 |& tee -a 1_1/1_1.log 1_1/timestamped_log/1_1-${ts}.log + +1_2: 0_0 + mkdir -p 1_2/Release + cp 0_0/Release/0_0.o 1_2/Release/1_2.o + set -o pipefail; (rm -f "1_2/Release/1_2.##") 2>&1 |& tee -a 1_2/1_2.log 1_2/timestamped_log/1_2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2/scripts/1_2.prx) 2>&1 |& tee -a 1_2/1_2.log 1_2/timestamped_log/1_2-${ts}.log + +1_3: 0_0 + mkdir -p 1_3/Release + cp 0_0/Release/0_0.o 1_3/Release/1_3.o + set -o pipefail; (rm -f "1_3/Release/1_3.##") 2>&1 |& tee -a 1_3/1_3.log 1_3/timestamped_log/1_3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3/scripts/1_3.prx) 2>&1 |& tee -a 1_3/1_3.log 1_3/timestamped_log/1_3-${ts}.log + +2_0: 0_0 + mkdir -p 2_0/Release + cp 0_0/Release/0_0.o 2_0/Release/2_0.o + set -o pipefail; (rm -f "2_0/Release/2_0.##") 2>&1 |& tee -a 2_0/2_0.log 2_0/timestamped_log/2_0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0/scripts/2_0.prx) 2>&1 |& tee -a 2_0/2_0.log 2_0/timestamped_log/2_0-${ts}.log + +2_1: 0_0 + mkdir -p 2_1/Release + cp 0_0/Release/0_0.o 2_1/Release/2_1.o + set -o pipefail; (rm -f "2_1/Release/2_1.##") 2>&1 |& tee -a 2_1/2_1.log 2_1/timestamped_log/2_1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1/scripts/2_1.prx) 2>&1 |& tee -a 2_1/2_1.log 2_1/timestamped_log/2_1-${ts}.log + +2_2: 0_0 + mkdir -p 2_2/Release + cp 0_0/Release/0_0.o 2_2/Release/2_2.o + set -o pipefail; (rm -f "2_2/Release/2_2.##") 2>&1 |& tee -a 2_2/2_2.log 2_2/timestamped_log/2_2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2/scripts/2_2.prx) 2>&1 |& tee -a 2_2/2_2.log 2_2/timestamped_log/2_2-${ts}.log + +2_3: 0_0 + mkdir -p 2_3/Release + cp 0_0/Release/0_0.o 2_3/Release/2_3.o + set -o pipefail; (rm -f "2_3/Release/2_3.##") 2>&1 |& tee -a 2_3/2_3.log 2_3/timestamped_log/2_3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3/scripts/2_3.prx) 2>&1 |& tee -a 2_3/2_3.log 2_3/timestamped_log/2_3-${ts}.log + +3_0: 0_0 + mkdir -p 3_0/Release + cp 0_0/Release/0_0.o 3_0/Release/3_0.o + set -o pipefail; (rm -f "3_0/Release/3_0.##") 2>&1 |& tee -a 3_0/3_0.log 3_0/timestamped_log/3_0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0/scripts/3_0.prx) 2>&1 |& tee -a 3_0/3_0.log 3_0/timestamped_log/3_0-${ts}.log + +3_1: 0_0 + mkdir -p 3_1/Release + cp 0_0/Release/0_0.o 3_1/Release/3_1.o + set -o pipefail; (rm -f "3_1/Release/3_1.##") 2>&1 |& tee -a 3_1/3_1.log 3_1/timestamped_log/3_1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1/scripts/3_1.prx) 2>&1 |& tee -a 3_1/3_1.log 3_1/timestamped_log/3_1-${ts}.log + +3_2: 0_0 + mkdir -p 3_2/Release + cp 0_0/Release/0_0.o 3_2/Release/3_2.o + set -o pipefail; (rm -f "3_2/Release/3_2.##") 2>&1 |& tee -a 3_2/3_2.log 3_2/timestamped_log/3_2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2/scripts/3_2.prx) 2>&1 |& tee -a 3_2/3_2.log 3_2/timestamped_log/3_2-${ts}.log + +3_3: 0_0 + mkdir -p 3_3/Release + cp 0_0/Release/0_0.o 3_3/Release/3_3.o + set -o pipefail; (rm -f "3_3/Release/3_3.##") 2>&1 |& tee -a 3_3/3_3.log 3_3/timestamped_log/3_3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3/scripts/3_3.prx) 2>&1 |& tee -a 3_3/3_3.log 3_3/timestamped_log/3_3-${ts}.log + +0_1_reloadable0: 0_0_reloadable0 + mkdir -p 0_1_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 0_1_reloadable0/Release/0_1_reloadable0.o + set -o pipefail; (rm -f "0_1_reloadable0/Release/0_1_reloadable0.##") 2>&1 |& tee -a 0_1_reloadable0/0_1_reloadable0.log 0_1_reloadable0/timestamped_log/0_1_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable0/scripts/0_1_reloadable0.prx) 2>&1 |& tee -a 0_1_reloadable0/0_1_reloadable0.log 0_1_reloadable0/timestamped_log/0_1_reloadable0-${ts}.log + +0_2_reloadable0: 0_0_reloadable0 + mkdir -p 0_2_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 0_2_reloadable0/Release/0_2_reloadable0.o + set -o pipefail; (rm -f "0_2_reloadable0/Release/0_2_reloadable0.##") 2>&1 |& tee -a 0_2_reloadable0/0_2_reloadable0.log 0_2_reloadable0/timestamped_log/0_2_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable0/scripts/0_2_reloadable0.prx) 2>&1 |& tee -a 0_2_reloadable0/0_2_reloadable0.log 0_2_reloadable0/timestamped_log/0_2_reloadable0-${ts}.log + +0_3_reloadable0: 0_0_reloadable0 + mkdir -p 0_3_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 0_3_reloadable0/Release/0_3_reloadable0.o + set -o pipefail; (rm -f "0_3_reloadable0/Release/0_3_reloadable0.##") 2>&1 |& tee -a 0_3_reloadable0/0_3_reloadable0.log 0_3_reloadable0/timestamped_log/0_3_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable0/scripts/0_3_reloadable0.prx) 2>&1 |& tee -a 0_3_reloadable0/0_3_reloadable0.log 0_3_reloadable0/timestamped_log/0_3_reloadable0-${ts}.log + +1_0_reloadable0: 0_0_reloadable0 + mkdir -p 1_0_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 1_0_reloadable0/Release/1_0_reloadable0.o + set -o pipefail; (rm -f "1_0_reloadable0/Release/1_0_reloadable0.##") 2>&1 |& tee -a 1_0_reloadable0/1_0_reloadable0.log 1_0_reloadable0/timestamped_log/1_0_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable0/scripts/1_0_reloadable0.prx) 2>&1 |& tee -a 1_0_reloadable0/1_0_reloadable0.log 1_0_reloadable0/timestamped_log/1_0_reloadable0-${ts}.log + +1_1_reloadable0: 0_0_reloadable0 + mkdir -p 1_1_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 1_1_reloadable0/Release/1_1_reloadable0.o + set -o pipefail; (rm -f "1_1_reloadable0/Release/1_1_reloadable0.##") 2>&1 |& tee -a 1_1_reloadable0/1_1_reloadable0.log 1_1_reloadable0/timestamped_log/1_1_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable0/scripts/1_1_reloadable0.prx) 2>&1 |& tee -a 1_1_reloadable0/1_1_reloadable0.log 1_1_reloadable0/timestamped_log/1_1_reloadable0-${ts}.log + +1_2_reloadable0: 0_0_reloadable0 + mkdir -p 1_2_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 1_2_reloadable0/Release/1_2_reloadable0.o + set -o pipefail; (rm -f "1_2_reloadable0/Release/1_2_reloadable0.##") 2>&1 |& tee -a 1_2_reloadable0/1_2_reloadable0.log 1_2_reloadable0/timestamped_log/1_2_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable0/scripts/1_2_reloadable0.prx) 2>&1 |& tee -a 1_2_reloadable0/1_2_reloadable0.log 1_2_reloadable0/timestamped_log/1_2_reloadable0-${ts}.log + +1_3_reloadable0: 0_0_reloadable0 + mkdir -p 1_3_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 1_3_reloadable0/Release/1_3_reloadable0.o + set -o pipefail; (rm -f "1_3_reloadable0/Release/1_3_reloadable0.##") 2>&1 |& tee -a 1_3_reloadable0/1_3_reloadable0.log 1_3_reloadable0/timestamped_log/1_3_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable0/scripts/1_3_reloadable0.prx) 2>&1 |& tee -a 1_3_reloadable0/1_3_reloadable0.log 1_3_reloadable0/timestamped_log/1_3_reloadable0-${ts}.log + +2_0_reloadable0: 0_0_reloadable0 + mkdir -p 2_0_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 2_0_reloadable0/Release/2_0_reloadable0.o + set -o pipefail; (rm -f "2_0_reloadable0/Release/2_0_reloadable0.##") 2>&1 |& tee -a 2_0_reloadable0/2_0_reloadable0.log 2_0_reloadable0/timestamped_log/2_0_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable0/scripts/2_0_reloadable0.prx) 2>&1 |& tee -a 2_0_reloadable0/2_0_reloadable0.log 2_0_reloadable0/timestamped_log/2_0_reloadable0-${ts}.log + +2_1_reloadable0: 0_0_reloadable0 + mkdir -p 2_1_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 2_1_reloadable0/Release/2_1_reloadable0.o + set -o pipefail; (rm -f "2_1_reloadable0/Release/2_1_reloadable0.##") 2>&1 |& tee -a 2_1_reloadable0/2_1_reloadable0.log 2_1_reloadable0/timestamped_log/2_1_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable0/scripts/2_1_reloadable0.prx) 2>&1 |& tee -a 2_1_reloadable0/2_1_reloadable0.log 2_1_reloadable0/timestamped_log/2_1_reloadable0-${ts}.log + +2_2_reloadable0: 0_0_reloadable0 + mkdir -p 2_2_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 2_2_reloadable0/Release/2_2_reloadable0.o + set -o pipefail; (rm -f "2_2_reloadable0/Release/2_2_reloadable0.##") 2>&1 |& tee -a 2_2_reloadable0/2_2_reloadable0.log 2_2_reloadable0/timestamped_log/2_2_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable0/scripts/2_2_reloadable0.prx) 2>&1 |& tee -a 2_2_reloadable0/2_2_reloadable0.log 2_2_reloadable0/timestamped_log/2_2_reloadable0-${ts}.log + +2_3_reloadable0: 0_0_reloadable0 + mkdir -p 2_3_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 2_3_reloadable0/Release/2_3_reloadable0.o + set -o pipefail; (rm -f "2_3_reloadable0/Release/2_3_reloadable0.##") 2>&1 |& tee -a 2_3_reloadable0/2_3_reloadable0.log 2_3_reloadable0/timestamped_log/2_3_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable0/scripts/2_3_reloadable0.prx) 2>&1 |& tee -a 2_3_reloadable0/2_3_reloadable0.log 2_3_reloadable0/timestamped_log/2_3_reloadable0-${ts}.log + +3_0_reloadable0: 0_0_reloadable0 + mkdir -p 3_0_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 3_0_reloadable0/Release/3_0_reloadable0.o + set -o pipefail; (rm -f "3_0_reloadable0/Release/3_0_reloadable0.##") 2>&1 |& tee -a 3_0_reloadable0/3_0_reloadable0.log 3_0_reloadable0/timestamped_log/3_0_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable0/scripts/3_0_reloadable0.prx) 2>&1 |& tee -a 3_0_reloadable0/3_0_reloadable0.log 3_0_reloadable0/timestamped_log/3_0_reloadable0-${ts}.log + +3_1_reloadable0: 0_0_reloadable0 + mkdir -p 3_1_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 3_1_reloadable0/Release/3_1_reloadable0.o + set -o pipefail; (rm -f "3_1_reloadable0/Release/3_1_reloadable0.##") 2>&1 |& tee -a 3_1_reloadable0/3_1_reloadable0.log 3_1_reloadable0/timestamped_log/3_1_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable0/scripts/3_1_reloadable0.prx) 2>&1 |& tee -a 3_1_reloadable0/3_1_reloadable0.log 3_1_reloadable0/timestamped_log/3_1_reloadable0-${ts}.log + +3_2_reloadable0: 0_0_reloadable0 + mkdir -p 3_2_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 3_2_reloadable0/Release/3_2_reloadable0.o + set -o pipefail; (rm -f "3_2_reloadable0/Release/3_2_reloadable0.##") 2>&1 |& tee -a 3_2_reloadable0/3_2_reloadable0.log 3_2_reloadable0/timestamped_log/3_2_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable0/scripts/3_2_reloadable0.prx) 2>&1 |& tee -a 3_2_reloadable0/3_2_reloadable0.log 3_2_reloadable0/timestamped_log/3_2_reloadable0-${ts}.log + +3_3_reloadable0: 0_0_reloadable0 + mkdir -p 3_3_reloadable0/Release + cp 0_0_reloadable0/Release/0_0_reloadable0.o 3_3_reloadable0/Release/3_3_reloadable0.o + set -o pipefail; (rm -f "3_3_reloadable0/Release/3_3_reloadable0.##") 2>&1 |& tee -a 3_3_reloadable0/3_3_reloadable0.log 3_3_reloadable0/timestamped_log/3_3_reloadable0-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable0/scripts/3_3_reloadable0.prx) 2>&1 |& tee -a 3_3_reloadable0/3_3_reloadable0.log 3_3_reloadable0/timestamped_log/3_3_reloadable0-${ts}.log + +0_1_reloadable1: 0_0_reloadable1 + mkdir -p 0_1_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 0_1_reloadable1/Release/0_1_reloadable1.o + set -o pipefail; (rm -f "0_1_reloadable1/Release/0_1_reloadable1.##") 2>&1 |& tee -a 0_1_reloadable1/0_1_reloadable1.log 0_1_reloadable1/timestamped_log/0_1_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable1/scripts/0_1_reloadable1.prx) 2>&1 |& tee -a 0_1_reloadable1/0_1_reloadable1.log 0_1_reloadable1/timestamped_log/0_1_reloadable1-${ts}.log + +0_2_reloadable1: 0_0_reloadable1 + mkdir -p 0_2_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 0_2_reloadable1/Release/0_2_reloadable1.o + set -o pipefail; (rm -f "0_2_reloadable1/Release/0_2_reloadable1.##") 2>&1 |& tee -a 0_2_reloadable1/0_2_reloadable1.log 0_2_reloadable1/timestamped_log/0_2_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable1/scripts/0_2_reloadable1.prx) 2>&1 |& tee -a 0_2_reloadable1/0_2_reloadable1.log 0_2_reloadable1/timestamped_log/0_2_reloadable1-${ts}.log + +0_3_reloadable1: 0_0_reloadable1 + mkdir -p 0_3_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 0_3_reloadable1/Release/0_3_reloadable1.o + set -o pipefail; (rm -f "0_3_reloadable1/Release/0_3_reloadable1.##") 2>&1 |& tee -a 0_3_reloadable1/0_3_reloadable1.log 0_3_reloadable1/timestamped_log/0_3_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable1/scripts/0_3_reloadable1.prx) 2>&1 |& tee -a 0_3_reloadable1/0_3_reloadable1.log 0_3_reloadable1/timestamped_log/0_3_reloadable1-${ts}.log + +1_0_reloadable1: 0_0_reloadable1 + mkdir -p 1_0_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 1_0_reloadable1/Release/1_0_reloadable1.o + set -o pipefail; (rm -f "1_0_reloadable1/Release/1_0_reloadable1.##") 2>&1 |& tee -a 1_0_reloadable1/1_0_reloadable1.log 1_0_reloadable1/timestamped_log/1_0_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable1/scripts/1_0_reloadable1.prx) 2>&1 |& tee -a 1_0_reloadable1/1_0_reloadable1.log 1_0_reloadable1/timestamped_log/1_0_reloadable1-${ts}.log + +1_1_reloadable1: 0_0_reloadable1 + mkdir -p 1_1_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 1_1_reloadable1/Release/1_1_reloadable1.o + set -o pipefail; (rm -f "1_1_reloadable1/Release/1_1_reloadable1.##") 2>&1 |& tee -a 1_1_reloadable1/1_1_reloadable1.log 1_1_reloadable1/timestamped_log/1_1_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable1/scripts/1_1_reloadable1.prx) 2>&1 |& tee -a 1_1_reloadable1/1_1_reloadable1.log 1_1_reloadable1/timestamped_log/1_1_reloadable1-${ts}.log + +1_2_reloadable1: 0_0_reloadable1 + mkdir -p 1_2_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 1_2_reloadable1/Release/1_2_reloadable1.o + set -o pipefail; (rm -f "1_2_reloadable1/Release/1_2_reloadable1.##") 2>&1 |& tee -a 1_2_reloadable1/1_2_reloadable1.log 1_2_reloadable1/timestamped_log/1_2_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable1/scripts/1_2_reloadable1.prx) 2>&1 |& tee -a 1_2_reloadable1/1_2_reloadable1.log 1_2_reloadable1/timestamped_log/1_2_reloadable1-${ts}.log + +1_3_reloadable1: 0_0_reloadable1 + mkdir -p 1_3_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 1_3_reloadable1/Release/1_3_reloadable1.o + set -o pipefail; (rm -f "1_3_reloadable1/Release/1_3_reloadable1.##") 2>&1 |& tee -a 1_3_reloadable1/1_3_reloadable1.log 1_3_reloadable1/timestamped_log/1_3_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable1/scripts/1_3_reloadable1.prx) 2>&1 |& tee -a 1_3_reloadable1/1_3_reloadable1.log 1_3_reloadable1/timestamped_log/1_3_reloadable1-${ts}.log + +2_0_reloadable1: 0_0_reloadable1 + mkdir -p 2_0_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 2_0_reloadable1/Release/2_0_reloadable1.o + set -o pipefail; (rm -f "2_0_reloadable1/Release/2_0_reloadable1.##") 2>&1 |& tee -a 2_0_reloadable1/2_0_reloadable1.log 2_0_reloadable1/timestamped_log/2_0_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable1/scripts/2_0_reloadable1.prx) 2>&1 |& tee -a 2_0_reloadable1/2_0_reloadable1.log 2_0_reloadable1/timestamped_log/2_0_reloadable1-${ts}.log + +2_1_reloadable1: 0_0_reloadable1 + mkdir -p 2_1_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 2_1_reloadable1/Release/2_1_reloadable1.o + set -o pipefail; (rm -f "2_1_reloadable1/Release/2_1_reloadable1.##") 2>&1 |& tee -a 2_1_reloadable1/2_1_reloadable1.log 2_1_reloadable1/timestamped_log/2_1_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable1/scripts/2_1_reloadable1.prx) 2>&1 |& tee -a 2_1_reloadable1/2_1_reloadable1.log 2_1_reloadable1/timestamped_log/2_1_reloadable1-${ts}.log + +2_2_reloadable1: 0_0_reloadable1 + mkdir -p 2_2_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 2_2_reloadable1/Release/2_2_reloadable1.o + set -o pipefail; (rm -f "2_2_reloadable1/Release/2_2_reloadable1.##") 2>&1 |& tee -a 2_2_reloadable1/2_2_reloadable1.log 2_2_reloadable1/timestamped_log/2_2_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable1/scripts/2_2_reloadable1.prx) 2>&1 |& tee -a 2_2_reloadable1/2_2_reloadable1.log 2_2_reloadable1/timestamped_log/2_2_reloadable1-${ts}.log + +2_3_reloadable1: 0_0_reloadable1 + mkdir -p 2_3_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 2_3_reloadable1/Release/2_3_reloadable1.o + set -o pipefail; (rm -f "2_3_reloadable1/Release/2_3_reloadable1.##") 2>&1 |& tee -a 2_3_reloadable1/2_3_reloadable1.log 2_3_reloadable1/timestamped_log/2_3_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable1/scripts/2_3_reloadable1.prx) 2>&1 |& tee -a 2_3_reloadable1/2_3_reloadable1.log 2_3_reloadable1/timestamped_log/2_3_reloadable1-${ts}.log + +3_0_reloadable1: 0_0_reloadable1 + mkdir -p 3_0_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 3_0_reloadable1/Release/3_0_reloadable1.o + set -o pipefail; (rm -f "3_0_reloadable1/Release/3_0_reloadable1.##") 2>&1 |& tee -a 3_0_reloadable1/3_0_reloadable1.log 3_0_reloadable1/timestamped_log/3_0_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable1/scripts/3_0_reloadable1.prx) 2>&1 |& tee -a 3_0_reloadable1/3_0_reloadable1.log 3_0_reloadable1/timestamped_log/3_0_reloadable1-${ts}.log + +3_1_reloadable1: 0_0_reloadable1 + mkdir -p 3_1_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 3_1_reloadable1/Release/3_1_reloadable1.o + set -o pipefail; (rm -f "3_1_reloadable1/Release/3_1_reloadable1.##") 2>&1 |& tee -a 3_1_reloadable1/3_1_reloadable1.log 3_1_reloadable1/timestamped_log/3_1_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable1/scripts/3_1_reloadable1.prx) 2>&1 |& tee -a 3_1_reloadable1/3_1_reloadable1.log 3_1_reloadable1/timestamped_log/3_1_reloadable1-${ts}.log + +3_2_reloadable1: 0_0_reloadable1 + mkdir -p 3_2_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 3_2_reloadable1/Release/3_2_reloadable1.o + set -o pipefail; (rm -f "3_2_reloadable1/Release/3_2_reloadable1.##") 2>&1 |& tee -a 3_2_reloadable1/3_2_reloadable1.log 3_2_reloadable1/timestamped_log/3_2_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable1/scripts/3_2_reloadable1.prx) 2>&1 |& tee -a 3_2_reloadable1/3_2_reloadable1.log 3_2_reloadable1/timestamped_log/3_2_reloadable1-${ts}.log + +3_3_reloadable1: 0_0_reloadable1 + mkdir -p 3_3_reloadable1/Release + cp 0_0_reloadable1/Release/0_0_reloadable1.o 3_3_reloadable1/Release/3_3_reloadable1.o + set -o pipefail; (rm -f "3_3_reloadable1/Release/3_3_reloadable1.##") 2>&1 |& tee -a 3_3_reloadable1/3_3_reloadable1.log 3_3_reloadable1/timestamped_log/3_3_reloadable1-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable1/scripts/3_3_reloadable1.prx) 2>&1 |& tee -a 3_3_reloadable1/3_3_reloadable1.log 3_3_reloadable1/timestamped_log/3_3_reloadable1-${ts}.log + +0_0_reloadable4: 0_0_reloadable2 + mkdir -p 0_0_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_0_reloadable4/Release/0_0_reloadable4.o + set -o pipefail; (rm -f "0_0_reloadable4/Release/0_0_reloadable4.##") 2>&1 |& tee -a 0_0_reloadable4/0_0_reloadable4.log 0_0_reloadable4/timestamped_log/0_0_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable4/scripts/0_0_reloadable4.prx) 2>&1 |& tee -a 0_0_reloadable4/0_0_reloadable4.log 0_0_reloadable4/timestamped_log/0_0_reloadable4-${ts}.log + +0_0_reloadable6: 0_0_reloadable2 + mkdir -p 0_0_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_0_reloadable6/Release/0_0_reloadable6.o + set -o pipefail; (rm -f "0_0_reloadable6/Release/0_0_reloadable6.##") 2>&1 |& tee -a 0_0_reloadable6/0_0_reloadable6.log 0_0_reloadable6/timestamped_log/0_0_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable6/scripts/0_0_reloadable6.prx) 2>&1 |& tee -a 0_0_reloadable6/0_0_reloadable6.log 0_0_reloadable6/timestamped_log/0_0_reloadable6-${ts}.log + +0_0_reloadable8: 0_0_reloadable2 + mkdir -p 0_0_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_0_reloadable8/Release/0_0_reloadable8.o + set -o pipefail; (rm -f "0_0_reloadable8/Release/0_0_reloadable8.##") 2>&1 |& tee -a 0_0_reloadable8/0_0_reloadable8.log 0_0_reloadable8/timestamped_log/0_0_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable8/scripts/0_0_reloadable8.prx) 2>&1 |& tee -a 0_0_reloadable8/0_0_reloadable8.log 0_0_reloadable8/timestamped_log/0_0_reloadable8-${ts}.log + +0_0_reloadable10: 0_0_reloadable2 + mkdir -p 0_0_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_0_reloadable10/Release/0_0_reloadable10.o + set -o pipefail; (rm -f "0_0_reloadable10/Release/0_0_reloadable10.##") 2>&1 |& tee -a 0_0_reloadable10/0_0_reloadable10.log 0_0_reloadable10/timestamped_log/0_0_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable10/scripts/0_0_reloadable10.prx) 2>&1 |& tee -a 0_0_reloadable10/0_0_reloadable10.log 0_0_reloadable10/timestamped_log/0_0_reloadable10-${ts}.log + +0_0_reloadable12: 0_0_reloadable2 + mkdir -p 0_0_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_0_reloadable12/Release/0_0_reloadable12.o + set -o pipefail; (rm -f "0_0_reloadable12/Release/0_0_reloadable12.##") 2>&1 |& tee -a 0_0_reloadable12/0_0_reloadable12.log 0_0_reloadable12/timestamped_log/0_0_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable12/scripts/0_0_reloadable12.prx) 2>&1 |& tee -a 0_0_reloadable12/0_0_reloadable12.log 0_0_reloadable12/timestamped_log/0_0_reloadable12-${ts}.log + +0_0_reloadable14: 0_0_reloadable2 + mkdir -p 0_0_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_0_reloadable14/Release/0_0_reloadable14.o + set -o pipefail; (rm -f "0_0_reloadable14/Release/0_0_reloadable14.##") 2>&1 |& tee -a 0_0_reloadable14/0_0_reloadable14.log 0_0_reloadable14/timestamped_log/0_0_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable14/scripts/0_0_reloadable14.prx) 2>&1 |& tee -a 0_0_reloadable14/0_0_reloadable14.log 0_0_reloadable14/timestamped_log/0_0_reloadable14-${ts}.log + +0_0_reloadable16: 0_0_reloadable2 + mkdir -p 0_0_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_0_reloadable16/Release/0_0_reloadable16.o + set -o pipefail; (rm -f "0_0_reloadable16/Release/0_0_reloadable16.##") 2>&1 |& tee -a 0_0_reloadable16/0_0_reloadable16.log 0_0_reloadable16/timestamped_log/0_0_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable16/scripts/0_0_reloadable16.prx) 2>&1 |& tee -a 0_0_reloadable16/0_0_reloadable16.log 0_0_reloadable16/timestamped_log/0_0_reloadable16-${ts}.log + +0_0_reloadable18: 0_0_reloadable2 + mkdir -p 0_0_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_0_reloadable18/Release/0_0_reloadable18.o + set -o pipefail; (rm -f "0_0_reloadable18/Release/0_0_reloadable18.##") 2>&1 |& tee -a 0_0_reloadable18/0_0_reloadable18.log 0_0_reloadable18/timestamped_log/0_0_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable18/scripts/0_0_reloadable18.prx) 2>&1 |& tee -a 0_0_reloadable18/0_0_reloadable18.log 0_0_reloadable18/timestamped_log/0_0_reloadable18-${ts}.log + +0_1_reloadable2: 0_0_reloadable2 + mkdir -p 0_1_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_1_reloadable2/Release/0_1_reloadable2.o + set -o pipefail; (rm -f "0_1_reloadable2/Release/0_1_reloadable2.##") 2>&1 |& tee -a 0_1_reloadable2/0_1_reloadable2.log 0_1_reloadable2/timestamped_log/0_1_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable2/scripts/0_1_reloadable2.prx) 2>&1 |& tee -a 0_1_reloadable2/0_1_reloadable2.log 0_1_reloadable2/timestamped_log/0_1_reloadable2-${ts}.log + +0_1_reloadable4: 0_0_reloadable2 + mkdir -p 0_1_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_1_reloadable4/Release/0_1_reloadable4.o + set -o pipefail; (rm -f "0_1_reloadable4/Release/0_1_reloadable4.##") 2>&1 |& tee -a 0_1_reloadable4/0_1_reloadable4.log 0_1_reloadable4/timestamped_log/0_1_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable4/scripts/0_1_reloadable4.prx) 2>&1 |& tee -a 0_1_reloadable4/0_1_reloadable4.log 0_1_reloadable4/timestamped_log/0_1_reloadable4-${ts}.log + +0_1_reloadable6: 0_0_reloadable2 + mkdir -p 0_1_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_1_reloadable6/Release/0_1_reloadable6.o + set -o pipefail; (rm -f "0_1_reloadable6/Release/0_1_reloadable6.##") 2>&1 |& tee -a 0_1_reloadable6/0_1_reloadable6.log 0_1_reloadable6/timestamped_log/0_1_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable6/scripts/0_1_reloadable6.prx) 2>&1 |& tee -a 0_1_reloadable6/0_1_reloadable6.log 0_1_reloadable6/timestamped_log/0_1_reloadable6-${ts}.log + +0_1_reloadable8: 0_0_reloadable2 + mkdir -p 0_1_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_1_reloadable8/Release/0_1_reloadable8.o + set -o pipefail; (rm -f "0_1_reloadable8/Release/0_1_reloadable8.##") 2>&1 |& tee -a 0_1_reloadable8/0_1_reloadable8.log 0_1_reloadable8/timestamped_log/0_1_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable8/scripts/0_1_reloadable8.prx) 2>&1 |& tee -a 0_1_reloadable8/0_1_reloadable8.log 0_1_reloadable8/timestamped_log/0_1_reloadable8-${ts}.log + +0_1_reloadable10: 0_0_reloadable2 + mkdir -p 0_1_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_1_reloadable10/Release/0_1_reloadable10.o + set -o pipefail; (rm -f "0_1_reloadable10/Release/0_1_reloadable10.##") 2>&1 |& tee -a 0_1_reloadable10/0_1_reloadable10.log 0_1_reloadable10/timestamped_log/0_1_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable10/scripts/0_1_reloadable10.prx) 2>&1 |& tee -a 0_1_reloadable10/0_1_reloadable10.log 0_1_reloadable10/timestamped_log/0_1_reloadable10-${ts}.log + +0_1_reloadable12: 0_0_reloadable2 + mkdir -p 0_1_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_1_reloadable12/Release/0_1_reloadable12.o + set -o pipefail; (rm -f "0_1_reloadable12/Release/0_1_reloadable12.##") 2>&1 |& tee -a 0_1_reloadable12/0_1_reloadable12.log 0_1_reloadable12/timestamped_log/0_1_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable12/scripts/0_1_reloadable12.prx) 2>&1 |& tee -a 0_1_reloadable12/0_1_reloadable12.log 0_1_reloadable12/timestamped_log/0_1_reloadable12-${ts}.log + +0_1_reloadable14: 0_0_reloadable2 + mkdir -p 0_1_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_1_reloadable14/Release/0_1_reloadable14.o + set -o pipefail; (rm -f "0_1_reloadable14/Release/0_1_reloadable14.##") 2>&1 |& tee -a 0_1_reloadable14/0_1_reloadable14.log 0_1_reloadable14/timestamped_log/0_1_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable14/scripts/0_1_reloadable14.prx) 2>&1 |& tee -a 0_1_reloadable14/0_1_reloadable14.log 0_1_reloadable14/timestamped_log/0_1_reloadable14-${ts}.log + +0_1_reloadable16: 0_0_reloadable2 + mkdir -p 0_1_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_1_reloadable16/Release/0_1_reloadable16.o + set -o pipefail; (rm -f "0_1_reloadable16/Release/0_1_reloadable16.##") 2>&1 |& tee -a 0_1_reloadable16/0_1_reloadable16.log 0_1_reloadable16/timestamped_log/0_1_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable16/scripts/0_1_reloadable16.prx) 2>&1 |& tee -a 0_1_reloadable16/0_1_reloadable16.log 0_1_reloadable16/timestamped_log/0_1_reloadable16-${ts}.log + +0_1_reloadable18: 0_0_reloadable2 + mkdir -p 0_1_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_1_reloadable18/Release/0_1_reloadable18.o + set -o pipefail; (rm -f "0_1_reloadable18/Release/0_1_reloadable18.##") 2>&1 |& tee -a 0_1_reloadable18/0_1_reloadable18.log 0_1_reloadable18/timestamped_log/0_1_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable18/scripts/0_1_reloadable18.prx) 2>&1 |& tee -a 0_1_reloadable18/0_1_reloadable18.log 0_1_reloadable18/timestamped_log/0_1_reloadable18-${ts}.log + +0_2_reloadable2: 0_0_reloadable2 + mkdir -p 0_2_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_2_reloadable2/Release/0_2_reloadable2.o + set -o pipefail; (rm -f "0_2_reloadable2/Release/0_2_reloadable2.##") 2>&1 |& tee -a 0_2_reloadable2/0_2_reloadable2.log 0_2_reloadable2/timestamped_log/0_2_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable2/scripts/0_2_reloadable2.prx) 2>&1 |& tee -a 0_2_reloadable2/0_2_reloadable2.log 0_2_reloadable2/timestamped_log/0_2_reloadable2-${ts}.log + +0_2_reloadable4: 0_0_reloadable2 + mkdir -p 0_2_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_2_reloadable4/Release/0_2_reloadable4.o + set -o pipefail; (rm -f "0_2_reloadable4/Release/0_2_reloadable4.##") 2>&1 |& tee -a 0_2_reloadable4/0_2_reloadable4.log 0_2_reloadable4/timestamped_log/0_2_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable4/scripts/0_2_reloadable4.prx) 2>&1 |& tee -a 0_2_reloadable4/0_2_reloadable4.log 0_2_reloadable4/timestamped_log/0_2_reloadable4-${ts}.log + +0_2_reloadable6: 0_0_reloadable2 + mkdir -p 0_2_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_2_reloadable6/Release/0_2_reloadable6.o + set -o pipefail; (rm -f "0_2_reloadable6/Release/0_2_reloadable6.##") 2>&1 |& tee -a 0_2_reloadable6/0_2_reloadable6.log 0_2_reloadable6/timestamped_log/0_2_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable6/scripts/0_2_reloadable6.prx) 2>&1 |& tee -a 0_2_reloadable6/0_2_reloadable6.log 0_2_reloadable6/timestamped_log/0_2_reloadable6-${ts}.log + +0_2_reloadable8: 0_0_reloadable2 + mkdir -p 0_2_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_2_reloadable8/Release/0_2_reloadable8.o + set -o pipefail; (rm -f "0_2_reloadable8/Release/0_2_reloadable8.##") 2>&1 |& tee -a 0_2_reloadable8/0_2_reloadable8.log 0_2_reloadable8/timestamped_log/0_2_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable8/scripts/0_2_reloadable8.prx) 2>&1 |& tee -a 0_2_reloadable8/0_2_reloadable8.log 0_2_reloadable8/timestamped_log/0_2_reloadable8-${ts}.log + +0_2_reloadable10: 0_0_reloadable2 + mkdir -p 0_2_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_2_reloadable10/Release/0_2_reloadable10.o + set -o pipefail; (rm -f "0_2_reloadable10/Release/0_2_reloadable10.##") 2>&1 |& tee -a 0_2_reloadable10/0_2_reloadable10.log 0_2_reloadable10/timestamped_log/0_2_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable10/scripts/0_2_reloadable10.prx) 2>&1 |& tee -a 0_2_reloadable10/0_2_reloadable10.log 0_2_reloadable10/timestamped_log/0_2_reloadable10-${ts}.log + +0_2_reloadable12: 0_0_reloadable2 + mkdir -p 0_2_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_2_reloadable12/Release/0_2_reloadable12.o + set -o pipefail; (rm -f "0_2_reloadable12/Release/0_2_reloadable12.##") 2>&1 |& tee -a 0_2_reloadable12/0_2_reloadable12.log 0_2_reloadable12/timestamped_log/0_2_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable12/scripts/0_2_reloadable12.prx) 2>&1 |& tee -a 0_2_reloadable12/0_2_reloadable12.log 0_2_reloadable12/timestamped_log/0_2_reloadable12-${ts}.log + +0_2_reloadable14: 0_0_reloadable2 + mkdir -p 0_2_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_2_reloadable14/Release/0_2_reloadable14.o + set -o pipefail; (rm -f "0_2_reloadable14/Release/0_2_reloadable14.##") 2>&1 |& tee -a 0_2_reloadable14/0_2_reloadable14.log 0_2_reloadable14/timestamped_log/0_2_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable14/scripts/0_2_reloadable14.prx) 2>&1 |& tee -a 0_2_reloadable14/0_2_reloadable14.log 0_2_reloadable14/timestamped_log/0_2_reloadable14-${ts}.log + +0_2_reloadable16: 0_0_reloadable2 + mkdir -p 0_2_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_2_reloadable16/Release/0_2_reloadable16.o + set -o pipefail; (rm -f "0_2_reloadable16/Release/0_2_reloadable16.##") 2>&1 |& tee -a 0_2_reloadable16/0_2_reloadable16.log 0_2_reloadable16/timestamped_log/0_2_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable16/scripts/0_2_reloadable16.prx) 2>&1 |& tee -a 0_2_reloadable16/0_2_reloadable16.log 0_2_reloadable16/timestamped_log/0_2_reloadable16-${ts}.log + +0_2_reloadable18: 0_0_reloadable2 + mkdir -p 0_2_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_2_reloadable18/Release/0_2_reloadable18.o + set -o pipefail; (rm -f "0_2_reloadable18/Release/0_2_reloadable18.##") 2>&1 |& tee -a 0_2_reloadable18/0_2_reloadable18.log 0_2_reloadable18/timestamped_log/0_2_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable18/scripts/0_2_reloadable18.prx) 2>&1 |& tee -a 0_2_reloadable18/0_2_reloadable18.log 0_2_reloadable18/timestamped_log/0_2_reloadable18-${ts}.log + +0_3_reloadable2: 0_0_reloadable2 + mkdir -p 0_3_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_3_reloadable2/Release/0_3_reloadable2.o + set -o pipefail; (rm -f "0_3_reloadable2/Release/0_3_reloadable2.##") 2>&1 |& tee -a 0_3_reloadable2/0_3_reloadable2.log 0_3_reloadable2/timestamped_log/0_3_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable2/scripts/0_3_reloadable2.prx) 2>&1 |& tee -a 0_3_reloadable2/0_3_reloadable2.log 0_3_reloadable2/timestamped_log/0_3_reloadable2-${ts}.log + +0_3_reloadable4: 0_0_reloadable2 + mkdir -p 0_3_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_3_reloadable4/Release/0_3_reloadable4.o + set -o pipefail; (rm -f "0_3_reloadable4/Release/0_3_reloadable4.##") 2>&1 |& tee -a 0_3_reloadable4/0_3_reloadable4.log 0_3_reloadable4/timestamped_log/0_3_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable4/scripts/0_3_reloadable4.prx) 2>&1 |& tee -a 0_3_reloadable4/0_3_reloadable4.log 0_3_reloadable4/timestamped_log/0_3_reloadable4-${ts}.log + +0_3_reloadable6: 0_0_reloadable2 + mkdir -p 0_3_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_3_reloadable6/Release/0_3_reloadable6.o + set -o pipefail; (rm -f "0_3_reloadable6/Release/0_3_reloadable6.##") 2>&1 |& tee -a 0_3_reloadable6/0_3_reloadable6.log 0_3_reloadable6/timestamped_log/0_3_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable6/scripts/0_3_reloadable6.prx) 2>&1 |& tee -a 0_3_reloadable6/0_3_reloadable6.log 0_3_reloadable6/timestamped_log/0_3_reloadable6-${ts}.log + +0_3_reloadable8: 0_0_reloadable2 + mkdir -p 0_3_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_3_reloadable8/Release/0_3_reloadable8.o + set -o pipefail; (rm -f "0_3_reloadable8/Release/0_3_reloadable8.##") 2>&1 |& tee -a 0_3_reloadable8/0_3_reloadable8.log 0_3_reloadable8/timestamped_log/0_3_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable8/scripts/0_3_reloadable8.prx) 2>&1 |& tee -a 0_3_reloadable8/0_3_reloadable8.log 0_3_reloadable8/timestamped_log/0_3_reloadable8-${ts}.log + +0_3_reloadable10: 0_0_reloadable2 + mkdir -p 0_3_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_3_reloadable10/Release/0_3_reloadable10.o + set -o pipefail; (rm -f "0_3_reloadable10/Release/0_3_reloadable10.##") 2>&1 |& tee -a 0_3_reloadable10/0_3_reloadable10.log 0_3_reloadable10/timestamped_log/0_3_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable10/scripts/0_3_reloadable10.prx) 2>&1 |& tee -a 0_3_reloadable10/0_3_reloadable10.log 0_3_reloadable10/timestamped_log/0_3_reloadable10-${ts}.log + +0_3_reloadable12: 0_0_reloadable2 + mkdir -p 0_3_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_3_reloadable12/Release/0_3_reloadable12.o + set -o pipefail; (rm -f "0_3_reloadable12/Release/0_3_reloadable12.##") 2>&1 |& tee -a 0_3_reloadable12/0_3_reloadable12.log 0_3_reloadable12/timestamped_log/0_3_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable12/scripts/0_3_reloadable12.prx) 2>&1 |& tee -a 0_3_reloadable12/0_3_reloadable12.log 0_3_reloadable12/timestamped_log/0_3_reloadable12-${ts}.log + +0_3_reloadable14: 0_0_reloadable2 + mkdir -p 0_3_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_3_reloadable14/Release/0_3_reloadable14.o + set -o pipefail; (rm -f "0_3_reloadable14/Release/0_3_reloadable14.##") 2>&1 |& tee -a 0_3_reloadable14/0_3_reloadable14.log 0_3_reloadable14/timestamped_log/0_3_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable14/scripts/0_3_reloadable14.prx) 2>&1 |& tee -a 0_3_reloadable14/0_3_reloadable14.log 0_3_reloadable14/timestamped_log/0_3_reloadable14-${ts}.log + +0_3_reloadable16: 0_0_reloadable2 + mkdir -p 0_3_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_3_reloadable16/Release/0_3_reloadable16.o + set -o pipefail; (rm -f "0_3_reloadable16/Release/0_3_reloadable16.##") 2>&1 |& tee -a 0_3_reloadable16/0_3_reloadable16.log 0_3_reloadable16/timestamped_log/0_3_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable16/scripts/0_3_reloadable16.prx) 2>&1 |& tee -a 0_3_reloadable16/0_3_reloadable16.log 0_3_reloadable16/timestamped_log/0_3_reloadable16-${ts}.log + +0_3_reloadable18: 0_0_reloadable2 + mkdir -p 0_3_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 0_3_reloadable18/Release/0_3_reloadable18.o + set -o pipefail; (rm -f "0_3_reloadable18/Release/0_3_reloadable18.##") 2>&1 |& tee -a 0_3_reloadable18/0_3_reloadable18.log 0_3_reloadable18/timestamped_log/0_3_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable18/scripts/0_3_reloadable18.prx) 2>&1 |& tee -a 0_3_reloadable18/0_3_reloadable18.log 0_3_reloadable18/timestamped_log/0_3_reloadable18-${ts}.log + +1_0_reloadable2: 0_0_reloadable2 + mkdir -p 1_0_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_0_reloadable2/Release/1_0_reloadable2.o + set -o pipefail; (rm -f "1_0_reloadable2/Release/1_0_reloadable2.##") 2>&1 |& tee -a 1_0_reloadable2/1_0_reloadable2.log 1_0_reloadable2/timestamped_log/1_0_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable2/scripts/1_0_reloadable2.prx) 2>&1 |& tee -a 1_0_reloadable2/1_0_reloadable2.log 1_0_reloadable2/timestamped_log/1_0_reloadable2-${ts}.log + +1_0_reloadable4: 0_0_reloadable2 + mkdir -p 1_0_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_0_reloadable4/Release/1_0_reloadable4.o + set -o pipefail; (rm -f "1_0_reloadable4/Release/1_0_reloadable4.##") 2>&1 |& tee -a 1_0_reloadable4/1_0_reloadable4.log 1_0_reloadable4/timestamped_log/1_0_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable4/scripts/1_0_reloadable4.prx) 2>&1 |& tee -a 1_0_reloadable4/1_0_reloadable4.log 1_0_reloadable4/timestamped_log/1_0_reloadable4-${ts}.log + +1_0_reloadable6: 0_0_reloadable2 + mkdir -p 1_0_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_0_reloadable6/Release/1_0_reloadable6.o + set -o pipefail; (rm -f "1_0_reloadable6/Release/1_0_reloadable6.##") 2>&1 |& tee -a 1_0_reloadable6/1_0_reloadable6.log 1_0_reloadable6/timestamped_log/1_0_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable6/scripts/1_0_reloadable6.prx) 2>&1 |& tee -a 1_0_reloadable6/1_0_reloadable6.log 1_0_reloadable6/timestamped_log/1_0_reloadable6-${ts}.log + +1_0_reloadable8: 0_0_reloadable2 + mkdir -p 1_0_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_0_reloadable8/Release/1_0_reloadable8.o + set -o pipefail; (rm -f "1_0_reloadable8/Release/1_0_reloadable8.##") 2>&1 |& tee -a 1_0_reloadable8/1_0_reloadable8.log 1_0_reloadable8/timestamped_log/1_0_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable8/scripts/1_0_reloadable8.prx) 2>&1 |& tee -a 1_0_reloadable8/1_0_reloadable8.log 1_0_reloadable8/timestamped_log/1_0_reloadable8-${ts}.log + +1_0_reloadable10: 0_0_reloadable2 + mkdir -p 1_0_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_0_reloadable10/Release/1_0_reloadable10.o + set -o pipefail; (rm -f "1_0_reloadable10/Release/1_0_reloadable10.##") 2>&1 |& tee -a 1_0_reloadable10/1_0_reloadable10.log 1_0_reloadable10/timestamped_log/1_0_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable10/scripts/1_0_reloadable10.prx) 2>&1 |& tee -a 1_0_reloadable10/1_0_reloadable10.log 1_0_reloadable10/timestamped_log/1_0_reloadable10-${ts}.log + +1_0_reloadable12: 0_0_reloadable2 + mkdir -p 1_0_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_0_reloadable12/Release/1_0_reloadable12.o + set -o pipefail; (rm -f "1_0_reloadable12/Release/1_0_reloadable12.##") 2>&1 |& tee -a 1_0_reloadable12/1_0_reloadable12.log 1_0_reloadable12/timestamped_log/1_0_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable12/scripts/1_0_reloadable12.prx) 2>&1 |& tee -a 1_0_reloadable12/1_0_reloadable12.log 1_0_reloadable12/timestamped_log/1_0_reloadable12-${ts}.log + +1_0_reloadable14: 0_0_reloadable2 + mkdir -p 1_0_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_0_reloadable14/Release/1_0_reloadable14.o + set -o pipefail; (rm -f "1_0_reloadable14/Release/1_0_reloadable14.##") 2>&1 |& tee -a 1_0_reloadable14/1_0_reloadable14.log 1_0_reloadable14/timestamped_log/1_0_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable14/scripts/1_0_reloadable14.prx) 2>&1 |& tee -a 1_0_reloadable14/1_0_reloadable14.log 1_0_reloadable14/timestamped_log/1_0_reloadable14-${ts}.log + +1_0_reloadable16: 0_0_reloadable2 + mkdir -p 1_0_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_0_reloadable16/Release/1_0_reloadable16.o + set -o pipefail; (rm -f "1_0_reloadable16/Release/1_0_reloadable16.##") 2>&1 |& tee -a 1_0_reloadable16/1_0_reloadable16.log 1_0_reloadable16/timestamped_log/1_0_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable16/scripts/1_0_reloadable16.prx) 2>&1 |& tee -a 1_0_reloadable16/1_0_reloadable16.log 1_0_reloadable16/timestamped_log/1_0_reloadable16-${ts}.log + +1_0_reloadable18: 0_0_reloadable2 + mkdir -p 1_0_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_0_reloadable18/Release/1_0_reloadable18.o + set -o pipefail; (rm -f "1_0_reloadable18/Release/1_0_reloadable18.##") 2>&1 |& tee -a 1_0_reloadable18/1_0_reloadable18.log 1_0_reloadable18/timestamped_log/1_0_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable18/scripts/1_0_reloadable18.prx) 2>&1 |& tee -a 1_0_reloadable18/1_0_reloadable18.log 1_0_reloadable18/timestamped_log/1_0_reloadable18-${ts}.log + +1_1_reloadable2: 0_0_reloadable2 + mkdir -p 1_1_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_1_reloadable2/Release/1_1_reloadable2.o + set -o pipefail; (rm -f "1_1_reloadable2/Release/1_1_reloadable2.##") 2>&1 |& tee -a 1_1_reloadable2/1_1_reloadable2.log 1_1_reloadable2/timestamped_log/1_1_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable2/scripts/1_1_reloadable2.prx) 2>&1 |& tee -a 1_1_reloadable2/1_1_reloadable2.log 1_1_reloadable2/timestamped_log/1_1_reloadable2-${ts}.log + +1_1_reloadable4: 0_0_reloadable2 + mkdir -p 1_1_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_1_reloadable4/Release/1_1_reloadable4.o + set -o pipefail; (rm -f "1_1_reloadable4/Release/1_1_reloadable4.##") 2>&1 |& tee -a 1_1_reloadable4/1_1_reloadable4.log 1_1_reloadable4/timestamped_log/1_1_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable4/scripts/1_1_reloadable4.prx) 2>&1 |& tee -a 1_1_reloadable4/1_1_reloadable4.log 1_1_reloadable4/timestamped_log/1_1_reloadable4-${ts}.log + +1_1_reloadable6: 0_0_reloadable2 + mkdir -p 1_1_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_1_reloadable6/Release/1_1_reloadable6.o + set -o pipefail; (rm -f "1_1_reloadable6/Release/1_1_reloadable6.##") 2>&1 |& tee -a 1_1_reloadable6/1_1_reloadable6.log 1_1_reloadable6/timestamped_log/1_1_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable6/scripts/1_1_reloadable6.prx) 2>&1 |& tee -a 1_1_reloadable6/1_1_reloadable6.log 1_1_reloadable6/timestamped_log/1_1_reloadable6-${ts}.log + +1_1_reloadable8: 0_0_reloadable2 + mkdir -p 1_1_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_1_reloadable8/Release/1_1_reloadable8.o + set -o pipefail; (rm -f "1_1_reloadable8/Release/1_1_reloadable8.##") 2>&1 |& tee -a 1_1_reloadable8/1_1_reloadable8.log 1_1_reloadable8/timestamped_log/1_1_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable8/scripts/1_1_reloadable8.prx) 2>&1 |& tee -a 1_1_reloadable8/1_1_reloadable8.log 1_1_reloadable8/timestamped_log/1_1_reloadable8-${ts}.log + +1_1_reloadable10: 0_0_reloadable2 + mkdir -p 1_1_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_1_reloadable10/Release/1_1_reloadable10.o + set -o pipefail; (rm -f "1_1_reloadable10/Release/1_1_reloadable10.##") 2>&1 |& tee -a 1_1_reloadable10/1_1_reloadable10.log 1_1_reloadable10/timestamped_log/1_1_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable10/scripts/1_1_reloadable10.prx) 2>&1 |& tee -a 1_1_reloadable10/1_1_reloadable10.log 1_1_reloadable10/timestamped_log/1_1_reloadable10-${ts}.log + +1_1_reloadable12: 0_0_reloadable2 + mkdir -p 1_1_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_1_reloadable12/Release/1_1_reloadable12.o + set -o pipefail; (rm -f "1_1_reloadable12/Release/1_1_reloadable12.##") 2>&1 |& tee -a 1_1_reloadable12/1_1_reloadable12.log 1_1_reloadable12/timestamped_log/1_1_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable12/scripts/1_1_reloadable12.prx) 2>&1 |& tee -a 1_1_reloadable12/1_1_reloadable12.log 1_1_reloadable12/timestamped_log/1_1_reloadable12-${ts}.log + +1_1_reloadable14: 0_0_reloadable2 + mkdir -p 1_1_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_1_reloadable14/Release/1_1_reloadable14.o + set -o pipefail; (rm -f "1_1_reloadable14/Release/1_1_reloadable14.##") 2>&1 |& tee -a 1_1_reloadable14/1_1_reloadable14.log 1_1_reloadable14/timestamped_log/1_1_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable14/scripts/1_1_reloadable14.prx) 2>&1 |& tee -a 1_1_reloadable14/1_1_reloadable14.log 1_1_reloadable14/timestamped_log/1_1_reloadable14-${ts}.log + +1_1_reloadable16: 0_0_reloadable2 + mkdir -p 1_1_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_1_reloadable16/Release/1_1_reloadable16.o + set -o pipefail; (rm -f "1_1_reloadable16/Release/1_1_reloadable16.##") 2>&1 |& tee -a 1_1_reloadable16/1_1_reloadable16.log 1_1_reloadable16/timestamped_log/1_1_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable16/scripts/1_1_reloadable16.prx) 2>&1 |& tee -a 1_1_reloadable16/1_1_reloadable16.log 1_1_reloadable16/timestamped_log/1_1_reloadable16-${ts}.log + +1_1_reloadable18: 0_0_reloadable2 + mkdir -p 1_1_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_1_reloadable18/Release/1_1_reloadable18.o + set -o pipefail; (rm -f "1_1_reloadable18/Release/1_1_reloadable18.##") 2>&1 |& tee -a 1_1_reloadable18/1_1_reloadable18.log 1_1_reloadable18/timestamped_log/1_1_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable18/scripts/1_1_reloadable18.prx) 2>&1 |& tee -a 1_1_reloadable18/1_1_reloadable18.log 1_1_reloadable18/timestamped_log/1_1_reloadable18-${ts}.log + +1_2_reloadable2: 0_0_reloadable2 + mkdir -p 1_2_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_2_reloadable2/Release/1_2_reloadable2.o + set -o pipefail; (rm -f "1_2_reloadable2/Release/1_2_reloadable2.##") 2>&1 |& tee -a 1_2_reloadable2/1_2_reloadable2.log 1_2_reloadable2/timestamped_log/1_2_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable2/scripts/1_2_reloadable2.prx) 2>&1 |& tee -a 1_2_reloadable2/1_2_reloadable2.log 1_2_reloadable2/timestamped_log/1_2_reloadable2-${ts}.log + +1_2_reloadable4: 0_0_reloadable2 + mkdir -p 1_2_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_2_reloadable4/Release/1_2_reloadable4.o + set -o pipefail; (rm -f "1_2_reloadable4/Release/1_2_reloadable4.##") 2>&1 |& tee -a 1_2_reloadable4/1_2_reloadable4.log 1_2_reloadable4/timestamped_log/1_2_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable4/scripts/1_2_reloadable4.prx) 2>&1 |& tee -a 1_2_reloadable4/1_2_reloadable4.log 1_2_reloadable4/timestamped_log/1_2_reloadable4-${ts}.log + +1_2_reloadable6: 0_0_reloadable2 + mkdir -p 1_2_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_2_reloadable6/Release/1_2_reloadable6.o + set -o pipefail; (rm -f "1_2_reloadable6/Release/1_2_reloadable6.##") 2>&1 |& tee -a 1_2_reloadable6/1_2_reloadable6.log 1_2_reloadable6/timestamped_log/1_2_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable6/scripts/1_2_reloadable6.prx) 2>&1 |& tee -a 1_2_reloadable6/1_2_reloadable6.log 1_2_reloadable6/timestamped_log/1_2_reloadable6-${ts}.log + +1_2_reloadable8: 0_0_reloadable2 + mkdir -p 1_2_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_2_reloadable8/Release/1_2_reloadable8.o + set -o pipefail; (rm -f "1_2_reloadable8/Release/1_2_reloadable8.##") 2>&1 |& tee -a 1_2_reloadable8/1_2_reloadable8.log 1_2_reloadable8/timestamped_log/1_2_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable8/scripts/1_2_reloadable8.prx) 2>&1 |& tee -a 1_2_reloadable8/1_2_reloadable8.log 1_2_reloadable8/timestamped_log/1_2_reloadable8-${ts}.log + +1_2_reloadable10: 0_0_reloadable2 + mkdir -p 1_2_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_2_reloadable10/Release/1_2_reloadable10.o + set -o pipefail; (rm -f "1_2_reloadable10/Release/1_2_reloadable10.##") 2>&1 |& tee -a 1_2_reloadable10/1_2_reloadable10.log 1_2_reloadable10/timestamped_log/1_2_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable10/scripts/1_2_reloadable10.prx) 2>&1 |& tee -a 1_2_reloadable10/1_2_reloadable10.log 1_2_reloadable10/timestamped_log/1_2_reloadable10-${ts}.log + +1_2_reloadable12: 0_0_reloadable2 + mkdir -p 1_2_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_2_reloadable12/Release/1_2_reloadable12.o + set -o pipefail; (rm -f "1_2_reloadable12/Release/1_2_reloadable12.##") 2>&1 |& tee -a 1_2_reloadable12/1_2_reloadable12.log 1_2_reloadable12/timestamped_log/1_2_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable12/scripts/1_2_reloadable12.prx) 2>&1 |& tee -a 1_2_reloadable12/1_2_reloadable12.log 1_2_reloadable12/timestamped_log/1_2_reloadable12-${ts}.log + +1_2_reloadable14: 0_0_reloadable2 + mkdir -p 1_2_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_2_reloadable14/Release/1_2_reloadable14.o + set -o pipefail; (rm -f "1_2_reloadable14/Release/1_2_reloadable14.##") 2>&1 |& tee -a 1_2_reloadable14/1_2_reloadable14.log 1_2_reloadable14/timestamped_log/1_2_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable14/scripts/1_2_reloadable14.prx) 2>&1 |& tee -a 1_2_reloadable14/1_2_reloadable14.log 1_2_reloadable14/timestamped_log/1_2_reloadable14-${ts}.log + +1_2_reloadable16: 0_0_reloadable2 + mkdir -p 1_2_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_2_reloadable16/Release/1_2_reloadable16.o + set -o pipefail; (rm -f "1_2_reloadable16/Release/1_2_reloadable16.##") 2>&1 |& tee -a 1_2_reloadable16/1_2_reloadable16.log 1_2_reloadable16/timestamped_log/1_2_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable16/scripts/1_2_reloadable16.prx) 2>&1 |& tee -a 1_2_reloadable16/1_2_reloadable16.log 1_2_reloadable16/timestamped_log/1_2_reloadable16-${ts}.log + +1_2_reloadable18: 0_0_reloadable2 + mkdir -p 1_2_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_2_reloadable18/Release/1_2_reloadable18.o + set -o pipefail; (rm -f "1_2_reloadable18/Release/1_2_reloadable18.##") 2>&1 |& tee -a 1_2_reloadable18/1_2_reloadable18.log 1_2_reloadable18/timestamped_log/1_2_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable18/scripts/1_2_reloadable18.prx) 2>&1 |& tee -a 1_2_reloadable18/1_2_reloadable18.log 1_2_reloadable18/timestamped_log/1_2_reloadable18-${ts}.log + +1_3_reloadable2: 0_0_reloadable2 + mkdir -p 1_3_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_3_reloadable2/Release/1_3_reloadable2.o + set -o pipefail; (rm -f "1_3_reloadable2/Release/1_3_reloadable2.##") 2>&1 |& tee -a 1_3_reloadable2/1_3_reloadable2.log 1_3_reloadable2/timestamped_log/1_3_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable2/scripts/1_3_reloadable2.prx) 2>&1 |& tee -a 1_3_reloadable2/1_3_reloadable2.log 1_3_reloadable2/timestamped_log/1_3_reloadable2-${ts}.log + +1_3_reloadable4: 0_0_reloadable2 + mkdir -p 1_3_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_3_reloadable4/Release/1_3_reloadable4.o + set -o pipefail; (rm -f "1_3_reloadable4/Release/1_3_reloadable4.##") 2>&1 |& tee -a 1_3_reloadable4/1_3_reloadable4.log 1_3_reloadable4/timestamped_log/1_3_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable4/scripts/1_3_reloadable4.prx) 2>&1 |& tee -a 1_3_reloadable4/1_3_reloadable4.log 1_3_reloadable4/timestamped_log/1_3_reloadable4-${ts}.log + +1_3_reloadable6: 0_0_reloadable2 + mkdir -p 1_3_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_3_reloadable6/Release/1_3_reloadable6.o + set -o pipefail; (rm -f "1_3_reloadable6/Release/1_3_reloadable6.##") 2>&1 |& tee -a 1_3_reloadable6/1_3_reloadable6.log 1_3_reloadable6/timestamped_log/1_3_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable6/scripts/1_3_reloadable6.prx) 2>&1 |& tee -a 1_3_reloadable6/1_3_reloadable6.log 1_3_reloadable6/timestamped_log/1_3_reloadable6-${ts}.log + +1_3_reloadable8: 0_0_reloadable2 + mkdir -p 1_3_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_3_reloadable8/Release/1_3_reloadable8.o + set -o pipefail; (rm -f "1_3_reloadable8/Release/1_3_reloadable8.##") 2>&1 |& tee -a 1_3_reloadable8/1_3_reloadable8.log 1_3_reloadable8/timestamped_log/1_3_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable8/scripts/1_3_reloadable8.prx) 2>&1 |& tee -a 1_3_reloadable8/1_3_reloadable8.log 1_3_reloadable8/timestamped_log/1_3_reloadable8-${ts}.log + +1_3_reloadable10: 0_0_reloadable2 + mkdir -p 1_3_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_3_reloadable10/Release/1_3_reloadable10.o + set -o pipefail; (rm -f "1_3_reloadable10/Release/1_3_reloadable10.##") 2>&1 |& tee -a 1_3_reloadable10/1_3_reloadable10.log 1_3_reloadable10/timestamped_log/1_3_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable10/scripts/1_3_reloadable10.prx) 2>&1 |& tee -a 1_3_reloadable10/1_3_reloadable10.log 1_3_reloadable10/timestamped_log/1_3_reloadable10-${ts}.log + +1_3_reloadable12: 0_0_reloadable2 + mkdir -p 1_3_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_3_reloadable12/Release/1_3_reloadable12.o + set -o pipefail; (rm -f "1_3_reloadable12/Release/1_3_reloadable12.##") 2>&1 |& tee -a 1_3_reloadable12/1_3_reloadable12.log 1_3_reloadable12/timestamped_log/1_3_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable12/scripts/1_3_reloadable12.prx) 2>&1 |& tee -a 1_3_reloadable12/1_3_reloadable12.log 1_3_reloadable12/timestamped_log/1_3_reloadable12-${ts}.log + +1_3_reloadable14: 0_0_reloadable2 + mkdir -p 1_3_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_3_reloadable14/Release/1_3_reloadable14.o + set -o pipefail; (rm -f "1_3_reloadable14/Release/1_3_reloadable14.##") 2>&1 |& tee -a 1_3_reloadable14/1_3_reloadable14.log 1_3_reloadable14/timestamped_log/1_3_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable14/scripts/1_3_reloadable14.prx) 2>&1 |& tee -a 1_3_reloadable14/1_3_reloadable14.log 1_3_reloadable14/timestamped_log/1_3_reloadable14-${ts}.log + +1_3_reloadable16: 0_0_reloadable2 + mkdir -p 1_3_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_3_reloadable16/Release/1_3_reloadable16.o + set -o pipefail; (rm -f "1_3_reloadable16/Release/1_3_reloadable16.##") 2>&1 |& tee -a 1_3_reloadable16/1_3_reloadable16.log 1_3_reloadable16/timestamped_log/1_3_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable16/scripts/1_3_reloadable16.prx) 2>&1 |& tee -a 1_3_reloadable16/1_3_reloadable16.log 1_3_reloadable16/timestamped_log/1_3_reloadable16-${ts}.log + +1_3_reloadable18: 0_0_reloadable2 + mkdir -p 1_3_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 1_3_reloadable18/Release/1_3_reloadable18.o + set -o pipefail; (rm -f "1_3_reloadable18/Release/1_3_reloadable18.##") 2>&1 |& tee -a 1_3_reloadable18/1_3_reloadable18.log 1_3_reloadable18/timestamped_log/1_3_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable18/scripts/1_3_reloadable18.prx) 2>&1 |& tee -a 1_3_reloadable18/1_3_reloadable18.log 1_3_reloadable18/timestamped_log/1_3_reloadable18-${ts}.log + +2_0_reloadable2: 0_0_reloadable2 + mkdir -p 2_0_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_0_reloadable2/Release/2_0_reloadable2.o + set -o pipefail; (rm -f "2_0_reloadable2/Release/2_0_reloadable2.##") 2>&1 |& tee -a 2_0_reloadable2/2_0_reloadable2.log 2_0_reloadable2/timestamped_log/2_0_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable2/scripts/2_0_reloadable2.prx) 2>&1 |& tee -a 2_0_reloadable2/2_0_reloadable2.log 2_0_reloadable2/timestamped_log/2_0_reloadable2-${ts}.log + +2_0_reloadable4: 0_0_reloadable2 + mkdir -p 2_0_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_0_reloadable4/Release/2_0_reloadable4.o + set -o pipefail; (rm -f "2_0_reloadable4/Release/2_0_reloadable4.##") 2>&1 |& tee -a 2_0_reloadable4/2_0_reloadable4.log 2_0_reloadable4/timestamped_log/2_0_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable4/scripts/2_0_reloadable4.prx) 2>&1 |& tee -a 2_0_reloadable4/2_0_reloadable4.log 2_0_reloadable4/timestamped_log/2_0_reloadable4-${ts}.log + +2_0_reloadable6: 0_0_reloadable2 + mkdir -p 2_0_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_0_reloadable6/Release/2_0_reloadable6.o + set -o pipefail; (rm -f "2_0_reloadable6/Release/2_0_reloadable6.##") 2>&1 |& tee -a 2_0_reloadable6/2_0_reloadable6.log 2_0_reloadable6/timestamped_log/2_0_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable6/scripts/2_0_reloadable6.prx) 2>&1 |& tee -a 2_0_reloadable6/2_0_reloadable6.log 2_0_reloadable6/timestamped_log/2_0_reloadable6-${ts}.log + +2_0_reloadable8: 0_0_reloadable2 + mkdir -p 2_0_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_0_reloadable8/Release/2_0_reloadable8.o + set -o pipefail; (rm -f "2_0_reloadable8/Release/2_0_reloadable8.##") 2>&1 |& tee -a 2_0_reloadable8/2_0_reloadable8.log 2_0_reloadable8/timestamped_log/2_0_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable8/scripts/2_0_reloadable8.prx) 2>&1 |& tee -a 2_0_reloadable8/2_0_reloadable8.log 2_0_reloadable8/timestamped_log/2_0_reloadable8-${ts}.log + +2_0_reloadable10: 0_0_reloadable2 + mkdir -p 2_0_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_0_reloadable10/Release/2_0_reloadable10.o + set -o pipefail; (rm -f "2_0_reloadable10/Release/2_0_reloadable10.##") 2>&1 |& tee -a 2_0_reloadable10/2_0_reloadable10.log 2_0_reloadable10/timestamped_log/2_0_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable10/scripts/2_0_reloadable10.prx) 2>&1 |& tee -a 2_0_reloadable10/2_0_reloadable10.log 2_0_reloadable10/timestamped_log/2_0_reloadable10-${ts}.log + +2_0_reloadable12: 0_0_reloadable2 + mkdir -p 2_0_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_0_reloadable12/Release/2_0_reloadable12.o + set -o pipefail; (rm -f "2_0_reloadable12/Release/2_0_reloadable12.##") 2>&1 |& tee -a 2_0_reloadable12/2_0_reloadable12.log 2_0_reloadable12/timestamped_log/2_0_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable12/scripts/2_0_reloadable12.prx) 2>&1 |& tee -a 2_0_reloadable12/2_0_reloadable12.log 2_0_reloadable12/timestamped_log/2_0_reloadable12-${ts}.log + +2_0_reloadable14: 0_0_reloadable2 + mkdir -p 2_0_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_0_reloadable14/Release/2_0_reloadable14.o + set -o pipefail; (rm -f "2_0_reloadable14/Release/2_0_reloadable14.##") 2>&1 |& tee -a 2_0_reloadable14/2_0_reloadable14.log 2_0_reloadable14/timestamped_log/2_0_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable14/scripts/2_0_reloadable14.prx) 2>&1 |& tee -a 2_0_reloadable14/2_0_reloadable14.log 2_0_reloadable14/timestamped_log/2_0_reloadable14-${ts}.log + +2_0_reloadable16: 0_0_reloadable2 + mkdir -p 2_0_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_0_reloadable16/Release/2_0_reloadable16.o + set -o pipefail; (rm -f "2_0_reloadable16/Release/2_0_reloadable16.##") 2>&1 |& tee -a 2_0_reloadable16/2_0_reloadable16.log 2_0_reloadable16/timestamped_log/2_0_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable16/scripts/2_0_reloadable16.prx) 2>&1 |& tee -a 2_0_reloadable16/2_0_reloadable16.log 2_0_reloadable16/timestamped_log/2_0_reloadable16-${ts}.log + +2_0_reloadable18: 0_0_reloadable2 + mkdir -p 2_0_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_0_reloadable18/Release/2_0_reloadable18.o + set -o pipefail; (rm -f "2_0_reloadable18/Release/2_0_reloadable18.##") 2>&1 |& tee -a 2_0_reloadable18/2_0_reloadable18.log 2_0_reloadable18/timestamped_log/2_0_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable18/scripts/2_0_reloadable18.prx) 2>&1 |& tee -a 2_0_reloadable18/2_0_reloadable18.log 2_0_reloadable18/timestamped_log/2_0_reloadable18-${ts}.log + +2_1_reloadable2: 0_0_reloadable2 + mkdir -p 2_1_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_1_reloadable2/Release/2_1_reloadable2.o + set -o pipefail; (rm -f "2_1_reloadable2/Release/2_1_reloadable2.##") 2>&1 |& tee -a 2_1_reloadable2/2_1_reloadable2.log 2_1_reloadable2/timestamped_log/2_1_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable2/scripts/2_1_reloadable2.prx) 2>&1 |& tee -a 2_1_reloadable2/2_1_reloadable2.log 2_1_reloadable2/timestamped_log/2_1_reloadable2-${ts}.log + +2_1_reloadable4: 0_0_reloadable2 + mkdir -p 2_1_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_1_reloadable4/Release/2_1_reloadable4.o + set -o pipefail; (rm -f "2_1_reloadable4/Release/2_1_reloadable4.##") 2>&1 |& tee -a 2_1_reloadable4/2_1_reloadable4.log 2_1_reloadable4/timestamped_log/2_1_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable4/scripts/2_1_reloadable4.prx) 2>&1 |& tee -a 2_1_reloadable4/2_1_reloadable4.log 2_1_reloadable4/timestamped_log/2_1_reloadable4-${ts}.log + +2_1_reloadable6: 0_0_reloadable2 + mkdir -p 2_1_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_1_reloadable6/Release/2_1_reloadable6.o + set -o pipefail; (rm -f "2_1_reloadable6/Release/2_1_reloadable6.##") 2>&1 |& tee -a 2_1_reloadable6/2_1_reloadable6.log 2_1_reloadable6/timestamped_log/2_1_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable6/scripts/2_1_reloadable6.prx) 2>&1 |& tee -a 2_1_reloadable6/2_1_reloadable6.log 2_1_reloadable6/timestamped_log/2_1_reloadable6-${ts}.log + +2_1_reloadable8: 0_0_reloadable2 + mkdir -p 2_1_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_1_reloadable8/Release/2_1_reloadable8.o + set -o pipefail; (rm -f "2_1_reloadable8/Release/2_1_reloadable8.##") 2>&1 |& tee -a 2_1_reloadable8/2_1_reloadable8.log 2_1_reloadable8/timestamped_log/2_1_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable8/scripts/2_1_reloadable8.prx) 2>&1 |& tee -a 2_1_reloadable8/2_1_reloadable8.log 2_1_reloadable8/timestamped_log/2_1_reloadable8-${ts}.log + +2_1_reloadable10: 0_0_reloadable2 + mkdir -p 2_1_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_1_reloadable10/Release/2_1_reloadable10.o + set -o pipefail; (rm -f "2_1_reloadable10/Release/2_1_reloadable10.##") 2>&1 |& tee -a 2_1_reloadable10/2_1_reloadable10.log 2_1_reloadable10/timestamped_log/2_1_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable10/scripts/2_1_reloadable10.prx) 2>&1 |& tee -a 2_1_reloadable10/2_1_reloadable10.log 2_1_reloadable10/timestamped_log/2_1_reloadable10-${ts}.log + +2_1_reloadable12: 0_0_reloadable2 + mkdir -p 2_1_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_1_reloadable12/Release/2_1_reloadable12.o + set -o pipefail; (rm -f "2_1_reloadable12/Release/2_1_reloadable12.##") 2>&1 |& tee -a 2_1_reloadable12/2_1_reloadable12.log 2_1_reloadable12/timestamped_log/2_1_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable12/scripts/2_1_reloadable12.prx) 2>&1 |& tee -a 2_1_reloadable12/2_1_reloadable12.log 2_1_reloadable12/timestamped_log/2_1_reloadable12-${ts}.log + +2_1_reloadable14: 0_0_reloadable2 + mkdir -p 2_1_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_1_reloadable14/Release/2_1_reloadable14.o + set -o pipefail; (rm -f "2_1_reloadable14/Release/2_1_reloadable14.##") 2>&1 |& tee -a 2_1_reloadable14/2_1_reloadable14.log 2_1_reloadable14/timestamped_log/2_1_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable14/scripts/2_1_reloadable14.prx) 2>&1 |& tee -a 2_1_reloadable14/2_1_reloadable14.log 2_1_reloadable14/timestamped_log/2_1_reloadable14-${ts}.log + +2_1_reloadable16: 0_0_reloadable2 + mkdir -p 2_1_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_1_reloadable16/Release/2_1_reloadable16.o + set -o pipefail; (rm -f "2_1_reloadable16/Release/2_1_reloadable16.##") 2>&1 |& tee -a 2_1_reloadable16/2_1_reloadable16.log 2_1_reloadable16/timestamped_log/2_1_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable16/scripts/2_1_reloadable16.prx) 2>&1 |& tee -a 2_1_reloadable16/2_1_reloadable16.log 2_1_reloadable16/timestamped_log/2_1_reloadable16-${ts}.log + +2_1_reloadable18: 0_0_reloadable2 + mkdir -p 2_1_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_1_reloadable18/Release/2_1_reloadable18.o + set -o pipefail; (rm -f "2_1_reloadable18/Release/2_1_reloadable18.##") 2>&1 |& tee -a 2_1_reloadable18/2_1_reloadable18.log 2_1_reloadable18/timestamped_log/2_1_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable18/scripts/2_1_reloadable18.prx) 2>&1 |& tee -a 2_1_reloadable18/2_1_reloadable18.log 2_1_reloadable18/timestamped_log/2_1_reloadable18-${ts}.log + +2_2_reloadable2: 0_0_reloadable2 + mkdir -p 2_2_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_2_reloadable2/Release/2_2_reloadable2.o + set -o pipefail; (rm -f "2_2_reloadable2/Release/2_2_reloadable2.##") 2>&1 |& tee -a 2_2_reloadable2/2_2_reloadable2.log 2_2_reloadable2/timestamped_log/2_2_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable2/scripts/2_2_reloadable2.prx) 2>&1 |& tee -a 2_2_reloadable2/2_2_reloadable2.log 2_2_reloadable2/timestamped_log/2_2_reloadable2-${ts}.log + +2_2_reloadable4: 0_0_reloadable2 + mkdir -p 2_2_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_2_reloadable4/Release/2_2_reloadable4.o + set -o pipefail; (rm -f "2_2_reloadable4/Release/2_2_reloadable4.##") 2>&1 |& tee -a 2_2_reloadable4/2_2_reloadable4.log 2_2_reloadable4/timestamped_log/2_2_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable4/scripts/2_2_reloadable4.prx) 2>&1 |& tee -a 2_2_reloadable4/2_2_reloadable4.log 2_2_reloadable4/timestamped_log/2_2_reloadable4-${ts}.log + +2_2_reloadable6: 0_0_reloadable2 + mkdir -p 2_2_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_2_reloadable6/Release/2_2_reloadable6.o + set -o pipefail; (rm -f "2_2_reloadable6/Release/2_2_reloadable6.##") 2>&1 |& tee -a 2_2_reloadable6/2_2_reloadable6.log 2_2_reloadable6/timestamped_log/2_2_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable6/scripts/2_2_reloadable6.prx) 2>&1 |& tee -a 2_2_reloadable6/2_2_reloadable6.log 2_2_reloadable6/timestamped_log/2_2_reloadable6-${ts}.log + +2_2_reloadable8: 0_0_reloadable2 + mkdir -p 2_2_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_2_reloadable8/Release/2_2_reloadable8.o + set -o pipefail; (rm -f "2_2_reloadable8/Release/2_2_reloadable8.##") 2>&1 |& tee -a 2_2_reloadable8/2_2_reloadable8.log 2_2_reloadable8/timestamped_log/2_2_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable8/scripts/2_2_reloadable8.prx) 2>&1 |& tee -a 2_2_reloadable8/2_2_reloadable8.log 2_2_reloadable8/timestamped_log/2_2_reloadable8-${ts}.log + +2_2_reloadable10: 0_0_reloadable2 + mkdir -p 2_2_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_2_reloadable10/Release/2_2_reloadable10.o + set -o pipefail; (rm -f "2_2_reloadable10/Release/2_2_reloadable10.##") 2>&1 |& tee -a 2_2_reloadable10/2_2_reloadable10.log 2_2_reloadable10/timestamped_log/2_2_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable10/scripts/2_2_reloadable10.prx) 2>&1 |& tee -a 2_2_reloadable10/2_2_reloadable10.log 2_2_reloadable10/timestamped_log/2_2_reloadable10-${ts}.log + +2_2_reloadable12: 0_0_reloadable2 + mkdir -p 2_2_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_2_reloadable12/Release/2_2_reloadable12.o + set -o pipefail; (rm -f "2_2_reloadable12/Release/2_2_reloadable12.##") 2>&1 |& tee -a 2_2_reloadable12/2_2_reloadable12.log 2_2_reloadable12/timestamped_log/2_2_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable12/scripts/2_2_reloadable12.prx) 2>&1 |& tee -a 2_2_reloadable12/2_2_reloadable12.log 2_2_reloadable12/timestamped_log/2_2_reloadable12-${ts}.log + +2_2_reloadable14: 0_0_reloadable2 + mkdir -p 2_2_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_2_reloadable14/Release/2_2_reloadable14.o + set -o pipefail; (rm -f "2_2_reloadable14/Release/2_2_reloadable14.##") 2>&1 |& tee -a 2_2_reloadable14/2_2_reloadable14.log 2_2_reloadable14/timestamped_log/2_2_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable14/scripts/2_2_reloadable14.prx) 2>&1 |& tee -a 2_2_reloadable14/2_2_reloadable14.log 2_2_reloadable14/timestamped_log/2_2_reloadable14-${ts}.log + +2_2_reloadable16: 0_0_reloadable2 + mkdir -p 2_2_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_2_reloadable16/Release/2_2_reloadable16.o + set -o pipefail; (rm -f "2_2_reloadable16/Release/2_2_reloadable16.##") 2>&1 |& tee -a 2_2_reloadable16/2_2_reloadable16.log 2_2_reloadable16/timestamped_log/2_2_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable16/scripts/2_2_reloadable16.prx) 2>&1 |& tee -a 2_2_reloadable16/2_2_reloadable16.log 2_2_reloadable16/timestamped_log/2_2_reloadable16-${ts}.log + +2_2_reloadable18: 0_0_reloadable2 + mkdir -p 2_2_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_2_reloadable18/Release/2_2_reloadable18.o + set -o pipefail; (rm -f "2_2_reloadable18/Release/2_2_reloadable18.##") 2>&1 |& tee -a 2_2_reloadable18/2_2_reloadable18.log 2_2_reloadable18/timestamped_log/2_2_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable18/scripts/2_2_reloadable18.prx) 2>&1 |& tee -a 2_2_reloadable18/2_2_reloadable18.log 2_2_reloadable18/timestamped_log/2_2_reloadable18-${ts}.log + +2_3_reloadable2: 0_0_reloadable2 + mkdir -p 2_3_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_3_reloadable2/Release/2_3_reloadable2.o + set -o pipefail; (rm -f "2_3_reloadable2/Release/2_3_reloadable2.##") 2>&1 |& tee -a 2_3_reloadable2/2_3_reloadable2.log 2_3_reloadable2/timestamped_log/2_3_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable2/scripts/2_3_reloadable2.prx) 2>&1 |& tee -a 2_3_reloadable2/2_3_reloadable2.log 2_3_reloadable2/timestamped_log/2_3_reloadable2-${ts}.log + +2_3_reloadable4: 0_0_reloadable2 + mkdir -p 2_3_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_3_reloadable4/Release/2_3_reloadable4.o + set -o pipefail; (rm -f "2_3_reloadable4/Release/2_3_reloadable4.##") 2>&1 |& tee -a 2_3_reloadable4/2_3_reloadable4.log 2_3_reloadable4/timestamped_log/2_3_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable4/scripts/2_3_reloadable4.prx) 2>&1 |& tee -a 2_3_reloadable4/2_3_reloadable4.log 2_3_reloadable4/timestamped_log/2_3_reloadable4-${ts}.log + +2_3_reloadable6: 0_0_reloadable2 + mkdir -p 2_3_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_3_reloadable6/Release/2_3_reloadable6.o + set -o pipefail; (rm -f "2_3_reloadable6/Release/2_3_reloadable6.##") 2>&1 |& tee -a 2_3_reloadable6/2_3_reloadable6.log 2_3_reloadable6/timestamped_log/2_3_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable6/scripts/2_3_reloadable6.prx) 2>&1 |& tee -a 2_3_reloadable6/2_3_reloadable6.log 2_3_reloadable6/timestamped_log/2_3_reloadable6-${ts}.log + +2_3_reloadable8: 0_0_reloadable2 + mkdir -p 2_3_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_3_reloadable8/Release/2_3_reloadable8.o + set -o pipefail; (rm -f "2_3_reloadable8/Release/2_3_reloadable8.##") 2>&1 |& tee -a 2_3_reloadable8/2_3_reloadable8.log 2_3_reloadable8/timestamped_log/2_3_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable8/scripts/2_3_reloadable8.prx) 2>&1 |& tee -a 2_3_reloadable8/2_3_reloadable8.log 2_3_reloadable8/timestamped_log/2_3_reloadable8-${ts}.log + +2_3_reloadable10: 0_0_reloadable2 + mkdir -p 2_3_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_3_reloadable10/Release/2_3_reloadable10.o + set -o pipefail; (rm -f "2_3_reloadable10/Release/2_3_reloadable10.##") 2>&1 |& tee -a 2_3_reloadable10/2_3_reloadable10.log 2_3_reloadable10/timestamped_log/2_3_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable10/scripts/2_3_reloadable10.prx) 2>&1 |& tee -a 2_3_reloadable10/2_3_reloadable10.log 2_3_reloadable10/timestamped_log/2_3_reloadable10-${ts}.log + +2_3_reloadable12: 0_0_reloadable2 + mkdir -p 2_3_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_3_reloadable12/Release/2_3_reloadable12.o + set -o pipefail; (rm -f "2_3_reloadable12/Release/2_3_reloadable12.##") 2>&1 |& tee -a 2_3_reloadable12/2_3_reloadable12.log 2_3_reloadable12/timestamped_log/2_3_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable12/scripts/2_3_reloadable12.prx) 2>&1 |& tee -a 2_3_reloadable12/2_3_reloadable12.log 2_3_reloadable12/timestamped_log/2_3_reloadable12-${ts}.log + +2_3_reloadable14: 0_0_reloadable2 + mkdir -p 2_3_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_3_reloadable14/Release/2_3_reloadable14.o + set -o pipefail; (rm -f "2_3_reloadable14/Release/2_3_reloadable14.##") 2>&1 |& tee -a 2_3_reloadable14/2_3_reloadable14.log 2_3_reloadable14/timestamped_log/2_3_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable14/scripts/2_3_reloadable14.prx) 2>&1 |& tee -a 2_3_reloadable14/2_3_reloadable14.log 2_3_reloadable14/timestamped_log/2_3_reloadable14-${ts}.log + +2_3_reloadable16: 0_0_reloadable2 + mkdir -p 2_3_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_3_reloadable16/Release/2_3_reloadable16.o + set -o pipefail; (rm -f "2_3_reloadable16/Release/2_3_reloadable16.##") 2>&1 |& tee -a 2_3_reloadable16/2_3_reloadable16.log 2_3_reloadable16/timestamped_log/2_3_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable16/scripts/2_3_reloadable16.prx) 2>&1 |& tee -a 2_3_reloadable16/2_3_reloadable16.log 2_3_reloadable16/timestamped_log/2_3_reloadable16-${ts}.log + +2_3_reloadable18: 0_0_reloadable2 + mkdir -p 2_3_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 2_3_reloadable18/Release/2_3_reloadable18.o + set -o pipefail; (rm -f "2_3_reloadable18/Release/2_3_reloadable18.##") 2>&1 |& tee -a 2_3_reloadable18/2_3_reloadable18.log 2_3_reloadable18/timestamped_log/2_3_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable18/scripts/2_3_reloadable18.prx) 2>&1 |& tee -a 2_3_reloadable18/2_3_reloadable18.log 2_3_reloadable18/timestamped_log/2_3_reloadable18-${ts}.log + +3_0_reloadable2: 0_0_reloadable2 + mkdir -p 3_0_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_0_reloadable2/Release/3_0_reloadable2.o + set -o pipefail; (rm -f "3_0_reloadable2/Release/3_0_reloadable2.##") 2>&1 |& tee -a 3_0_reloadable2/3_0_reloadable2.log 3_0_reloadable2/timestamped_log/3_0_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable2/scripts/3_0_reloadable2.prx) 2>&1 |& tee -a 3_0_reloadable2/3_0_reloadable2.log 3_0_reloadable2/timestamped_log/3_0_reloadable2-${ts}.log + +3_0_reloadable4: 0_0_reloadable2 + mkdir -p 3_0_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_0_reloadable4/Release/3_0_reloadable4.o + set -o pipefail; (rm -f "3_0_reloadable4/Release/3_0_reloadable4.##") 2>&1 |& tee -a 3_0_reloadable4/3_0_reloadable4.log 3_0_reloadable4/timestamped_log/3_0_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable4/scripts/3_0_reloadable4.prx) 2>&1 |& tee -a 3_0_reloadable4/3_0_reloadable4.log 3_0_reloadable4/timestamped_log/3_0_reloadable4-${ts}.log + +3_0_reloadable6: 0_0_reloadable2 + mkdir -p 3_0_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_0_reloadable6/Release/3_0_reloadable6.o + set -o pipefail; (rm -f "3_0_reloadable6/Release/3_0_reloadable6.##") 2>&1 |& tee -a 3_0_reloadable6/3_0_reloadable6.log 3_0_reloadable6/timestamped_log/3_0_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable6/scripts/3_0_reloadable6.prx) 2>&1 |& tee -a 3_0_reloadable6/3_0_reloadable6.log 3_0_reloadable6/timestamped_log/3_0_reloadable6-${ts}.log + +3_0_reloadable8: 0_0_reloadable2 + mkdir -p 3_0_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_0_reloadable8/Release/3_0_reloadable8.o + set -o pipefail; (rm -f "3_0_reloadable8/Release/3_0_reloadable8.##") 2>&1 |& tee -a 3_0_reloadable8/3_0_reloadable8.log 3_0_reloadable8/timestamped_log/3_0_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable8/scripts/3_0_reloadable8.prx) 2>&1 |& tee -a 3_0_reloadable8/3_0_reloadable8.log 3_0_reloadable8/timestamped_log/3_0_reloadable8-${ts}.log + +3_0_reloadable10: 0_0_reloadable2 + mkdir -p 3_0_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_0_reloadable10/Release/3_0_reloadable10.o + set -o pipefail; (rm -f "3_0_reloadable10/Release/3_0_reloadable10.##") 2>&1 |& tee -a 3_0_reloadable10/3_0_reloadable10.log 3_0_reloadable10/timestamped_log/3_0_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable10/scripts/3_0_reloadable10.prx) 2>&1 |& tee -a 3_0_reloadable10/3_0_reloadable10.log 3_0_reloadable10/timestamped_log/3_0_reloadable10-${ts}.log + +3_0_reloadable12: 0_0_reloadable2 + mkdir -p 3_0_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_0_reloadable12/Release/3_0_reloadable12.o + set -o pipefail; (rm -f "3_0_reloadable12/Release/3_0_reloadable12.##") 2>&1 |& tee -a 3_0_reloadable12/3_0_reloadable12.log 3_0_reloadable12/timestamped_log/3_0_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable12/scripts/3_0_reloadable12.prx) 2>&1 |& tee -a 3_0_reloadable12/3_0_reloadable12.log 3_0_reloadable12/timestamped_log/3_0_reloadable12-${ts}.log + +3_0_reloadable14: 0_0_reloadable2 + mkdir -p 3_0_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_0_reloadable14/Release/3_0_reloadable14.o + set -o pipefail; (rm -f "3_0_reloadable14/Release/3_0_reloadable14.##") 2>&1 |& tee -a 3_0_reloadable14/3_0_reloadable14.log 3_0_reloadable14/timestamped_log/3_0_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable14/scripts/3_0_reloadable14.prx) 2>&1 |& tee -a 3_0_reloadable14/3_0_reloadable14.log 3_0_reloadable14/timestamped_log/3_0_reloadable14-${ts}.log + +3_0_reloadable16: 0_0_reloadable2 + mkdir -p 3_0_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_0_reloadable16/Release/3_0_reloadable16.o + set -o pipefail; (rm -f "3_0_reloadable16/Release/3_0_reloadable16.##") 2>&1 |& tee -a 3_0_reloadable16/3_0_reloadable16.log 3_0_reloadable16/timestamped_log/3_0_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable16/scripts/3_0_reloadable16.prx) 2>&1 |& tee -a 3_0_reloadable16/3_0_reloadable16.log 3_0_reloadable16/timestamped_log/3_0_reloadable16-${ts}.log + +3_0_reloadable18: 0_0_reloadable2 + mkdir -p 3_0_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_0_reloadable18/Release/3_0_reloadable18.o + set -o pipefail; (rm -f "3_0_reloadable18/Release/3_0_reloadable18.##") 2>&1 |& tee -a 3_0_reloadable18/3_0_reloadable18.log 3_0_reloadable18/timestamped_log/3_0_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable18/scripts/3_0_reloadable18.prx) 2>&1 |& tee -a 3_0_reloadable18/3_0_reloadable18.log 3_0_reloadable18/timestamped_log/3_0_reloadable18-${ts}.log + +3_1_reloadable2: 0_0_reloadable2 + mkdir -p 3_1_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_1_reloadable2/Release/3_1_reloadable2.o + set -o pipefail; (rm -f "3_1_reloadable2/Release/3_1_reloadable2.##") 2>&1 |& tee -a 3_1_reloadable2/3_1_reloadable2.log 3_1_reloadable2/timestamped_log/3_1_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable2/scripts/3_1_reloadable2.prx) 2>&1 |& tee -a 3_1_reloadable2/3_1_reloadable2.log 3_1_reloadable2/timestamped_log/3_1_reloadable2-${ts}.log + +3_1_reloadable4: 0_0_reloadable2 + mkdir -p 3_1_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_1_reloadable4/Release/3_1_reloadable4.o + set -o pipefail; (rm -f "3_1_reloadable4/Release/3_1_reloadable4.##") 2>&1 |& tee -a 3_1_reloadable4/3_1_reloadable4.log 3_1_reloadable4/timestamped_log/3_1_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable4/scripts/3_1_reloadable4.prx) 2>&1 |& tee -a 3_1_reloadable4/3_1_reloadable4.log 3_1_reloadable4/timestamped_log/3_1_reloadable4-${ts}.log + +3_1_reloadable6: 0_0_reloadable2 + mkdir -p 3_1_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_1_reloadable6/Release/3_1_reloadable6.o + set -o pipefail; (rm -f "3_1_reloadable6/Release/3_1_reloadable6.##") 2>&1 |& tee -a 3_1_reloadable6/3_1_reloadable6.log 3_1_reloadable6/timestamped_log/3_1_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable6/scripts/3_1_reloadable6.prx) 2>&1 |& tee -a 3_1_reloadable6/3_1_reloadable6.log 3_1_reloadable6/timestamped_log/3_1_reloadable6-${ts}.log + +3_1_reloadable8: 0_0_reloadable2 + mkdir -p 3_1_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_1_reloadable8/Release/3_1_reloadable8.o + set -o pipefail; (rm -f "3_1_reloadable8/Release/3_1_reloadable8.##") 2>&1 |& tee -a 3_1_reloadable8/3_1_reloadable8.log 3_1_reloadable8/timestamped_log/3_1_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable8/scripts/3_1_reloadable8.prx) 2>&1 |& tee -a 3_1_reloadable8/3_1_reloadable8.log 3_1_reloadable8/timestamped_log/3_1_reloadable8-${ts}.log + +3_1_reloadable10: 0_0_reloadable2 + mkdir -p 3_1_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_1_reloadable10/Release/3_1_reloadable10.o + set -o pipefail; (rm -f "3_1_reloadable10/Release/3_1_reloadable10.##") 2>&1 |& tee -a 3_1_reloadable10/3_1_reloadable10.log 3_1_reloadable10/timestamped_log/3_1_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable10/scripts/3_1_reloadable10.prx) 2>&1 |& tee -a 3_1_reloadable10/3_1_reloadable10.log 3_1_reloadable10/timestamped_log/3_1_reloadable10-${ts}.log + +3_1_reloadable12: 0_0_reloadable2 + mkdir -p 3_1_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_1_reloadable12/Release/3_1_reloadable12.o + set -o pipefail; (rm -f "3_1_reloadable12/Release/3_1_reloadable12.##") 2>&1 |& tee -a 3_1_reloadable12/3_1_reloadable12.log 3_1_reloadable12/timestamped_log/3_1_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable12/scripts/3_1_reloadable12.prx) 2>&1 |& tee -a 3_1_reloadable12/3_1_reloadable12.log 3_1_reloadable12/timestamped_log/3_1_reloadable12-${ts}.log + +3_1_reloadable14: 0_0_reloadable2 + mkdir -p 3_1_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_1_reloadable14/Release/3_1_reloadable14.o + set -o pipefail; (rm -f "3_1_reloadable14/Release/3_1_reloadable14.##") 2>&1 |& tee -a 3_1_reloadable14/3_1_reloadable14.log 3_1_reloadable14/timestamped_log/3_1_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable14/scripts/3_1_reloadable14.prx) 2>&1 |& tee -a 3_1_reloadable14/3_1_reloadable14.log 3_1_reloadable14/timestamped_log/3_1_reloadable14-${ts}.log + +3_1_reloadable16: 0_0_reloadable2 + mkdir -p 3_1_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_1_reloadable16/Release/3_1_reloadable16.o + set -o pipefail; (rm -f "3_1_reloadable16/Release/3_1_reloadable16.##") 2>&1 |& tee -a 3_1_reloadable16/3_1_reloadable16.log 3_1_reloadable16/timestamped_log/3_1_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable16/scripts/3_1_reloadable16.prx) 2>&1 |& tee -a 3_1_reloadable16/3_1_reloadable16.log 3_1_reloadable16/timestamped_log/3_1_reloadable16-${ts}.log + +3_1_reloadable18: 0_0_reloadable2 + mkdir -p 3_1_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_1_reloadable18/Release/3_1_reloadable18.o + set -o pipefail; (rm -f "3_1_reloadable18/Release/3_1_reloadable18.##") 2>&1 |& tee -a 3_1_reloadable18/3_1_reloadable18.log 3_1_reloadable18/timestamped_log/3_1_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable18/scripts/3_1_reloadable18.prx) 2>&1 |& tee -a 3_1_reloadable18/3_1_reloadable18.log 3_1_reloadable18/timestamped_log/3_1_reloadable18-${ts}.log + +3_2_reloadable2: 0_0_reloadable2 + mkdir -p 3_2_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_2_reloadable2/Release/3_2_reloadable2.o + set -o pipefail; (rm -f "3_2_reloadable2/Release/3_2_reloadable2.##") 2>&1 |& tee -a 3_2_reloadable2/3_2_reloadable2.log 3_2_reloadable2/timestamped_log/3_2_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable2/scripts/3_2_reloadable2.prx) 2>&1 |& tee -a 3_2_reloadable2/3_2_reloadable2.log 3_2_reloadable2/timestamped_log/3_2_reloadable2-${ts}.log + +3_2_reloadable4: 0_0_reloadable2 + mkdir -p 3_2_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_2_reloadable4/Release/3_2_reloadable4.o + set -o pipefail; (rm -f "3_2_reloadable4/Release/3_2_reloadable4.##") 2>&1 |& tee -a 3_2_reloadable4/3_2_reloadable4.log 3_2_reloadable4/timestamped_log/3_2_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable4/scripts/3_2_reloadable4.prx) 2>&1 |& tee -a 3_2_reloadable4/3_2_reloadable4.log 3_2_reloadable4/timestamped_log/3_2_reloadable4-${ts}.log + +3_2_reloadable6: 0_0_reloadable2 + mkdir -p 3_2_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_2_reloadable6/Release/3_2_reloadable6.o + set -o pipefail; (rm -f "3_2_reloadable6/Release/3_2_reloadable6.##") 2>&1 |& tee -a 3_2_reloadable6/3_2_reloadable6.log 3_2_reloadable6/timestamped_log/3_2_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable6/scripts/3_2_reloadable6.prx) 2>&1 |& tee -a 3_2_reloadable6/3_2_reloadable6.log 3_2_reloadable6/timestamped_log/3_2_reloadable6-${ts}.log + +3_2_reloadable8: 0_0_reloadable2 + mkdir -p 3_2_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_2_reloadable8/Release/3_2_reloadable8.o + set -o pipefail; (rm -f "3_2_reloadable8/Release/3_2_reloadable8.##") 2>&1 |& tee -a 3_2_reloadable8/3_2_reloadable8.log 3_2_reloadable8/timestamped_log/3_2_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable8/scripts/3_2_reloadable8.prx) 2>&1 |& tee -a 3_2_reloadable8/3_2_reloadable8.log 3_2_reloadable8/timestamped_log/3_2_reloadable8-${ts}.log + +3_2_reloadable10: 0_0_reloadable2 + mkdir -p 3_2_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_2_reloadable10/Release/3_2_reloadable10.o + set -o pipefail; (rm -f "3_2_reloadable10/Release/3_2_reloadable10.##") 2>&1 |& tee -a 3_2_reloadable10/3_2_reloadable10.log 3_2_reloadable10/timestamped_log/3_2_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable10/scripts/3_2_reloadable10.prx) 2>&1 |& tee -a 3_2_reloadable10/3_2_reloadable10.log 3_2_reloadable10/timestamped_log/3_2_reloadable10-${ts}.log + +3_2_reloadable12: 0_0_reloadable2 + mkdir -p 3_2_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_2_reloadable12/Release/3_2_reloadable12.o + set -o pipefail; (rm -f "3_2_reloadable12/Release/3_2_reloadable12.##") 2>&1 |& tee -a 3_2_reloadable12/3_2_reloadable12.log 3_2_reloadable12/timestamped_log/3_2_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable12/scripts/3_2_reloadable12.prx) 2>&1 |& tee -a 3_2_reloadable12/3_2_reloadable12.log 3_2_reloadable12/timestamped_log/3_2_reloadable12-${ts}.log + +3_2_reloadable14: 0_0_reloadable2 + mkdir -p 3_2_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_2_reloadable14/Release/3_2_reloadable14.o + set -o pipefail; (rm -f "3_2_reloadable14/Release/3_2_reloadable14.##") 2>&1 |& tee -a 3_2_reloadable14/3_2_reloadable14.log 3_2_reloadable14/timestamped_log/3_2_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable14/scripts/3_2_reloadable14.prx) 2>&1 |& tee -a 3_2_reloadable14/3_2_reloadable14.log 3_2_reloadable14/timestamped_log/3_2_reloadable14-${ts}.log + +3_2_reloadable16: 0_0_reloadable2 + mkdir -p 3_2_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_2_reloadable16/Release/3_2_reloadable16.o + set -o pipefail; (rm -f "3_2_reloadable16/Release/3_2_reloadable16.##") 2>&1 |& tee -a 3_2_reloadable16/3_2_reloadable16.log 3_2_reloadable16/timestamped_log/3_2_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable16/scripts/3_2_reloadable16.prx) 2>&1 |& tee -a 3_2_reloadable16/3_2_reloadable16.log 3_2_reloadable16/timestamped_log/3_2_reloadable16-${ts}.log + +3_2_reloadable18: 0_0_reloadable2 + mkdir -p 3_2_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_2_reloadable18/Release/3_2_reloadable18.o + set -o pipefail; (rm -f "3_2_reloadable18/Release/3_2_reloadable18.##") 2>&1 |& tee -a 3_2_reloadable18/3_2_reloadable18.log 3_2_reloadable18/timestamped_log/3_2_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable18/scripts/3_2_reloadable18.prx) 2>&1 |& tee -a 3_2_reloadable18/3_2_reloadable18.log 3_2_reloadable18/timestamped_log/3_2_reloadable18-${ts}.log + +3_3_reloadable2: 0_0_reloadable2 + mkdir -p 3_3_reloadable2/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_3_reloadable2/Release/3_3_reloadable2.o + set -o pipefail; (rm -f "3_3_reloadable2/Release/3_3_reloadable2.##") 2>&1 |& tee -a 3_3_reloadable2/3_3_reloadable2.log 3_3_reloadable2/timestamped_log/3_3_reloadable2-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable2/scripts/3_3_reloadable2.prx) 2>&1 |& tee -a 3_3_reloadable2/3_3_reloadable2.log 3_3_reloadable2/timestamped_log/3_3_reloadable2-${ts}.log + +3_3_reloadable4: 0_0_reloadable2 + mkdir -p 3_3_reloadable4/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_3_reloadable4/Release/3_3_reloadable4.o + set -o pipefail; (rm -f "3_3_reloadable4/Release/3_3_reloadable4.##") 2>&1 |& tee -a 3_3_reloadable4/3_3_reloadable4.log 3_3_reloadable4/timestamped_log/3_3_reloadable4-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable4/scripts/3_3_reloadable4.prx) 2>&1 |& tee -a 3_3_reloadable4/3_3_reloadable4.log 3_3_reloadable4/timestamped_log/3_3_reloadable4-${ts}.log + +3_3_reloadable6: 0_0_reloadable2 + mkdir -p 3_3_reloadable6/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_3_reloadable6/Release/3_3_reloadable6.o + set -o pipefail; (rm -f "3_3_reloadable6/Release/3_3_reloadable6.##") 2>&1 |& tee -a 3_3_reloadable6/3_3_reloadable6.log 3_3_reloadable6/timestamped_log/3_3_reloadable6-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable6/scripts/3_3_reloadable6.prx) 2>&1 |& tee -a 3_3_reloadable6/3_3_reloadable6.log 3_3_reloadable6/timestamped_log/3_3_reloadable6-${ts}.log + +3_3_reloadable8: 0_0_reloadable2 + mkdir -p 3_3_reloadable8/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_3_reloadable8/Release/3_3_reloadable8.o + set -o pipefail; (rm -f "3_3_reloadable8/Release/3_3_reloadable8.##") 2>&1 |& tee -a 3_3_reloadable8/3_3_reloadable8.log 3_3_reloadable8/timestamped_log/3_3_reloadable8-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable8/scripts/3_3_reloadable8.prx) 2>&1 |& tee -a 3_3_reloadable8/3_3_reloadable8.log 3_3_reloadable8/timestamped_log/3_3_reloadable8-${ts}.log + +3_3_reloadable10: 0_0_reloadable2 + mkdir -p 3_3_reloadable10/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_3_reloadable10/Release/3_3_reloadable10.o + set -o pipefail; (rm -f "3_3_reloadable10/Release/3_3_reloadable10.##") 2>&1 |& tee -a 3_3_reloadable10/3_3_reloadable10.log 3_3_reloadable10/timestamped_log/3_3_reloadable10-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable10/scripts/3_3_reloadable10.prx) 2>&1 |& tee -a 3_3_reloadable10/3_3_reloadable10.log 3_3_reloadable10/timestamped_log/3_3_reloadable10-${ts}.log + +3_3_reloadable12: 0_0_reloadable2 + mkdir -p 3_3_reloadable12/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_3_reloadable12/Release/3_3_reloadable12.o + set -o pipefail; (rm -f "3_3_reloadable12/Release/3_3_reloadable12.##") 2>&1 |& tee -a 3_3_reloadable12/3_3_reloadable12.log 3_3_reloadable12/timestamped_log/3_3_reloadable12-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable12/scripts/3_3_reloadable12.prx) 2>&1 |& tee -a 3_3_reloadable12/3_3_reloadable12.log 3_3_reloadable12/timestamped_log/3_3_reloadable12-${ts}.log + +3_3_reloadable14: 0_0_reloadable2 + mkdir -p 3_3_reloadable14/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_3_reloadable14/Release/3_3_reloadable14.o + set -o pipefail; (rm -f "3_3_reloadable14/Release/3_3_reloadable14.##") 2>&1 |& tee -a 3_3_reloadable14/3_3_reloadable14.log 3_3_reloadable14/timestamped_log/3_3_reloadable14-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable14/scripts/3_3_reloadable14.prx) 2>&1 |& tee -a 3_3_reloadable14/3_3_reloadable14.log 3_3_reloadable14/timestamped_log/3_3_reloadable14-${ts}.log + +3_3_reloadable16: 0_0_reloadable2 + mkdir -p 3_3_reloadable16/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_3_reloadable16/Release/3_3_reloadable16.o + set -o pipefail; (rm -f "3_3_reloadable16/Release/3_3_reloadable16.##") 2>&1 |& tee -a 3_3_reloadable16/3_3_reloadable16.log 3_3_reloadable16/timestamped_log/3_3_reloadable16-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable16/scripts/3_3_reloadable16.prx) 2>&1 |& tee -a 3_3_reloadable16/3_3_reloadable16.log 3_3_reloadable16/timestamped_log/3_3_reloadable16-${ts}.log + +3_3_reloadable18: 0_0_reloadable2 + mkdir -p 3_3_reloadable18/Release + cp 0_0_reloadable2/Release/0_0_reloadable2.o 3_3_reloadable18/Release/3_3_reloadable18.o + set -o pipefail; (rm -f "3_3_reloadable18/Release/3_3_reloadable18.##") 2>&1 |& tee -a 3_3_reloadable18/3_3_reloadable18.log 3_3_reloadable18/timestamped_log/3_3_reloadable18-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable18/scripts/3_3_reloadable18.prx) 2>&1 |& tee -a 3_3_reloadable18/3_3_reloadable18.log 3_3_reloadable18/timestamped_log/3_3_reloadable18-${ts}.log + +0_0_reloadable9: 0_0_reloadable3 + mkdir -p 0_0_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_0_reloadable9/Release/0_0_reloadable9.o + set -o pipefail; (rm -f "0_0_reloadable9/Release/0_0_reloadable9.##") 2>&1 |& tee -a 0_0_reloadable9/0_0_reloadable9.log 0_0_reloadable9/timestamped_log/0_0_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable9/scripts/0_0_reloadable9.prx) 2>&1 |& tee -a 0_0_reloadable9/0_0_reloadable9.log 0_0_reloadable9/timestamped_log/0_0_reloadable9-${ts}.log + +0_0_reloadable13: 0_0_reloadable3 + mkdir -p 0_0_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_0_reloadable13/Release/0_0_reloadable13.o + set -o pipefail; (rm -f "0_0_reloadable13/Release/0_0_reloadable13.##") 2>&1 |& tee -a 0_0_reloadable13/0_0_reloadable13.log 0_0_reloadable13/timestamped_log/0_0_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable13/scripts/0_0_reloadable13.prx) 2>&1 |& tee -a 0_0_reloadable13/0_0_reloadable13.log 0_0_reloadable13/timestamped_log/0_0_reloadable13-${ts}.log + +0_1_reloadable3: 0_0_reloadable3 + mkdir -p 0_1_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_1_reloadable3/Release/0_1_reloadable3.o + set -o pipefail; (rm -f "0_1_reloadable3/Release/0_1_reloadable3.##") 2>&1 |& tee -a 0_1_reloadable3/0_1_reloadable3.log 0_1_reloadable3/timestamped_log/0_1_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable3/scripts/0_1_reloadable3.prx) 2>&1 |& tee -a 0_1_reloadable3/0_1_reloadable3.log 0_1_reloadable3/timestamped_log/0_1_reloadable3-${ts}.log + +0_1_reloadable9: 0_0_reloadable3 + mkdir -p 0_1_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_1_reloadable9/Release/0_1_reloadable9.o + set -o pipefail; (rm -f "0_1_reloadable9/Release/0_1_reloadable9.##") 2>&1 |& tee -a 0_1_reloadable9/0_1_reloadable9.log 0_1_reloadable9/timestamped_log/0_1_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable9/scripts/0_1_reloadable9.prx) 2>&1 |& tee -a 0_1_reloadable9/0_1_reloadable9.log 0_1_reloadable9/timestamped_log/0_1_reloadable9-${ts}.log + +0_1_reloadable13: 0_0_reloadable3 + mkdir -p 0_1_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_1_reloadable13/Release/0_1_reloadable13.o + set -o pipefail; (rm -f "0_1_reloadable13/Release/0_1_reloadable13.##") 2>&1 |& tee -a 0_1_reloadable13/0_1_reloadable13.log 0_1_reloadable13/timestamped_log/0_1_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable13/scripts/0_1_reloadable13.prx) 2>&1 |& tee -a 0_1_reloadable13/0_1_reloadable13.log 0_1_reloadable13/timestamped_log/0_1_reloadable13-${ts}.log + +0_2_reloadable3: 0_0_reloadable3 + mkdir -p 0_2_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_2_reloadable3/Release/0_2_reloadable3.o + set -o pipefail; (rm -f "0_2_reloadable3/Release/0_2_reloadable3.##") 2>&1 |& tee -a 0_2_reloadable3/0_2_reloadable3.log 0_2_reloadable3/timestamped_log/0_2_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable3/scripts/0_2_reloadable3.prx) 2>&1 |& tee -a 0_2_reloadable3/0_2_reloadable3.log 0_2_reloadable3/timestamped_log/0_2_reloadable3-${ts}.log + +0_2_reloadable9: 0_0_reloadable3 + mkdir -p 0_2_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_2_reloadable9/Release/0_2_reloadable9.o + set -o pipefail; (rm -f "0_2_reloadable9/Release/0_2_reloadable9.##") 2>&1 |& tee -a 0_2_reloadable9/0_2_reloadable9.log 0_2_reloadable9/timestamped_log/0_2_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable9/scripts/0_2_reloadable9.prx) 2>&1 |& tee -a 0_2_reloadable9/0_2_reloadable9.log 0_2_reloadable9/timestamped_log/0_2_reloadable9-${ts}.log + +0_2_reloadable13: 0_0_reloadable3 + mkdir -p 0_2_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_2_reloadable13/Release/0_2_reloadable13.o + set -o pipefail; (rm -f "0_2_reloadable13/Release/0_2_reloadable13.##") 2>&1 |& tee -a 0_2_reloadable13/0_2_reloadable13.log 0_2_reloadable13/timestamped_log/0_2_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable13/scripts/0_2_reloadable13.prx) 2>&1 |& tee -a 0_2_reloadable13/0_2_reloadable13.log 0_2_reloadable13/timestamped_log/0_2_reloadable13-${ts}.log + +0_3_reloadable3: 0_0_reloadable3 + mkdir -p 0_3_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_3_reloadable3/Release/0_3_reloadable3.o + set -o pipefail; (rm -f "0_3_reloadable3/Release/0_3_reloadable3.##") 2>&1 |& tee -a 0_3_reloadable3/0_3_reloadable3.log 0_3_reloadable3/timestamped_log/0_3_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable3/scripts/0_3_reloadable3.prx) 2>&1 |& tee -a 0_3_reloadable3/0_3_reloadable3.log 0_3_reloadable3/timestamped_log/0_3_reloadable3-${ts}.log + +0_3_reloadable9: 0_0_reloadable3 + mkdir -p 0_3_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_3_reloadable9/Release/0_3_reloadable9.o + set -o pipefail; (rm -f "0_3_reloadable9/Release/0_3_reloadable9.##") 2>&1 |& tee -a 0_3_reloadable9/0_3_reloadable9.log 0_3_reloadable9/timestamped_log/0_3_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable9/scripts/0_3_reloadable9.prx) 2>&1 |& tee -a 0_3_reloadable9/0_3_reloadable9.log 0_3_reloadable9/timestamped_log/0_3_reloadable9-${ts}.log + +0_3_reloadable13: 0_0_reloadable3 + mkdir -p 0_3_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 0_3_reloadable13/Release/0_3_reloadable13.o + set -o pipefail; (rm -f "0_3_reloadable13/Release/0_3_reloadable13.##") 2>&1 |& tee -a 0_3_reloadable13/0_3_reloadable13.log 0_3_reloadable13/timestamped_log/0_3_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable13/scripts/0_3_reloadable13.prx) 2>&1 |& tee -a 0_3_reloadable13/0_3_reloadable13.log 0_3_reloadable13/timestamped_log/0_3_reloadable13-${ts}.log + +1_0_reloadable3: 0_0_reloadable3 + mkdir -p 1_0_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_0_reloadable3/Release/1_0_reloadable3.o + set -o pipefail; (rm -f "1_0_reloadable3/Release/1_0_reloadable3.##") 2>&1 |& tee -a 1_0_reloadable3/1_0_reloadable3.log 1_0_reloadable3/timestamped_log/1_0_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable3/scripts/1_0_reloadable3.prx) 2>&1 |& tee -a 1_0_reloadable3/1_0_reloadable3.log 1_0_reloadable3/timestamped_log/1_0_reloadable3-${ts}.log + +1_0_reloadable9: 0_0_reloadable3 + mkdir -p 1_0_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_0_reloadable9/Release/1_0_reloadable9.o + set -o pipefail; (rm -f "1_0_reloadable9/Release/1_0_reloadable9.##") 2>&1 |& tee -a 1_0_reloadable9/1_0_reloadable9.log 1_0_reloadable9/timestamped_log/1_0_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable9/scripts/1_0_reloadable9.prx) 2>&1 |& tee -a 1_0_reloadable9/1_0_reloadable9.log 1_0_reloadable9/timestamped_log/1_0_reloadable9-${ts}.log + +1_0_reloadable13: 0_0_reloadable3 + mkdir -p 1_0_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_0_reloadable13/Release/1_0_reloadable13.o + set -o pipefail; (rm -f "1_0_reloadable13/Release/1_0_reloadable13.##") 2>&1 |& tee -a 1_0_reloadable13/1_0_reloadable13.log 1_0_reloadable13/timestamped_log/1_0_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable13/scripts/1_0_reloadable13.prx) 2>&1 |& tee -a 1_0_reloadable13/1_0_reloadable13.log 1_0_reloadable13/timestamped_log/1_0_reloadable13-${ts}.log + +1_1_reloadable3: 0_0_reloadable3 + mkdir -p 1_1_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_1_reloadable3/Release/1_1_reloadable3.o + set -o pipefail; (rm -f "1_1_reloadable3/Release/1_1_reloadable3.##") 2>&1 |& tee -a 1_1_reloadable3/1_1_reloadable3.log 1_1_reloadable3/timestamped_log/1_1_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable3/scripts/1_1_reloadable3.prx) 2>&1 |& tee -a 1_1_reloadable3/1_1_reloadable3.log 1_1_reloadable3/timestamped_log/1_1_reloadable3-${ts}.log + +1_1_reloadable9: 0_0_reloadable3 + mkdir -p 1_1_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_1_reloadable9/Release/1_1_reloadable9.o + set -o pipefail; (rm -f "1_1_reloadable9/Release/1_1_reloadable9.##") 2>&1 |& tee -a 1_1_reloadable9/1_1_reloadable9.log 1_1_reloadable9/timestamped_log/1_1_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable9/scripts/1_1_reloadable9.prx) 2>&1 |& tee -a 1_1_reloadable9/1_1_reloadable9.log 1_1_reloadable9/timestamped_log/1_1_reloadable9-${ts}.log + +1_1_reloadable13: 0_0_reloadable3 + mkdir -p 1_1_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_1_reloadable13/Release/1_1_reloadable13.o + set -o pipefail; (rm -f "1_1_reloadable13/Release/1_1_reloadable13.##") 2>&1 |& tee -a 1_1_reloadable13/1_1_reloadable13.log 1_1_reloadable13/timestamped_log/1_1_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable13/scripts/1_1_reloadable13.prx) 2>&1 |& tee -a 1_1_reloadable13/1_1_reloadable13.log 1_1_reloadable13/timestamped_log/1_1_reloadable13-${ts}.log + +1_2_reloadable3: 0_0_reloadable3 + mkdir -p 1_2_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_2_reloadable3/Release/1_2_reloadable3.o + set -o pipefail; (rm -f "1_2_reloadable3/Release/1_2_reloadable3.##") 2>&1 |& tee -a 1_2_reloadable3/1_2_reloadable3.log 1_2_reloadable3/timestamped_log/1_2_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable3/scripts/1_2_reloadable3.prx) 2>&1 |& tee -a 1_2_reloadable3/1_2_reloadable3.log 1_2_reloadable3/timestamped_log/1_2_reloadable3-${ts}.log + +1_2_reloadable9: 0_0_reloadable3 + mkdir -p 1_2_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_2_reloadable9/Release/1_2_reloadable9.o + set -o pipefail; (rm -f "1_2_reloadable9/Release/1_2_reloadable9.##") 2>&1 |& tee -a 1_2_reloadable9/1_2_reloadable9.log 1_2_reloadable9/timestamped_log/1_2_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable9/scripts/1_2_reloadable9.prx) 2>&1 |& tee -a 1_2_reloadable9/1_2_reloadable9.log 1_2_reloadable9/timestamped_log/1_2_reloadable9-${ts}.log + +1_2_reloadable13: 0_0_reloadable3 + mkdir -p 1_2_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_2_reloadable13/Release/1_2_reloadable13.o + set -o pipefail; (rm -f "1_2_reloadable13/Release/1_2_reloadable13.##") 2>&1 |& tee -a 1_2_reloadable13/1_2_reloadable13.log 1_2_reloadable13/timestamped_log/1_2_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable13/scripts/1_2_reloadable13.prx) 2>&1 |& tee -a 1_2_reloadable13/1_2_reloadable13.log 1_2_reloadable13/timestamped_log/1_2_reloadable13-${ts}.log + +1_3_reloadable3: 0_0_reloadable3 + mkdir -p 1_3_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_3_reloadable3/Release/1_3_reloadable3.o + set -o pipefail; (rm -f "1_3_reloadable3/Release/1_3_reloadable3.##") 2>&1 |& tee -a 1_3_reloadable3/1_3_reloadable3.log 1_3_reloadable3/timestamped_log/1_3_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable3/scripts/1_3_reloadable3.prx) 2>&1 |& tee -a 1_3_reloadable3/1_3_reloadable3.log 1_3_reloadable3/timestamped_log/1_3_reloadable3-${ts}.log + +1_3_reloadable9: 0_0_reloadable3 + mkdir -p 1_3_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_3_reloadable9/Release/1_3_reloadable9.o + set -o pipefail; (rm -f "1_3_reloadable9/Release/1_3_reloadable9.##") 2>&1 |& tee -a 1_3_reloadable9/1_3_reloadable9.log 1_3_reloadable9/timestamped_log/1_3_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable9/scripts/1_3_reloadable9.prx) 2>&1 |& tee -a 1_3_reloadable9/1_3_reloadable9.log 1_3_reloadable9/timestamped_log/1_3_reloadable9-${ts}.log + +1_3_reloadable13: 0_0_reloadable3 + mkdir -p 1_3_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 1_3_reloadable13/Release/1_3_reloadable13.o + set -o pipefail; (rm -f "1_3_reloadable13/Release/1_3_reloadable13.##") 2>&1 |& tee -a 1_3_reloadable13/1_3_reloadable13.log 1_3_reloadable13/timestamped_log/1_3_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable13/scripts/1_3_reloadable13.prx) 2>&1 |& tee -a 1_3_reloadable13/1_3_reloadable13.log 1_3_reloadable13/timestamped_log/1_3_reloadable13-${ts}.log + +2_0_reloadable3: 0_0_reloadable3 + mkdir -p 2_0_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_0_reloadable3/Release/2_0_reloadable3.o + set -o pipefail; (rm -f "2_0_reloadable3/Release/2_0_reloadable3.##") 2>&1 |& tee -a 2_0_reloadable3/2_0_reloadable3.log 2_0_reloadable3/timestamped_log/2_0_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable3/scripts/2_0_reloadable3.prx) 2>&1 |& tee -a 2_0_reloadable3/2_0_reloadable3.log 2_0_reloadable3/timestamped_log/2_0_reloadable3-${ts}.log + +2_0_reloadable9: 0_0_reloadable3 + mkdir -p 2_0_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_0_reloadable9/Release/2_0_reloadable9.o + set -o pipefail; (rm -f "2_0_reloadable9/Release/2_0_reloadable9.##") 2>&1 |& tee -a 2_0_reloadable9/2_0_reloadable9.log 2_0_reloadable9/timestamped_log/2_0_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable9/scripts/2_0_reloadable9.prx) 2>&1 |& tee -a 2_0_reloadable9/2_0_reloadable9.log 2_0_reloadable9/timestamped_log/2_0_reloadable9-${ts}.log + +2_0_reloadable13: 0_0_reloadable3 + mkdir -p 2_0_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_0_reloadable13/Release/2_0_reloadable13.o + set -o pipefail; (rm -f "2_0_reloadable13/Release/2_0_reloadable13.##") 2>&1 |& tee -a 2_0_reloadable13/2_0_reloadable13.log 2_0_reloadable13/timestamped_log/2_0_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable13/scripts/2_0_reloadable13.prx) 2>&1 |& tee -a 2_0_reloadable13/2_0_reloadable13.log 2_0_reloadable13/timestamped_log/2_0_reloadable13-${ts}.log + +2_1_reloadable3: 0_0_reloadable3 + mkdir -p 2_1_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_1_reloadable3/Release/2_1_reloadable3.o + set -o pipefail; (rm -f "2_1_reloadable3/Release/2_1_reloadable3.##") 2>&1 |& tee -a 2_1_reloadable3/2_1_reloadable3.log 2_1_reloadable3/timestamped_log/2_1_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable3/scripts/2_1_reloadable3.prx) 2>&1 |& tee -a 2_1_reloadable3/2_1_reloadable3.log 2_1_reloadable3/timestamped_log/2_1_reloadable3-${ts}.log + +2_1_reloadable9: 0_0_reloadable3 + mkdir -p 2_1_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_1_reloadable9/Release/2_1_reloadable9.o + set -o pipefail; (rm -f "2_1_reloadable9/Release/2_1_reloadable9.##") 2>&1 |& tee -a 2_1_reloadable9/2_1_reloadable9.log 2_1_reloadable9/timestamped_log/2_1_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable9/scripts/2_1_reloadable9.prx) 2>&1 |& tee -a 2_1_reloadable9/2_1_reloadable9.log 2_1_reloadable9/timestamped_log/2_1_reloadable9-${ts}.log + +2_1_reloadable13: 0_0_reloadable3 + mkdir -p 2_1_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_1_reloadable13/Release/2_1_reloadable13.o + set -o pipefail; (rm -f "2_1_reloadable13/Release/2_1_reloadable13.##") 2>&1 |& tee -a 2_1_reloadable13/2_1_reloadable13.log 2_1_reloadable13/timestamped_log/2_1_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable13/scripts/2_1_reloadable13.prx) 2>&1 |& tee -a 2_1_reloadable13/2_1_reloadable13.log 2_1_reloadable13/timestamped_log/2_1_reloadable13-${ts}.log + +2_2_reloadable3: 0_0_reloadable3 + mkdir -p 2_2_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_2_reloadable3/Release/2_2_reloadable3.o + set -o pipefail; (rm -f "2_2_reloadable3/Release/2_2_reloadable3.##") 2>&1 |& tee -a 2_2_reloadable3/2_2_reloadable3.log 2_2_reloadable3/timestamped_log/2_2_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable3/scripts/2_2_reloadable3.prx) 2>&1 |& tee -a 2_2_reloadable3/2_2_reloadable3.log 2_2_reloadable3/timestamped_log/2_2_reloadable3-${ts}.log + +2_2_reloadable9: 0_0_reloadable3 + mkdir -p 2_2_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_2_reloadable9/Release/2_2_reloadable9.o + set -o pipefail; (rm -f "2_2_reloadable9/Release/2_2_reloadable9.##") 2>&1 |& tee -a 2_2_reloadable9/2_2_reloadable9.log 2_2_reloadable9/timestamped_log/2_2_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable9/scripts/2_2_reloadable9.prx) 2>&1 |& tee -a 2_2_reloadable9/2_2_reloadable9.log 2_2_reloadable9/timestamped_log/2_2_reloadable9-${ts}.log + +2_2_reloadable13: 0_0_reloadable3 + mkdir -p 2_2_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_2_reloadable13/Release/2_2_reloadable13.o + set -o pipefail; (rm -f "2_2_reloadable13/Release/2_2_reloadable13.##") 2>&1 |& tee -a 2_2_reloadable13/2_2_reloadable13.log 2_2_reloadable13/timestamped_log/2_2_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable13/scripts/2_2_reloadable13.prx) 2>&1 |& tee -a 2_2_reloadable13/2_2_reloadable13.log 2_2_reloadable13/timestamped_log/2_2_reloadable13-${ts}.log + +2_3_reloadable3: 0_0_reloadable3 + mkdir -p 2_3_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_3_reloadable3/Release/2_3_reloadable3.o + set -o pipefail; (rm -f "2_3_reloadable3/Release/2_3_reloadable3.##") 2>&1 |& tee -a 2_3_reloadable3/2_3_reloadable3.log 2_3_reloadable3/timestamped_log/2_3_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable3/scripts/2_3_reloadable3.prx) 2>&1 |& tee -a 2_3_reloadable3/2_3_reloadable3.log 2_3_reloadable3/timestamped_log/2_3_reloadable3-${ts}.log + +2_3_reloadable9: 0_0_reloadable3 + mkdir -p 2_3_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_3_reloadable9/Release/2_3_reloadable9.o + set -o pipefail; (rm -f "2_3_reloadable9/Release/2_3_reloadable9.##") 2>&1 |& tee -a 2_3_reloadable9/2_3_reloadable9.log 2_3_reloadable9/timestamped_log/2_3_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable9/scripts/2_3_reloadable9.prx) 2>&1 |& tee -a 2_3_reloadable9/2_3_reloadable9.log 2_3_reloadable9/timestamped_log/2_3_reloadable9-${ts}.log + +2_3_reloadable13: 0_0_reloadable3 + mkdir -p 2_3_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 2_3_reloadable13/Release/2_3_reloadable13.o + set -o pipefail; (rm -f "2_3_reloadable13/Release/2_3_reloadable13.##") 2>&1 |& tee -a 2_3_reloadable13/2_3_reloadable13.log 2_3_reloadable13/timestamped_log/2_3_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable13/scripts/2_3_reloadable13.prx) 2>&1 |& tee -a 2_3_reloadable13/2_3_reloadable13.log 2_3_reloadable13/timestamped_log/2_3_reloadable13-${ts}.log + +3_0_reloadable3: 0_0_reloadable3 + mkdir -p 3_0_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_0_reloadable3/Release/3_0_reloadable3.o + set -o pipefail; (rm -f "3_0_reloadable3/Release/3_0_reloadable3.##") 2>&1 |& tee -a 3_0_reloadable3/3_0_reloadable3.log 3_0_reloadable3/timestamped_log/3_0_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable3/scripts/3_0_reloadable3.prx) 2>&1 |& tee -a 3_0_reloadable3/3_0_reloadable3.log 3_0_reloadable3/timestamped_log/3_0_reloadable3-${ts}.log + +3_0_reloadable9: 0_0_reloadable3 + mkdir -p 3_0_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_0_reloadable9/Release/3_0_reloadable9.o + set -o pipefail; (rm -f "3_0_reloadable9/Release/3_0_reloadable9.##") 2>&1 |& tee -a 3_0_reloadable9/3_0_reloadable9.log 3_0_reloadable9/timestamped_log/3_0_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable9/scripts/3_0_reloadable9.prx) 2>&1 |& tee -a 3_0_reloadable9/3_0_reloadable9.log 3_0_reloadable9/timestamped_log/3_0_reloadable9-${ts}.log + +3_0_reloadable13: 0_0_reloadable3 + mkdir -p 3_0_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_0_reloadable13/Release/3_0_reloadable13.o + set -o pipefail; (rm -f "3_0_reloadable13/Release/3_0_reloadable13.##") 2>&1 |& tee -a 3_0_reloadable13/3_0_reloadable13.log 3_0_reloadable13/timestamped_log/3_0_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable13/scripts/3_0_reloadable13.prx) 2>&1 |& tee -a 3_0_reloadable13/3_0_reloadable13.log 3_0_reloadable13/timestamped_log/3_0_reloadable13-${ts}.log + +3_1_reloadable3: 0_0_reloadable3 + mkdir -p 3_1_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_1_reloadable3/Release/3_1_reloadable3.o + set -o pipefail; (rm -f "3_1_reloadable3/Release/3_1_reloadable3.##") 2>&1 |& tee -a 3_1_reloadable3/3_1_reloadable3.log 3_1_reloadable3/timestamped_log/3_1_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable3/scripts/3_1_reloadable3.prx) 2>&1 |& tee -a 3_1_reloadable3/3_1_reloadable3.log 3_1_reloadable3/timestamped_log/3_1_reloadable3-${ts}.log + +3_1_reloadable9: 0_0_reloadable3 + mkdir -p 3_1_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_1_reloadable9/Release/3_1_reloadable9.o + set -o pipefail; (rm -f "3_1_reloadable9/Release/3_1_reloadable9.##") 2>&1 |& tee -a 3_1_reloadable9/3_1_reloadable9.log 3_1_reloadable9/timestamped_log/3_1_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable9/scripts/3_1_reloadable9.prx) 2>&1 |& tee -a 3_1_reloadable9/3_1_reloadable9.log 3_1_reloadable9/timestamped_log/3_1_reloadable9-${ts}.log + +3_1_reloadable13: 0_0_reloadable3 + mkdir -p 3_1_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_1_reloadable13/Release/3_1_reloadable13.o + set -o pipefail; (rm -f "3_1_reloadable13/Release/3_1_reloadable13.##") 2>&1 |& tee -a 3_1_reloadable13/3_1_reloadable13.log 3_1_reloadable13/timestamped_log/3_1_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable13/scripts/3_1_reloadable13.prx) 2>&1 |& tee -a 3_1_reloadable13/3_1_reloadable13.log 3_1_reloadable13/timestamped_log/3_1_reloadable13-${ts}.log + +3_2_reloadable3: 0_0_reloadable3 + mkdir -p 3_2_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_2_reloadable3/Release/3_2_reloadable3.o + set -o pipefail; (rm -f "3_2_reloadable3/Release/3_2_reloadable3.##") 2>&1 |& tee -a 3_2_reloadable3/3_2_reloadable3.log 3_2_reloadable3/timestamped_log/3_2_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable3/scripts/3_2_reloadable3.prx) 2>&1 |& tee -a 3_2_reloadable3/3_2_reloadable3.log 3_2_reloadable3/timestamped_log/3_2_reloadable3-${ts}.log + +3_2_reloadable9: 0_0_reloadable3 + mkdir -p 3_2_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_2_reloadable9/Release/3_2_reloadable9.o + set -o pipefail; (rm -f "3_2_reloadable9/Release/3_2_reloadable9.##") 2>&1 |& tee -a 3_2_reloadable9/3_2_reloadable9.log 3_2_reloadable9/timestamped_log/3_2_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable9/scripts/3_2_reloadable9.prx) 2>&1 |& tee -a 3_2_reloadable9/3_2_reloadable9.log 3_2_reloadable9/timestamped_log/3_2_reloadable9-${ts}.log + +3_2_reloadable13: 0_0_reloadable3 + mkdir -p 3_2_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_2_reloadable13/Release/3_2_reloadable13.o + set -o pipefail; (rm -f "3_2_reloadable13/Release/3_2_reloadable13.##") 2>&1 |& tee -a 3_2_reloadable13/3_2_reloadable13.log 3_2_reloadable13/timestamped_log/3_2_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable13/scripts/3_2_reloadable13.prx) 2>&1 |& tee -a 3_2_reloadable13/3_2_reloadable13.log 3_2_reloadable13/timestamped_log/3_2_reloadable13-${ts}.log + +3_3_reloadable3: 0_0_reloadable3 + mkdir -p 3_3_reloadable3/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_3_reloadable3/Release/3_3_reloadable3.o + set -o pipefail; (rm -f "3_3_reloadable3/Release/3_3_reloadable3.##") 2>&1 |& tee -a 3_3_reloadable3/3_3_reloadable3.log 3_3_reloadable3/timestamped_log/3_3_reloadable3-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable3/scripts/3_3_reloadable3.prx) 2>&1 |& tee -a 3_3_reloadable3/3_3_reloadable3.log 3_3_reloadable3/timestamped_log/3_3_reloadable3-${ts}.log + +3_3_reloadable9: 0_0_reloadable3 + mkdir -p 3_3_reloadable9/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_3_reloadable9/Release/3_3_reloadable9.o + set -o pipefail; (rm -f "3_3_reloadable9/Release/3_3_reloadable9.##") 2>&1 |& tee -a 3_3_reloadable9/3_3_reloadable9.log 3_3_reloadable9/timestamped_log/3_3_reloadable9-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable9/scripts/3_3_reloadable9.prx) 2>&1 |& tee -a 3_3_reloadable9/3_3_reloadable9.log 3_3_reloadable9/timestamped_log/3_3_reloadable9-${ts}.log + +3_3_reloadable13: 0_0_reloadable3 + mkdir -p 3_3_reloadable13/Release + cp 0_0_reloadable3/Release/0_0_reloadable3.o 3_3_reloadable13/Release/3_3_reloadable13.o + set -o pipefail; (rm -f "3_3_reloadable13/Release/3_3_reloadable13.##") 2>&1 |& tee -a 3_3_reloadable13/3_3_reloadable13.log 3_3_reloadable13/timestamped_log/3_3_reloadable13-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable13/scripts/3_3_reloadable13.prx) 2>&1 |& tee -a 3_3_reloadable13/3_3_reloadable13.log 3_3_reloadable13/timestamped_log/3_3_reloadable13-${ts}.log + +0_0_reloadable7: 0_0_reloadable5 + mkdir -p 0_0_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_0_reloadable7/Release/0_0_reloadable7.o + set -o pipefail; (rm -f "0_0_reloadable7/Release/0_0_reloadable7.##") 2>&1 |& tee -a 0_0_reloadable7/0_0_reloadable7.log 0_0_reloadable7/timestamped_log/0_0_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable7/scripts/0_0_reloadable7.prx) 2>&1 |& tee -a 0_0_reloadable7/0_0_reloadable7.log 0_0_reloadable7/timestamped_log/0_0_reloadable7-${ts}.log + +0_0_reloadable11: 0_0_reloadable5 + mkdir -p 0_0_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_0_reloadable11/Release/0_0_reloadable11.o + set -o pipefail; (rm -f "0_0_reloadable11/Release/0_0_reloadable11.##") 2>&1 |& tee -a 0_0_reloadable11/0_0_reloadable11.log 0_0_reloadable11/timestamped_log/0_0_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable11/scripts/0_0_reloadable11.prx) 2>&1 |& tee -a 0_0_reloadable11/0_0_reloadable11.log 0_0_reloadable11/timestamped_log/0_0_reloadable11-${ts}.log + +0_0_reloadable15: 0_0_reloadable5 + mkdir -p 0_0_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_0_reloadable15/Release/0_0_reloadable15.o + set -o pipefail; (rm -f "0_0_reloadable15/Release/0_0_reloadable15.##") 2>&1 |& tee -a 0_0_reloadable15/0_0_reloadable15.log 0_0_reloadable15/timestamped_log/0_0_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_0_reloadable15/scripts/0_0_reloadable15.prx) 2>&1 |& tee -a 0_0_reloadable15/0_0_reloadable15.log 0_0_reloadable15/timestamped_log/0_0_reloadable15-${ts}.log + +0_1_reloadable5: 0_0_reloadable5 + mkdir -p 0_1_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_1_reloadable5/Release/0_1_reloadable5.o + set -o pipefail; (rm -f "0_1_reloadable5/Release/0_1_reloadable5.##") 2>&1 |& tee -a 0_1_reloadable5/0_1_reloadable5.log 0_1_reloadable5/timestamped_log/0_1_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable5/scripts/0_1_reloadable5.prx) 2>&1 |& tee -a 0_1_reloadable5/0_1_reloadable5.log 0_1_reloadable5/timestamped_log/0_1_reloadable5-${ts}.log + +0_1_reloadable7: 0_0_reloadable5 + mkdir -p 0_1_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_1_reloadable7/Release/0_1_reloadable7.o + set -o pipefail; (rm -f "0_1_reloadable7/Release/0_1_reloadable7.##") 2>&1 |& tee -a 0_1_reloadable7/0_1_reloadable7.log 0_1_reloadable7/timestamped_log/0_1_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable7/scripts/0_1_reloadable7.prx) 2>&1 |& tee -a 0_1_reloadable7/0_1_reloadable7.log 0_1_reloadable7/timestamped_log/0_1_reloadable7-${ts}.log + +0_1_reloadable11: 0_0_reloadable5 + mkdir -p 0_1_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_1_reloadable11/Release/0_1_reloadable11.o + set -o pipefail; (rm -f "0_1_reloadable11/Release/0_1_reloadable11.##") 2>&1 |& tee -a 0_1_reloadable11/0_1_reloadable11.log 0_1_reloadable11/timestamped_log/0_1_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable11/scripts/0_1_reloadable11.prx) 2>&1 |& tee -a 0_1_reloadable11/0_1_reloadable11.log 0_1_reloadable11/timestamped_log/0_1_reloadable11-${ts}.log + +0_1_reloadable15: 0_0_reloadable5 + mkdir -p 0_1_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_1_reloadable15/Release/0_1_reloadable15.o + set -o pipefail; (rm -f "0_1_reloadable15/Release/0_1_reloadable15.##") 2>&1 |& tee -a 0_1_reloadable15/0_1_reloadable15.log 0_1_reloadable15/timestamped_log/0_1_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable15/scripts/0_1_reloadable15.prx) 2>&1 |& tee -a 0_1_reloadable15/0_1_reloadable15.log 0_1_reloadable15/timestamped_log/0_1_reloadable15-${ts}.log + +0_2_reloadable5: 0_0_reloadable5 + mkdir -p 0_2_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_2_reloadable5/Release/0_2_reloadable5.o + set -o pipefail; (rm -f "0_2_reloadable5/Release/0_2_reloadable5.##") 2>&1 |& tee -a 0_2_reloadable5/0_2_reloadable5.log 0_2_reloadable5/timestamped_log/0_2_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable5/scripts/0_2_reloadable5.prx) 2>&1 |& tee -a 0_2_reloadable5/0_2_reloadable5.log 0_2_reloadable5/timestamped_log/0_2_reloadable5-${ts}.log + +0_2_reloadable7: 0_0_reloadable5 + mkdir -p 0_2_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_2_reloadable7/Release/0_2_reloadable7.o + set -o pipefail; (rm -f "0_2_reloadable7/Release/0_2_reloadable7.##") 2>&1 |& tee -a 0_2_reloadable7/0_2_reloadable7.log 0_2_reloadable7/timestamped_log/0_2_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable7/scripts/0_2_reloadable7.prx) 2>&1 |& tee -a 0_2_reloadable7/0_2_reloadable7.log 0_2_reloadable7/timestamped_log/0_2_reloadable7-${ts}.log + +0_2_reloadable11: 0_0_reloadable5 + mkdir -p 0_2_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_2_reloadable11/Release/0_2_reloadable11.o + set -o pipefail; (rm -f "0_2_reloadable11/Release/0_2_reloadable11.##") 2>&1 |& tee -a 0_2_reloadable11/0_2_reloadable11.log 0_2_reloadable11/timestamped_log/0_2_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable11/scripts/0_2_reloadable11.prx) 2>&1 |& tee -a 0_2_reloadable11/0_2_reloadable11.log 0_2_reloadable11/timestamped_log/0_2_reloadable11-${ts}.log + +0_2_reloadable15: 0_0_reloadable5 + mkdir -p 0_2_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_2_reloadable15/Release/0_2_reloadable15.o + set -o pipefail; (rm -f "0_2_reloadable15/Release/0_2_reloadable15.##") 2>&1 |& tee -a 0_2_reloadable15/0_2_reloadable15.log 0_2_reloadable15/timestamped_log/0_2_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable15/scripts/0_2_reloadable15.prx) 2>&1 |& tee -a 0_2_reloadable15/0_2_reloadable15.log 0_2_reloadable15/timestamped_log/0_2_reloadable15-${ts}.log + +0_3_reloadable5: 0_0_reloadable5 + mkdir -p 0_3_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_3_reloadable5/Release/0_3_reloadable5.o + set -o pipefail; (rm -f "0_3_reloadable5/Release/0_3_reloadable5.##") 2>&1 |& tee -a 0_3_reloadable5/0_3_reloadable5.log 0_3_reloadable5/timestamped_log/0_3_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable5/scripts/0_3_reloadable5.prx) 2>&1 |& tee -a 0_3_reloadable5/0_3_reloadable5.log 0_3_reloadable5/timestamped_log/0_3_reloadable5-${ts}.log + +0_3_reloadable7: 0_0_reloadable5 + mkdir -p 0_3_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_3_reloadable7/Release/0_3_reloadable7.o + set -o pipefail; (rm -f "0_3_reloadable7/Release/0_3_reloadable7.##") 2>&1 |& tee -a 0_3_reloadable7/0_3_reloadable7.log 0_3_reloadable7/timestamped_log/0_3_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable7/scripts/0_3_reloadable7.prx) 2>&1 |& tee -a 0_3_reloadable7/0_3_reloadable7.log 0_3_reloadable7/timestamped_log/0_3_reloadable7-${ts}.log + +0_3_reloadable11: 0_0_reloadable5 + mkdir -p 0_3_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_3_reloadable11/Release/0_3_reloadable11.o + set -o pipefail; (rm -f "0_3_reloadable11/Release/0_3_reloadable11.##") 2>&1 |& tee -a 0_3_reloadable11/0_3_reloadable11.log 0_3_reloadable11/timestamped_log/0_3_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable11/scripts/0_3_reloadable11.prx) 2>&1 |& tee -a 0_3_reloadable11/0_3_reloadable11.log 0_3_reloadable11/timestamped_log/0_3_reloadable11-${ts}.log + +0_3_reloadable15: 0_0_reloadable5 + mkdir -p 0_3_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 0_3_reloadable15/Release/0_3_reloadable15.o + set -o pipefail; (rm -f "0_3_reloadable15/Release/0_3_reloadable15.##") 2>&1 |& tee -a 0_3_reloadable15/0_3_reloadable15.log 0_3_reloadable15/timestamped_log/0_3_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable15/scripts/0_3_reloadable15.prx) 2>&1 |& tee -a 0_3_reloadable15/0_3_reloadable15.log 0_3_reloadable15/timestamped_log/0_3_reloadable15-${ts}.log + +1_0_reloadable5: 0_0_reloadable5 + mkdir -p 1_0_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_0_reloadable5/Release/1_0_reloadable5.o + set -o pipefail; (rm -f "1_0_reloadable5/Release/1_0_reloadable5.##") 2>&1 |& tee -a 1_0_reloadable5/1_0_reloadable5.log 1_0_reloadable5/timestamped_log/1_0_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable5/scripts/1_0_reloadable5.prx) 2>&1 |& tee -a 1_0_reloadable5/1_0_reloadable5.log 1_0_reloadable5/timestamped_log/1_0_reloadable5-${ts}.log + +1_0_reloadable7: 0_0_reloadable5 + mkdir -p 1_0_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_0_reloadable7/Release/1_0_reloadable7.o + set -o pipefail; (rm -f "1_0_reloadable7/Release/1_0_reloadable7.##") 2>&1 |& tee -a 1_0_reloadable7/1_0_reloadable7.log 1_0_reloadable7/timestamped_log/1_0_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable7/scripts/1_0_reloadable7.prx) 2>&1 |& tee -a 1_0_reloadable7/1_0_reloadable7.log 1_0_reloadable7/timestamped_log/1_0_reloadable7-${ts}.log + +1_0_reloadable11: 0_0_reloadable5 + mkdir -p 1_0_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_0_reloadable11/Release/1_0_reloadable11.o + set -o pipefail; (rm -f "1_0_reloadable11/Release/1_0_reloadable11.##") 2>&1 |& tee -a 1_0_reloadable11/1_0_reloadable11.log 1_0_reloadable11/timestamped_log/1_0_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable11/scripts/1_0_reloadable11.prx) 2>&1 |& tee -a 1_0_reloadable11/1_0_reloadable11.log 1_0_reloadable11/timestamped_log/1_0_reloadable11-${ts}.log + +1_0_reloadable15: 0_0_reloadable5 + mkdir -p 1_0_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_0_reloadable15/Release/1_0_reloadable15.o + set -o pipefail; (rm -f "1_0_reloadable15/Release/1_0_reloadable15.##") 2>&1 |& tee -a 1_0_reloadable15/1_0_reloadable15.log 1_0_reloadable15/timestamped_log/1_0_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable15/scripts/1_0_reloadable15.prx) 2>&1 |& tee -a 1_0_reloadable15/1_0_reloadable15.log 1_0_reloadable15/timestamped_log/1_0_reloadable15-${ts}.log + +1_1_reloadable5: 0_0_reloadable5 + mkdir -p 1_1_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_1_reloadable5/Release/1_1_reloadable5.o + set -o pipefail; (rm -f "1_1_reloadable5/Release/1_1_reloadable5.##") 2>&1 |& tee -a 1_1_reloadable5/1_1_reloadable5.log 1_1_reloadable5/timestamped_log/1_1_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable5/scripts/1_1_reloadable5.prx) 2>&1 |& tee -a 1_1_reloadable5/1_1_reloadable5.log 1_1_reloadable5/timestamped_log/1_1_reloadable5-${ts}.log + +1_1_reloadable7: 0_0_reloadable5 + mkdir -p 1_1_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_1_reloadable7/Release/1_1_reloadable7.o + set -o pipefail; (rm -f "1_1_reloadable7/Release/1_1_reloadable7.##") 2>&1 |& tee -a 1_1_reloadable7/1_1_reloadable7.log 1_1_reloadable7/timestamped_log/1_1_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable7/scripts/1_1_reloadable7.prx) 2>&1 |& tee -a 1_1_reloadable7/1_1_reloadable7.log 1_1_reloadable7/timestamped_log/1_1_reloadable7-${ts}.log + +1_1_reloadable11: 0_0_reloadable5 + mkdir -p 1_1_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_1_reloadable11/Release/1_1_reloadable11.o + set -o pipefail; (rm -f "1_1_reloadable11/Release/1_1_reloadable11.##") 2>&1 |& tee -a 1_1_reloadable11/1_1_reloadable11.log 1_1_reloadable11/timestamped_log/1_1_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable11/scripts/1_1_reloadable11.prx) 2>&1 |& tee -a 1_1_reloadable11/1_1_reloadable11.log 1_1_reloadable11/timestamped_log/1_1_reloadable11-${ts}.log + +1_1_reloadable15: 0_0_reloadable5 + mkdir -p 1_1_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_1_reloadable15/Release/1_1_reloadable15.o + set -o pipefail; (rm -f "1_1_reloadable15/Release/1_1_reloadable15.##") 2>&1 |& tee -a 1_1_reloadable15/1_1_reloadable15.log 1_1_reloadable15/timestamped_log/1_1_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable15/scripts/1_1_reloadable15.prx) 2>&1 |& tee -a 1_1_reloadable15/1_1_reloadable15.log 1_1_reloadable15/timestamped_log/1_1_reloadable15-${ts}.log + +1_2_reloadable5: 0_0_reloadable5 + mkdir -p 1_2_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_2_reloadable5/Release/1_2_reloadable5.o + set -o pipefail; (rm -f "1_2_reloadable5/Release/1_2_reloadable5.##") 2>&1 |& tee -a 1_2_reloadable5/1_2_reloadable5.log 1_2_reloadable5/timestamped_log/1_2_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable5/scripts/1_2_reloadable5.prx) 2>&1 |& tee -a 1_2_reloadable5/1_2_reloadable5.log 1_2_reloadable5/timestamped_log/1_2_reloadable5-${ts}.log + +1_2_reloadable7: 0_0_reloadable5 + mkdir -p 1_2_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_2_reloadable7/Release/1_2_reloadable7.o + set -o pipefail; (rm -f "1_2_reloadable7/Release/1_2_reloadable7.##") 2>&1 |& tee -a 1_2_reloadable7/1_2_reloadable7.log 1_2_reloadable7/timestamped_log/1_2_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable7/scripts/1_2_reloadable7.prx) 2>&1 |& tee -a 1_2_reloadable7/1_2_reloadable7.log 1_2_reloadable7/timestamped_log/1_2_reloadable7-${ts}.log + +1_2_reloadable11: 0_0_reloadable5 + mkdir -p 1_2_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_2_reloadable11/Release/1_2_reloadable11.o + set -o pipefail; (rm -f "1_2_reloadable11/Release/1_2_reloadable11.##") 2>&1 |& tee -a 1_2_reloadable11/1_2_reloadable11.log 1_2_reloadable11/timestamped_log/1_2_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable11/scripts/1_2_reloadable11.prx) 2>&1 |& tee -a 1_2_reloadable11/1_2_reloadable11.log 1_2_reloadable11/timestamped_log/1_2_reloadable11-${ts}.log + +1_2_reloadable15: 0_0_reloadable5 + mkdir -p 1_2_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_2_reloadable15/Release/1_2_reloadable15.o + set -o pipefail; (rm -f "1_2_reloadable15/Release/1_2_reloadable15.##") 2>&1 |& tee -a 1_2_reloadable15/1_2_reloadable15.log 1_2_reloadable15/timestamped_log/1_2_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable15/scripts/1_2_reloadable15.prx) 2>&1 |& tee -a 1_2_reloadable15/1_2_reloadable15.log 1_2_reloadable15/timestamped_log/1_2_reloadable15-${ts}.log + +1_3_reloadable5: 0_0_reloadable5 + mkdir -p 1_3_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_3_reloadable5/Release/1_3_reloadable5.o + set -o pipefail; (rm -f "1_3_reloadable5/Release/1_3_reloadable5.##") 2>&1 |& tee -a 1_3_reloadable5/1_3_reloadable5.log 1_3_reloadable5/timestamped_log/1_3_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable5/scripts/1_3_reloadable5.prx) 2>&1 |& tee -a 1_3_reloadable5/1_3_reloadable5.log 1_3_reloadable5/timestamped_log/1_3_reloadable5-${ts}.log + +1_3_reloadable7: 0_0_reloadable5 + mkdir -p 1_3_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_3_reloadable7/Release/1_3_reloadable7.o + set -o pipefail; (rm -f "1_3_reloadable7/Release/1_3_reloadable7.##") 2>&1 |& tee -a 1_3_reloadable7/1_3_reloadable7.log 1_3_reloadable7/timestamped_log/1_3_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable7/scripts/1_3_reloadable7.prx) 2>&1 |& tee -a 1_3_reloadable7/1_3_reloadable7.log 1_3_reloadable7/timestamped_log/1_3_reloadable7-${ts}.log + +1_3_reloadable11: 0_0_reloadable5 + mkdir -p 1_3_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_3_reloadable11/Release/1_3_reloadable11.o + set -o pipefail; (rm -f "1_3_reloadable11/Release/1_3_reloadable11.##") 2>&1 |& tee -a 1_3_reloadable11/1_3_reloadable11.log 1_3_reloadable11/timestamped_log/1_3_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable11/scripts/1_3_reloadable11.prx) 2>&1 |& tee -a 1_3_reloadable11/1_3_reloadable11.log 1_3_reloadable11/timestamped_log/1_3_reloadable11-${ts}.log + +1_3_reloadable15: 0_0_reloadable5 + mkdir -p 1_3_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 1_3_reloadable15/Release/1_3_reloadable15.o + set -o pipefail; (rm -f "1_3_reloadable15/Release/1_3_reloadable15.##") 2>&1 |& tee -a 1_3_reloadable15/1_3_reloadable15.log 1_3_reloadable15/timestamped_log/1_3_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable15/scripts/1_3_reloadable15.prx) 2>&1 |& tee -a 1_3_reloadable15/1_3_reloadable15.log 1_3_reloadable15/timestamped_log/1_3_reloadable15-${ts}.log + +2_0_reloadable5: 0_0_reloadable5 + mkdir -p 2_0_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_0_reloadable5/Release/2_0_reloadable5.o + set -o pipefail; (rm -f "2_0_reloadable5/Release/2_0_reloadable5.##") 2>&1 |& tee -a 2_0_reloadable5/2_0_reloadable5.log 2_0_reloadable5/timestamped_log/2_0_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable5/scripts/2_0_reloadable5.prx) 2>&1 |& tee -a 2_0_reloadable5/2_0_reloadable5.log 2_0_reloadable5/timestamped_log/2_0_reloadable5-${ts}.log + +2_0_reloadable7: 0_0_reloadable5 + mkdir -p 2_0_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_0_reloadable7/Release/2_0_reloadable7.o + set -o pipefail; (rm -f "2_0_reloadable7/Release/2_0_reloadable7.##") 2>&1 |& tee -a 2_0_reloadable7/2_0_reloadable7.log 2_0_reloadable7/timestamped_log/2_0_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable7/scripts/2_0_reloadable7.prx) 2>&1 |& tee -a 2_0_reloadable7/2_0_reloadable7.log 2_0_reloadable7/timestamped_log/2_0_reloadable7-${ts}.log + +2_0_reloadable11: 0_0_reloadable5 + mkdir -p 2_0_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_0_reloadable11/Release/2_0_reloadable11.o + set -o pipefail; (rm -f "2_0_reloadable11/Release/2_0_reloadable11.##") 2>&1 |& tee -a 2_0_reloadable11/2_0_reloadable11.log 2_0_reloadable11/timestamped_log/2_0_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable11/scripts/2_0_reloadable11.prx) 2>&1 |& tee -a 2_0_reloadable11/2_0_reloadable11.log 2_0_reloadable11/timestamped_log/2_0_reloadable11-${ts}.log + +2_0_reloadable15: 0_0_reloadable5 + mkdir -p 2_0_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_0_reloadable15/Release/2_0_reloadable15.o + set -o pipefail; (rm -f "2_0_reloadable15/Release/2_0_reloadable15.##") 2>&1 |& tee -a 2_0_reloadable15/2_0_reloadable15.log 2_0_reloadable15/timestamped_log/2_0_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable15/scripts/2_0_reloadable15.prx) 2>&1 |& tee -a 2_0_reloadable15/2_0_reloadable15.log 2_0_reloadable15/timestamped_log/2_0_reloadable15-${ts}.log + +2_1_reloadable5: 0_0_reloadable5 + mkdir -p 2_1_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_1_reloadable5/Release/2_1_reloadable5.o + set -o pipefail; (rm -f "2_1_reloadable5/Release/2_1_reloadable5.##") 2>&1 |& tee -a 2_1_reloadable5/2_1_reloadable5.log 2_1_reloadable5/timestamped_log/2_1_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable5/scripts/2_1_reloadable5.prx) 2>&1 |& tee -a 2_1_reloadable5/2_1_reloadable5.log 2_1_reloadable5/timestamped_log/2_1_reloadable5-${ts}.log + +2_1_reloadable7: 0_0_reloadable5 + mkdir -p 2_1_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_1_reloadable7/Release/2_1_reloadable7.o + set -o pipefail; (rm -f "2_1_reloadable7/Release/2_1_reloadable7.##") 2>&1 |& tee -a 2_1_reloadable7/2_1_reloadable7.log 2_1_reloadable7/timestamped_log/2_1_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable7/scripts/2_1_reloadable7.prx) 2>&1 |& tee -a 2_1_reloadable7/2_1_reloadable7.log 2_1_reloadable7/timestamped_log/2_1_reloadable7-${ts}.log + +2_1_reloadable11: 0_0_reloadable5 + mkdir -p 2_1_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_1_reloadable11/Release/2_1_reloadable11.o + set -o pipefail; (rm -f "2_1_reloadable11/Release/2_1_reloadable11.##") 2>&1 |& tee -a 2_1_reloadable11/2_1_reloadable11.log 2_1_reloadable11/timestamped_log/2_1_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable11/scripts/2_1_reloadable11.prx) 2>&1 |& tee -a 2_1_reloadable11/2_1_reloadable11.log 2_1_reloadable11/timestamped_log/2_1_reloadable11-${ts}.log + +2_1_reloadable15: 0_0_reloadable5 + mkdir -p 2_1_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_1_reloadable15/Release/2_1_reloadable15.o + set -o pipefail; (rm -f "2_1_reloadable15/Release/2_1_reloadable15.##") 2>&1 |& tee -a 2_1_reloadable15/2_1_reloadable15.log 2_1_reloadable15/timestamped_log/2_1_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable15/scripts/2_1_reloadable15.prx) 2>&1 |& tee -a 2_1_reloadable15/2_1_reloadable15.log 2_1_reloadable15/timestamped_log/2_1_reloadable15-${ts}.log + +2_2_reloadable5: 0_0_reloadable5 + mkdir -p 2_2_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_2_reloadable5/Release/2_2_reloadable5.o + set -o pipefail; (rm -f "2_2_reloadable5/Release/2_2_reloadable5.##") 2>&1 |& tee -a 2_2_reloadable5/2_2_reloadable5.log 2_2_reloadable5/timestamped_log/2_2_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable5/scripts/2_2_reloadable5.prx) 2>&1 |& tee -a 2_2_reloadable5/2_2_reloadable5.log 2_2_reloadable5/timestamped_log/2_2_reloadable5-${ts}.log + +2_2_reloadable7: 0_0_reloadable5 + mkdir -p 2_2_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_2_reloadable7/Release/2_2_reloadable7.o + set -o pipefail; (rm -f "2_2_reloadable7/Release/2_2_reloadable7.##") 2>&1 |& tee -a 2_2_reloadable7/2_2_reloadable7.log 2_2_reloadable7/timestamped_log/2_2_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable7/scripts/2_2_reloadable7.prx) 2>&1 |& tee -a 2_2_reloadable7/2_2_reloadable7.log 2_2_reloadable7/timestamped_log/2_2_reloadable7-${ts}.log + +2_2_reloadable11: 0_0_reloadable5 + mkdir -p 2_2_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_2_reloadable11/Release/2_2_reloadable11.o + set -o pipefail; (rm -f "2_2_reloadable11/Release/2_2_reloadable11.##") 2>&1 |& tee -a 2_2_reloadable11/2_2_reloadable11.log 2_2_reloadable11/timestamped_log/2_2_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable11/scripts/2_2_reloadable11.prx) 2>&1 |& tee -a 2_2_reloadable11/2_2_reloadable11.log 2_2_reloadable11/timestamped_log/2_2_reloadable11-${ts}.log + +2_2_reloadable15: 0_0_reloadable5 + mkdir -p 2_2_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_2_reloadable15/Release/2_2_reloadable15.o + set -o pipefail; (rm -f "2_2_reloadable15/Release/2_2_reloadable15.##") 2>&1 |& tee -a 2_2_reloadable15/2_2_reloadable15.log 2_2_reloadable15/timestamped_log/2_2_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable15/scripts/2_2_reloadable15.prx) 2>&1 |& tee -a 2_2_reloadable15/2_2_reloadable15.log 2_2_reloadable15/timestamped_log/2_2_reloadable15-${ts}.log + +2_3_reloadable5: 0_0_reloadable5 + mkdir -p 2_3_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_3_reloadable5/Release/2_3_reloadable5.o + set -o pipefail; (rm -f "2_3_reloadable5/Release/2_3_reloadable5.##") 2>&1 |& tee -a 2_3_reloadable5/2_3_reloadable5.log 2_3_reloadable5/timestamped_log/2_3_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable5/scripts/2_3_reloadable5.prx) 2>&1 |& tee -a 2_3_reloadable5/2_3_reloadable5.log 2_3_reloadable5/timestamped_log/2_3_reloadable5-${ts}.log + +2_3_reloadable7: 0_0_reloadable5 + mkdir -p 2_3_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_3_reloadable7/Release/2_3_reloadable7.o + set -o pipefail; (rm -f "2_3_reloadable7/Release/2_3_reloadable7.##") 2>&1 |& tee -a 2_3_reloadable7/2_3_reloadable7.log 2_3_reloadable7/timestamped_log/2_3_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable7/scripts/2_3_reloadable7.prx) 2>&1 |& tee -a 2_3_reloadable7/2_3_reloadable7.log 2_3_reloadable7/timestamped_log/2_3_reloadable7-${ts}.log + +2_3_reloadable11: 0_0_reloadable5 + mkdir -p 2_3_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_3_reloadable11/Release/2_3_reloadable11.o + set -o pipefail; (rm -f "2_3_reloadable11/Release/2_3_reloadable11.##") 2>&1 |& tee -a 2_3_reloadable11/2_3_reloadable11.log 2_3_reloadable11/timestamped_log/2_3_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable11/scripts/2_3_reloadable11.prx) 2>&1 |& tee -a 2_3_reloadable11/2_3_reloadable11.log 2_3_reloadable11/timestamped_log/2_3_reloadable11-${ts}.log + +2_3_reloadable15: 0_0_reloadable5 + mkdir -p 2_3_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 2_3_reloadable15/Release/2_3_reloadable15.o + set -o pipefail; (rm -f "2_3_reloadable15/Release/2_3_reloadable15.##") 2>&1 |& tee -a 2_3_reloadable15/2_3_reloadable15.log 2_3_reloadable15/timestamped_log/2_3_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable15/scripts/2_3_reloadable15.prx) 2>&1 |& tee -a 2_3_reloadable15/2_3_reloadable15.log 2_3_reloadable15/timestamped_log/2_3_reloadable15-${ts}.log + +3_0_reloadable5: 0_0_reloadable5 + mkdir -p 3_0_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_0_reloadable5/Release/3_0_reloadable5.o + set -o pipefail; (rm -f "3_0_reloadable5/Release/3_0_reloadable5.##") 2>&1 |& tee -a 3_0_reloadable5/3_0_reloadable5.log 3_0_reloadable5/timestamped_log/3_0_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable5/scripts/3_0_reloadable5.prx) 2>&1 |& tee -a 3_0_reloadable5/3_0_reloadable5.log 3_0_reloadable5/timestamped_log/3_0_reloadable5-${ts}.log + +3_0_reloadable7: 0_0_reloadable5 + mkdir -p 3_0_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_0_reloadable7/Release/3_0_reloadable7.o + set -o pipefail; (rm -f "3_0_reloadable7/Release/3_0_reloadable7.##") 2>&1 |& tee -a 3_0_reloadable7/3_0_reloadable7.log 3_0_reloadable7/timestamped_log/3_0_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable7/scripts/3_0_reloadable7.prx) 2>&1 |& tee -a 3_0_reloadable7/3_0_reloadable7.log 3_0_reloadable7/timestamped_log/3_0_reloadable7-${ts}.log + +3_0_reloadable11: 0_0_reloadable5 + mkdir -p 3_0_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_0_reloadable11/Release/3_0_reloadable11.o + set -o pipefail; (rm -f "3_0_reloadable11/Release/3_0_reloadable11.##") 2>&1 |& tee -a 3_0_reloadable11/3_0_reloadable11.log 3_0_reloadable11/timestamped_log/3_0_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable11/scripts/3_0_reloadable11.prx) 2>&1 |& tee -a 3_0_reloadable11/3_0_reloadable11.log 3_0_reloadable11/timestamped_log/3_0_reloadable11-${ts}.log + +3_0_reloadable15: 0_0_reloadable5 + mkdir -p 3_0_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_0_reloadable15/Release/3_0_reloadable15.o + set -o pipefail; (rm -f "3_0_reloadable15/Release/3_0_reloadable15.##") 2>&1 |& tee -a 3_0_reloadable15/3_0_reloadable15.log 3_0_reloadable15/timestamped_log/3_0_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable15/scripts/3_0_reloadable15.prx) 2>&1 |& tee -a 3_0_reloadable15/3_0_reloadable15.log 3_0_reloadable15/timestamped_log/3_0_reloadable15-${ts}.log + +3_1_reloadable5: 0_0_reloadable5 + mkdir -p 3_1_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_1_reloadable5/Release/3_1_reloadable5.o + set -o pipefail; (rm -f "3_1_reloadable5/Release/3_1_reloadable5.##") 2>&1 |& tee -a 3_1_reloadable5/3_1_reloadable5.log 3_1_reloadable5/timestamped_log/3_1_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable5/scripts/3_1_reloadable5.prx) 2>&1 |& tee -a 3_1_reloadable5/3_1_reloadable5.log 3_1_reloadable5/timestamped_log/3_1_reloadable5-${ts}.log + +3_1_reloadable7: 0_0_reloadable5 + mkdir -p 3_1_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_1_reloadable7/Release/3_1_reloadable7.o + set -o pipefail; (rm -f "3_1_reloadable7/Release/3_1_reloadable7.##") 2>&1 |& tee -a 3_1_reloadable7/3_1_reloadable7.log 3_1_reloadable7/timestamped_log/3_1_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable7/scripts/3_1_reloadable7.prx) 2>&1 |& tee -a 3_1_reloadable7/3_1_reloadable7.log 3_1_reloadable7/timestamped_log/3_1_reloadable7-${ts}.log + +3_1_reloadable11: 0_0_reloadable5 + mkdir -p 3_1_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_1_reloadable11/Release/3_1_reloadable11.o + set -o pipefail; (rm -f "3_1_reloadable11/Release/3_1_reloadable11.##") 2>&1 |& tee -a 3_1_reloadable11/3_1_reloadable11.log 3_1_reloadable11/timestamped_log/3_1_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable11/scripts/3_1_reloadable11.prx) 2>&1 |& tee -a 3_1_reloadable11/3_1_reloadable11.log 3_1_reloadable11/timestamped_log/3_1_reloadable11-${ts}.log + +3_1_reloadable15: 0_0_reloadable5 + mkdir -p 3_1_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_1_reloadable15/Release/3_1_reloadable15.o + set -o pipefail; (rm -f "3_1_reloadable15/Release/3_1_reloadable15.##") 2>&1 |& tee -a 3_1_reloadable15/3_1_reloadable15.log 3_1_reloadable15/timestamped_log/3_1_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable15/scripts/3_1_reloadable15.prx) 2>&1 |& tee -a 3_1_reloadable15/3_1_reloadable15.log 3_1_reloadable15/timestamped_log/3_1_reloadable15-${ts}.log + +3_2_reloadable5: 0_0_reloadable5 + mkdir -p 3_2_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_2_reloadable5/Release/3_2_reloadable5.o + set -o pipefail; (rm -f "3_2_reloadable5/Release/3_2_reloadable5.##") 2>&1 |& tee -a 3_2_reloadable5/3_2_reloadable5.log 3_2_reloadable5/timestamped_log/3_2_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable5/scripts/3_2_reloadable5.prx) 2>&1 |& tee -a 3_2_reloadable5/3_2_reloadable5.log 3_2_reloadable5/timestamped_log/3_2_reloadable5-${ts}.log + +3_2_reloadable7: 0_0_reloadable5 + mkdir -p 3_2_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_2_reloadable7/Release/3_2_reloadable7.o + set -o pipefail; (rm -f "3_2_reloadable7/Release/3_2_reloadable7.##") 2>&1 |& tee -a 3_2_reloadable7/3_2_reloadable7.log 3_2_reloadable7/timestamped_log/3_2_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable7/scripts/3_2_reloadable7.prx) 2>&1 |& tee -a 3_2_reloadable7/3_2_reloadable7.log 3_2_reloadable7/timestamped_log/3_2_reloadable7-${ts}.log + +3_2_reloadable11: 0_0_reloadable5 + mkdir -p 3_2_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_2_reloadable11/Release/3_2_reloadable11.o + set -o pipefail; (rm -f "3_2_reloadable11/Release/3_2_reloadable11.##") 2>&1 |& tee -a 3_2_reloadable11/3_2_reloadable11.log 3_2_reloadable11/timestamped_log/3_2_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable11/scripts/3_2_reloadable11.prx) 2>&1 |& tee -a 3_2_reloadable11/3_2_reloadable11.log 3_2_reloadable11/timestamped_log/3_2_reloadable11-${ts}.log + +3_2_reloadable15: 0_0_reloadable5 + mkdir -p 3_2_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_2_reloadable15/Release/3_2_reloadable15.o + set -o pipefail; (rm -f "3_2_reloadable15/Release/3_2_reloadable15.##") 2>&1 |& tee -a 3_2_reloadable15/3_2_reloadable15.log 3_2_reloadable15/timestamped_log/3_2_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable15/scripts/3_2_reloadable15.prx) 2>&1 |& tee -a 3_2_reloadable15/3_2_reloadable15.log 3_2_reloadable15/timestamped_log/3_2_reloadable15-${ts}.log + +3_3_reloadable5: 0_0_reloadable5 + mkdir -p 3_3_reloadable5/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_3_reloadable5/Release/3_3_reloadable5.o + set -o pipefail; (rm -f "3_3_reloadable5/Release/3_3_reloadable5.##") 2>&1 |& tee -a 3_3_reloadable5/3_3_reloadable5.log 3_3_reloadable5/timestamped_log/3_3_reloadable5-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable5/scripts/3_3_reloadable5.prx) 2>&1 |& tee -a 3_3_reloadable5/3_3_reloadable5.log 3_3_reloadable5/timestamped_log/3_3_reloadable5-${ts}.log + +3_3_reloadable7: 0_0_reloadable5 + mkdir -p 3_3_reloadable7/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_3_reloadable7/Release/3_3_reloadable7.o + set -o pipefail; (rm -f "3_3_reloadable7/Release/3_3_reloadable7.##") 2>&1 |& tee -a 3_3_reloadable7/3_3_reloadable7.log 3_3_reloadable7/timestamped_log/3_3_reloadable7-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable7/scripts/3_3_reloadable7.prx) 2>&1 |& tee -a 3_3_reloadable7/3_3_reloadable7.log 3_3_reloadable7/timestamped_log/3_3_reloadable7-${ts}.log + +3_3_reloadable11: 0_0_reloadable5 + mkdir -p 3_3_reloadable11/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_3_reloadable11/Release/3_3_reloadable11.o + set -o pipefail; (rm -f "3_3_reloadable11/Release/3_3_reloadable11.##") 2>&1 |& tee -a 3_3_reloadable11/3_3_reloadable11.log 3_3_reloadable11/timestamped_log/3_3_reloadable11-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable11/scripts/3_3_reloadable11.prx) 2>&1 |& tee -a 3_3_reloadable11/3_3_reloadable11.log 3_3_reloadable11/timestamped_log/3_3_reloadable11-${ts}.log + +3_3_reloadable15: 0_0_reloadable5 + mkdir -p 3_3_reloadable15/Release + cp 0_0_reloadable5/Release/0_0_reloadable5.o 3_3_reloadable15/Release/3_3_reloadable15.o + set -o pipefail; (rm -f "3_3_reloadable15/Release/3_3_reloadable15.##") 2>&1 |& tee -a 3_3_reloadable15/3_3_reloadable15.log 3_3_reloadable15/timestamped_log/3_3_reloadable15-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable15/scripts/3_3_reloadable15.prx) 2>&1 |& tee -a 3_3_reloadable15/3_3_reloadable15.log 3_3_reloadable15/timestamped_log/3_3_reloadable15-${ts}.log + +0_1_reloadable17: 0_0_reloadable17 + mkdir -p 0_1_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 0_1_reloadable17/Release/0_1_reloadable17.o + set -o pipefail; (rm -f "0_1_reloadable17/Release/0_1_reloadable17.##") 2>&1 |& tee -a 0_1_reloadable17/0_1_reloadable17.log 0_1_reloadable17/timestamped_log/0_1_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable17/scripts/0_1_reloadable17.prx) 2>&1 |& tee -a 0_1_reloadable17/0_1_reloadable17.log 0_1_reloadable17/timestamped_log/0_1_reloadable17-${ts}.log + +0_2_reloadable17: 0_0_reloadable17 + mkdir -p 0_2_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 0_2_reloadable17/Release/0_2_reloadable17.o + set -o pipefail; (rm -f "0_2_reloadable17/Release/0_2_reloadable17.##") 2>&1 |& tee -a 0_2_reloadable17/0_2_reloadable17.log 0_2_reloadable17/timestamped_log/0_2_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable17/scripts/0_2_reloadable17.prx) 2>&1 |& tee -a 0_2_reloadable17/0_2_reloadable17.log 0_2_reloadable17/timestamped_log/0_2_reloadable17-${ts}.log + +0_3_reloadable17: 0_0_reloadable17 + mkdir -p 0_3_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 0_3_reloadable17/Release/0_3_reloadable17.o + set -o pipefail; (rm -f "0_3_reloadable17/Release/0_3_reloadable17.##") 2>&1 |& tee -a 0_3_reloadable17/0_3_reloadable17.log 0_3_reloadable17/timestamped_log/0_3_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable17/scripts/0_3_reloadable17.prx) 2>&1 |& tee -a 0_3_reloadable17/0_3_reloadable17.log 0_3_reloadable17/timestamped_log/0_3_reloadable17-${ts}.log + +1_0_reloadable17: 0_0_reloadable17 + mkdir -p 1_0_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 1_0_reloadable17/Release/1_0_reloadable17.o + set -o pipefail; (rm -f "1_0_reloadable17/Release/1_0_reloadable17.##") 2>&1 |& tee -a 1_0_reloadable17/1_0_reloadable17.log 1_0_reloadable17/timestamped_log/1_0_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable17/scripts/1_0_reloadable17.prx) 2>&1 |& tee -a 1_0_reloadable17/1_0_reloadable17.log 1_0_reloadable17/timestamped_log/1_0_reloadable17-${ts}.log + +1_1_reloadable17: 0_0_reloadable17 + mkdir -p 1_1_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 1_1_reloadable17/Release/1_1_reloadable17.o + set -o pipefail; (rm -f "1_1_reloadable17/Release/1_1_reloadable17.##") 2>&1 |& tee -a 1_1_reloadable17/1_1_reloadable17.log 1_1_reloadable17/timestamped_log/1_1_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable17/scripts/1_1_reloadable17.prx) 2>&1 |& tee -a 1_1_reloadable17/1_1_reloadable17.log 1_1_reloadable17/timestamped_log/1_1_reloadable17-${ts}.log + +1_2_reloadable17: 0_0_reloadable17 + mkdir -p 1_2_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 1_2_reloadable17/Release/1_2_reloadable17.o + set -o pipefail; (rm -f "1_2_reloadable17/Release/1_2_reloadable17.##") 2>&1 |& tee -a 1_2_reloadable17/1_2_reloadable17.log 1_2_reloadable17/timestamped_log/1_2_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable17/scripts/1_2_reloadable17.prx) 2>&1 |& tee -a 1_2_reloadable17/1_2_reloadable17.log 1_2_reloadable17/timestamped_log/1_2_reloadable17-${ts}.log + +1_3_reloadable17: 0_0_reloadable17 + mkdir -p 1_3_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 1_3_reloadable17/Release/1_3_reloadable17.o + set -o pipefail; (rm -f "1_3_reloadable17/Release/1_3_reloadable17.##") 2>&1 |& tee -a 1_3_reloadable17/1_3_reloadable17.log 1_3_reloadable17/timestamped_log/1_3_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable17/scripts/1_3_reloadable17.prx) 2>&1 |& tee -a 1_3_reloadable17/1_3_reloadable17.log 1_3_reloadable17/timestamped_log/1_3_reloadable17-${ts}.log + +2_0_reloadable17: 0_0_reloadable17 + mkdir -p 2_0_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 2_0_reloadable17/Release/2_0_reloadable17.o + set -o pipefail; (rm -f "2_0_reloadable17/Release/2_0_reloadable17.##") 2>&1 |& tee -a 2_0_reloadable17/2_0_reloadable17.log 2_0_reloadable17/timestamped_log/2_0_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable17/scripts/2_0_reloadable17.prx) 2>&1 |& tee -a 2_0_reloadable17/2_0_reloadable17.log 2_0_reloadable17/timestamped_log/2_0_reloadable17-${ts}.log + +2_1_reloadable17: 0_0_reloadable17 + mkdir -p 2_1_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 2_1_reloadable17/Release/2_1_reloadable17.o + set -o pipefail; (rm -f "2_1_reloadable17/Release/2_1_reloadable17.##") 2>&1 |& tee -a 2_1_reloadable17/2_1_reloadable17.log 2_1_reloadable17/timestamped_log/2_1_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable17/scripts/2_1_reloadable17.prx) 2>&1 |& tee -a 2_1_reloadable17/2_1_reloadable17.log 2_1_reloadable17/timestamped_log/2_1_reloadable17-${ts}.log + +2_2_reloadable17: 0_0_reloadable17 + mkdir -p 2_2_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 2_2_reloadable17/Release/2_2_reloadable17.o + set -o pipefail; (rm -f "2_2_reloadable17/Release/2_2_reloadable17.##") 2>&1 |& tee -a 2_2_reloadable17/2_2_reloadable17.log 2_2_reloadable17/timestamped_log/2_2_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable17/scripts/2_2_reloadable17.prx) 2>&1 |& tee -a 2_2_reloadable17/2_2_reloadable17.log 2_2_reloadable17/timestamped_log/2_2_reloadable17-${ts}.log + +2_3_reloadable17: 0_0_reloadable17 + mkdir -p 2_3_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 2_3_reloadable17/Release/2_3_reloadable17.o + set -o pipefail; (rm -f "2_3_reloadable17/Release/2_3_reloadable17.##") 2>&1 |& tee -a 2_3_reloadable17/2_3_reloadable17.log 2_3_reloadable17/timestamped_log/2_3_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable17/scripts/2_3_reloadable17.prx) 2>&1 |& tee -a 2_3_reloadable17/2_3_reloadable17.log 2_3_reloadable17/timestamped_log/2_3_reloadable17-${ts}.log + +3_0_reloadable17: 0_0_reloadable17 + mkdir -p 3_0_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 3_0_reloadable17/Release/3_0_reloadable17.o + set -o pipefail; (rm -f "3_0_reloadable17/Release/3_0_reloadable17.##") 2>&1 |& tee -a 3_0_reloadable17/3_0_reloadable17.log 3_0_reloadable17/timestamped_log/3_0_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable17/scripts/3_0_reloadable17.prx) 2>&1 |& tee -a 3_0_reloadable17/3_0_reloadable17.log 3_0_reloadable17/timestamped_log/3_0_reloadable17-${ts}.log + +3_1_reloadable17: 0_0_reloadable17 + mkdir -p 3_1_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 3_1_reloadable17/Release/3_1_reloadable17.o + set -o pipefail; (rm -f "3_1_reloadable17/Release/3_1_reloadable17.##") 2>&1 |& tee -a 3_1_reloadable17/3_1_reloadable17.log 3_1_reloadable17/timestamped_log/3_1_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable17/scripts/3_1_reloadable17.prx) 2>&1 |& tee -a 3_1_reloadable17/3_1_reloadable17.log 3_1_reloadable17/timestamped_log/3_1_reloadable17-${ts}.log + +3_2_reloadable17: 0_0_reloadable17 + mkdir -p 3_2_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 3_2_reloadable17/Release/3_2_reloadable17.o + set -o pipefail; (rm -f "3_2_reloadable17/Release/3_2_reloadable17.##") 2>&1 |& tee -a 3_2_reloadable17/3_2_reloadable17.log 3_2_reloadable17/timestamped_log/3_2_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable17/scripts/3_2_reloadable17.prx) 2>&1 |& tee -a 3_2_reloadable17/3_2_reloadable17.log 3_2_reloadable17/timestamped_log/3_2_reloadable17-${ts}.log + +3_3_reloadable17: 0_0_reloadable17 + mkdir -p 3_3_reloadable17/Release + cp 0_0_reloadable17/Release/0_0_reloadable17.o 3_3_reloadable17/Release/3_3_reloadable17.o + set -o pipefail; (rm -f "3_3_reloadable17/Release/3_3_reloadable17.##") 2>&1 |& tee -a 3_3_reloadable17/3_3_reloadable17.log 3_3_reloadable17/timestamped_log/3_3_reloadable17-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable17/scripts/3_3_reloadable17.prx) 2>&1 |& tee -a 3_3_reloadable17/3_3_reloadable17.log 3_3_reloadable17/timestamped_log/3_3_reloadable17-${ts}.log + +0_1_reloadable19: 0_0_reloadable19 + mkdir -p 0_1_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 0_1_reloadable19/Release/0_1_reloadable19.o + set -o pipefail; (rm -f "0_1_reloadable19/Release/0_1_reloadable19.##") 2>&1 |& tee -a 0_1_reloadable19/0_1_reloadable19.log 0_1_reloadable19/timestamped_log/0_1_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable19/scripts/0_1_reloadable19.prx) 2>&1 |& tee -a 0_1_reloadable19/0_1_reloadable19.log 0_1_reloadable19/timestamped_log/0_1_reloadable19-${ts}.log + +0_2_reloadable19: 0_0_reloadable19 + mkdir -p 0_2_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 0_2_reloadable19/Release/0_2_reloadable19.o + set -o pipefail; (rm -f "0_2_reloadable19/Release/0_2_reloadable19.##") 2>&1 |& tee -a 0_2_reloadable19/0_2_reloadable19.log 0_2_reloadable19/timestamped_log/0_2_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable19/scripts/0_2_reloadable19.prx) 2>&1 |& tee -a 0_2_reloadable19/0_2_reloadable19.log 0_2_reloadable19/timestamped_log/0_2_reloadable19-${ts}.log + +0_3_reloadable19: 0_0_reloadable19 + mkdir -p 0_3_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 0_3_reloadable19/Release/0_3_reloadable19.o + set -o pipefail; (rm -f "0_3_reloadable19/Release/0_3_reloadable19.##") 2>&1 |& tee -a 0_3_reloadable19/0_3_reloadable19.log 0_3_reloadable19/timestamped_log/0_3_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable19/scripts/0_3_reloadable19.prx) 2>&1 |& tee -a 0_3_reloadable19/0_3_reloadable19.log 0_3_reloadable19/timestamped_log/0_3_reloadable19-${ts}.log + +1_0_reloadable19: 0_0_reloadable19 + mkdir -p 1_0_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 1_0_reloadable19/Release/1_0_reloadable19.o + set -o pipefail; (rm -f "1_0_reloadable19/Release/1_0_reloadable19.##") 2>&1 |& tee -a 1_0_reloadable19/1_0_reloadable19.log 1_0_reloadable19/timestamped_log/1_0_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable19/scripts/1_0_reloadable19.prx) 2>&1 |& tee -a 1_0_reloadable19/1_0_reloadable19.log 1_0_reloadable19/timestamped_log/1_0_reloadable19-${ts}.log + +1_1_reloadable19: 0_0_reloadable19 + mkdir -p 1_1_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 1_1_reloadable19/Release/1_1_reloadable19.o + set -o pipefail; (rm -f "1_1_reloadable19/Release/1_1_reloadable19.##") 2>&1 |& tee -a 1_1_reloadable19/1_1_reloadable19.log 1_1_reloadable19/timestamped_log/1_1_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable19/scripts/1_1_reloadable19.prx) 2>&1 |& tee -a 1_1_reloadable19/1_1_reloadable19.log 1_1_reloadable19/timestamped_log/1_1_reloadable19-${ts}.log + +1_2_reloadable19: 0_0_reloadable19 + mkdir -p 1_2_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 1_2_reloadable19/Release/1_2_reloadable19.o + set -o pipefail; (rm -f "1_2_reloadable19/Release/1_2_reloadable19.##") 2>&1 |& tee -a 1_2_reloadable19/1_2_reloadable19.log 1_2_reloadable19/timestamped_log/1_2_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable19/scripts/1_2_reloadable19.prx) 2>&1 |& tee -a 1_2_reloadable19/1_2_reloadable19.log 1_2_reloadable19/timestamped_log/1_2_reloadable19-${ts}.log + +1_3_reloadable19: 0_0_reloadable19 + mkdir -p 1_3_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 1_3_reloadable19/Release/1_3_reloadable19.o + set -o pipefail; (rm -f "1_3_reloadable19/Release/1_3_reloadable19.##") 2>&1 |& tee -a 1_3_reloadable19/1_3_reloadable19.log 1_3_reloadable19/timestamped_log/1_3_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable19/scripts/1_3_reloadable19.prx) 2>&1 |& tee -a 1_3_reloadable19/1_3_reloadable19.log 1_3_reloadable19/timestamped_log/1_3_reloadable19-${ts}.log + +2_0_reloadable19: 0_0_reloadable19 + mkdir -p 2_0_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 2_0_reloadable19/Release/2_0_reloadable19.o + set -o pipefail; (rm -f "2_0_reloadable19/Release/2_0_reloadable19.##") 2>&1 |& tee -a 2_0_reloadable19/2_0_reloadable19.log 2_0_reloadable19/timestamped_log/2_0_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable19/scripts/2_0_reloadable19.prx) 2>&1 |& tee -a 2_0_reloadable19/2_0_reloadable19.log 2_0_reloadable19/timestamped_log/2_0_reloadable19-${ts}.log + +2_1_reloadable19: 0_0_reloadable19 + mkdir -p 2_1_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 2_1_reloadable19/Release/2_1_reloadable19.o + set -o pipefail; (rm -f "2_1_reloadable19/Release/2_1_reloadable19.##") 2>&1 |& tee -a 2_1_reloadable19/2_1_reloadable19.log 2_1_reloadable19/timestamped_log/2_1_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable19/scripts/2_1_reloadable19.prx) 2>&1 |& tee -a 2_1_reloadable19/2_1_reloadable19.log 2_1_reloadable19/timestamped_log/2_1_reloadable19-${ts}.log + +2_2_reloadable19: 0_0_reloadable19 + mkdir -p 2_2_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 2_2_reloadable19/Release/2_2_reloadable19.o + set -o pipefail; (rm -f "2_2_reloadable19/Release/2_2_reloadable19.##") 2>&1 |& tee -a 2_2_reloadable19/2_2_reloadable19.log 2_2_reloadable19/timestamped_log/2_2_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable19/scripts/2_2_reloadable19.prx) 2>&1 |& tee -a 2_2_reloadable19/2_2_reloadable19.log 2_2_reloadable19/timestamped_log/2_2_reloadable19-${ts}.log + +2_3_reloadable19: 0_0_reloadable19 + mkdir -p 2_3_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 2_3_reloadable19/Release/2_3_reloadable19.o + set -o pipefail; (rm -f "2_3_reloadable19/Release/2_3_reloadable19.##") 2>&1 |& tee -a 2_3_reloadable19/2_3_reloadable19.log 2_3_reloadable19/timestamped_log/2_3_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable19/scripts/2_3_reloadable19.prx) 2>&1 |& tee -a 2_3_reloadable19/2_3_reloadable19.log 2_3_reloadable19/timestamped_log/2_3_reloadable19-${ts}.log + +3_0_reloadable19: 0_0_reloadable19 + mkdir -p 3_0_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 3_0_reloadable19/Release/3_0_reloadable19.o + set -o pipefail; (rm -f "3_0_reloadable19/Release/3_0_reloadable19.##") 2>&1 |& tee -a 3_0_reloadable19/3_0_reloadable19.log 3_0_reloadable19/timestamped_log/3_0_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable19/scripts/3_0_reloadable19.prx) 2>&1 |& tee -a 3_0_reloadable19/3_0_reloadable19.log 3_0_reloadable19/timestamped_log/3_0_reloadable19-${ts}.log + +3_1_reloadable19: 0_0_reloadable19 + mkdir -p 3_1_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 3_1_reloadable19/Release/3_1_reloadable19.o + set -o pipefail; (rm -f "3_1_reloadable19/Release/3_1_reloadable19.##") 2>&1 |& tee -a 3_1_reloadable19/3_1_reloadable19.log 3_1_reloadable19/timestamped_log/3_1_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable19/scripts/3_1_reloadable19.prx) 2>&1 |& tee -a 3_1_reloadable19/3_1_reloadable19.log 3_1_reloadable19/timestamped_log/3_1_reloadable19-${ts}.log + +3_2_reloadable19: 0_0_reloadable19 + mkdir -p 3_2_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 3_2_reloadable19/Release/3_2_reloadable19.o + set -o pipefail; (rm -f "3_2_reloadable19/Release/3_2_reloadable19.##") 2>&1 |& tee -a 3_2_reloadable19/3_2_reloadable19.log 3_2_reloadable19/timestamped_log/3_2_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable19/scripts/3_2_reloadable19.prx) 2>&1 |& tee -a 3_2_reloadable19/3_2_reloadable19.log 3_2_reloadable19/timestamped_log/3_2_reloadable19-${ts}.log + +3_3_reloadable19: 0_0_reloadable19 + mkdir -p 3_3_reloadable19/Release + cp 0_0_reloadable19/Release/0_0_reloadable19.o 3_3_reloadable19/Release/3_3_reloadable19.o + set -o pipefail; (rm -f "3_3_reloadable19/Release/3_3_reloadable19.##") 2>&1 |& tee -a 3_3_reloadable19/3_3_reloadable19.log 3_3_reloadable19/timestamped_log/3_3_reloadable19-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable19/scripts/3_3_reloadable19.prx) 2>&1 |& tee -a 3_3_reloadable19/3_3_reloadable19.log 3_3_reloadable19/timestamped_log/3_3_reloadable19-${ts}.log + +0_1_reloadable20: 0_0_reloadable20 + mkdir -p 0_1_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 0_1_reloadable20/Release/0_1_reloadable20.o + set -o pipefail; (rm -f "0_1_reloadable20/Release/0_1_reloadable20.##") 2>&1 |& tee -a 0_1_reloadable20/0_1_reloadable20.log 0_1_reloadable20/timestamped_log/0_1_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable20/scripts/0_1_reloadable20.prx) 2>&1 |& tee -a 0_1_reloadable20/0_1_reloadable20.log 0_1_reloadable20/timestamped_log/0_1_reloadable20-${ts}.log + +0_2_reloadable20: 0_0_reloadable20 + mkdir -p 0_2_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 0_2_reloadable20/Release/0_2_reloadable20.o + set -o pipefail; (rm -f "0_2_reloadable20/Release/0_2_reloadable20.##") 2>&1 |& tee -a 0_2_reloadable20/0_2_reloadable20.log 0_2_reloadable20/timestamped_log/0_2_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable20/scripts/0_2_reloadable20.prx) 2>&1 |& tee -a 0_2_reloadable20/0_2_reloadable20.log 0_2_reloadable20/timestamped_log/0_2_reloadable20-${ts}.log + +0_3_reloadable20: 0_0_reloadable20 + mkdir -p 0_3_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 0_3_reloadable20/Release/0_3_reloadable20.o + set -o pipefail; (rm -f "0_3_reloadable20/Release/0_3_reloadable20.##") 2>&1 |& tee -a 0_3_reloadable20/0_3_reloadable20.log 0_3_reloadable20/timestamped_log/0_3_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable20/scripts/0_3_reloadable20.prx) 2>&1 |& tee -a 0_3_reloadable20/0_3_reloadable20.log 0_3_reloadable20/timestamped_log/0_3_reloadable20-${ts}.log + +1_0_reloadable20: 0_0_reloadable20 + mkdir -p 1_0_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 1_0_reloadable20/Release/1_0_reloadable20.o + set -o pipefail; (rm -f "1_0_reloadable20/Release/1_0_reloadable20.##") 2>&1 |& tee -a 1_0_reloadable20/1_0_reloadable20.log 1_0_reloadable20/timestamped_log/1_0_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable20/scripts/1_0_reloadable20.prx) 2>&1 |& tee -a 1_0_reloadable20/1_0_reloadable20.log 1_0_reloadable20/timestamped_log/1_0_reloadable20-${ts}.log + +1_1_reloadable20: 0_0_reloadable20 + mkdir -p 1_1_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 1_1_reloadable20/Release/1_1_reloadable20.o + set -o pipefail; (rm -f "1_1_reloadable20/Release/1_1_reloadable20.##") 2>&1 |& tee -a 1_1_reloadable20/1_1_reloadable20.log 1_1_reloadable20/timestamped_log/1_1_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable20/scripts/1_1_reloadable20.prx) 2>&1 |& tee -a 1_1_reloadable20/1_1_reloadable20.log 1_1_reloadable20/timestamped_log/1_1_reloadable20-${ts}.log + +1_2_reloadable20: 0_0_reloadable20 + mkdir -p 1_2_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 1_2_reloadable20/Release/1_2_reloadable20.o + set -o pipefail; (rm -f "1_2_reloadable20/Release/1_2_reloadable20.##") 2>&1 |& tee -a 1_2_reloadable20/1_2_reloadable20.log 1_2_reloadable20/timestamped_log/1_2_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable20/scripts/1_2_reloadable20.prx) 2>&1 |& tee -a 1_2_reloadable20/1_2_reloadable20.log 1_2_reloadable20/timestamped_log/1_2_reloadable20-${ts}.log + +1_3_reloadable20: 0_0_reloadable20 + mkdir -p 1_3_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 1_3_reloadable20/Release/1_3_reloadable20.o + set -o pipefail; (rm -f "1_3_reloadable20/Release/1_3_reloadable20.##") 2>&1 |& tee -a 1_3_reloadable20/1_3_reloadable20.log 1_3_reloadable20/timestamped_log/1_3_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable20/scripts/1_3_reloadable20.prx) 2>&1 |& tee -a 1_3_reloadable20/1_3_reloadable20.log 1_3_reloadable20/timestamped_log/1_3_reloadable20-${ts}.log + +2_0_reloadable20: 0_0_reloadable20 + mkdir -p 2_0_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 2_0_reloadable20/Release/2_0_reloadable20.o + set -o pipefail; (rm -f "2_0_reloadable20/Release/2_0_reloadable20.##") 2>&1 |& tee -a 2_0_reloadable20/2_0_reloadable20.log 2_0_reloadable20/timestamped_log/2_0_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable20/scripts/2_0_reloadable20.prx) 2>&1 |& tee -a 2_0_reloadable20/2_0_reloadable20.log 2_0_reloadable20/timestamped_log/2_0_reloadable20-${ts}.log + +2_1_reloadable20: 0_0_reloadable20 + mkdir -p 2_1_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 2_1_reloadable20/Release/2_1_reloadable20.o + set -o pipefail; (rm -f "2_1_reloadable20/Release/2_1_reloadable20.##") 2>&1 |& tee -a 2_1_reloadable20/2_1_reloadable20.log 2_1_reloadable20/timestamped_log/2_1_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable20/scripts/2_1_reloadable20.prx) 2>&1 |& tee -a 2_1_reloadable20/2_1_reloadable20.log 2_1_reloadable20/timestamped_log/2_1_reloadable20-${ts}.log + +2_2_reloadable20: 0_0_reloadable20 + mkdir -p 2_2_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 2_2_reloadable20/Release/2_2_reloadable20.o + set -o pipefail; (rm -f "2_2_reloadable20/Release/2_2_reloadable20.##") 2>&1 |& tee -a 2_2_reloadable20/2_2_reloadable20.log 2_2_reloadable20/timestamped_log/2_2_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable20/scripts/2_2_reloadable20.prx) 2>&1 |& tee -a 2_2_reloadable20/2_2_reloadable20.log 2_2_reloadable20/timestamped_log/2_2_reloadable20-${ts}.log + +2_3_reloadable20: 0_0_reloadable20 + mkdir -p 2_3_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 2_3_reloadable20/Release/2_3_reloadable20.o + set -o pipefail; (rm -f "2_3_reloadable20/Release/2_3_reloadable20.##") 2>&1 |& tee -a 2_3_reloadable20/2_3_reloadable20.log 2_3_reloadable20/timestamped_log/2_3_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable20/scripts/2_3_reloadable20.prx) 2>&1 |& tee -a 2_3_reloadable20/2_3_reloadable20.log 2_3_reloadable20/timestamped_log/2_3_reloadable20-${ts}.log + +3_0_reloadable20: 0_0_reloadable20 + mkdir -p 3_0_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 3_0_reloadable20/Release/3_0_reloadable20.o + set -o pipefail; (rm -f "3_0_reloadable20/Release/3_0_reloadable20.##") 2>&1 |& tee -a 3_0_reloadable20/3_0_reloadable20.log 3_0_reloadable20/timestamped_log/3_0_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable20/scripts/3_0_reloadable20.prx) 2>&1 |& tee -a 3_0_reloadable20/3_0_reloadable20.log 3_0_reloadable20/timestamped_log/3_0_reloadable20-${ts}.log + +3_1_reloadable20: 0_0_reloadable20 + mkdir -p 3_1_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 3_1_reloadable20/Release/3_1_reloadable20.o + set -o pipefail; (rm -f "3_1_reloadable20/Release/3_1_reloadable20.##") 2>&1 |& tee -a 3_1_reloadable20/3_1_reloadable20.log 3_1_reloadable20/timestamped_log/3_1_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable20/scripts/3_1_reloadable20.prx) 2>&1 |& tee -a 3_1_reloadable20/3_1_reloadable20.log 3_1_reloadable20/timestamped_log/3_1_reloadable20-${ts}.log + +3_2_reloadable20: 0_0_reloadable20 + mkdir -p 3_2_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 3_2_reloadable20/Release/3_2_reloadable20.o + set -o pipefail; (rm -f "3_2_reloadable20/Release/3_2_reloadable20.##") 2>&1 |& tee -a 3_2_reloadable20/3_2_reloadable20.log 3_2_reloadable20/timestamped_log/3_2_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable20/scripts/3_2_reloadable20.prx) 2>&1 |& tee -a 3_2_reloadable20/3_2_reloadable20.log 3_2_reloadable20/timestamped_log/3_2_reloadable20-${ts}.log + +3_3_reloadable20: 0_0_reloadable20 + mkdir -p 3_3_reloadable20/Release + cp 0_0_reloadable20/Release/0_0_reloadable20.o 3_3_reloadable20/Release/3_3_reloadable20.o + set -o pipefail; (rm -f "3_3_reloadable20/Release/3_3_reloadable20.##") 2>&1 |& tee -a 3_3_reloadable20/3_3_reloadable20.log 3_3_reloadable20/timestamped_log/3_3_reloadable20-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable20/scripts/3_3_reloadable20.prx) 2>&1 |& tee -a 3_3_reloadable20/3_3_reloadable20.log 3_3_reloadable20/timestamped_log/3_3_reloadable20-${ts}.log + +0_1_reloadable21: 0_0_reloadable21 + mkdir -p 0_1_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 0_1_reloadable21/Release/0_1_reloadable21.o + set -o pipefail; (rm -f "0_1_reloadable21/Release/0_1_reloadable21.##") 2>&1 |& tee -a 0_1_reloadable21/0_1_reloadable21.log 0_1_reloadable21/timestamped_log/0_1_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable21/scripts/0_1_reloadable21.prx) 2>&1 |& tee -a 0_1_reloadable21/0_1_reloadable21.log 0_1_reloadable21/timestamped_log/0_1_reloadable21-${ts}.log + +0_2_reloadable21: 0_0_reloadable21 + mkdir -p 0_2_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 0_2_reloadable21/Release/0_2_reloadable21.o + set -o pipefail; (rm -f "0_2_reloadable21/Release/0_2_reloadable21.##") 2>&1 |& tee -a 0_2_reloadable21/0_2_reloadable21.log 0_2_reloadable21/timestamped_log/0_2_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable21/scripts/0_2_reloadable21.prx) 2>&1 |& tee -a 0_2_reloadable21/0_2_reloadable21.log 0_2_reloadable21/timestamped_log/0_2_reloadable21-${ts}.log + +0_3_reloadable21: 0_0_reloadable21 + mkdir -p 0_3_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 0_3_reloadable21/Release/0_3_reloadable21.o + set -o pipefail; (rm -f "0_3_reloadable21/Release/0_3_reloadable21.##") 2>&1 |& tee -a 0_3_reloadable21/0_3_reloadable21.log 0_3_reloadable21/timestamped_log/0_3_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable21/scripts/0_3_reloadable21.prx) 2>&1 |& tee -a 0_3_reloadable21/0_3_reloadable21.log 0_3_reloadable21/timestamped_log/0_3_reloadable21-${ts}.log + +1_0_reloadable21: 0_0_reloadable21 + mkdir -p 1_0_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 1_0_reloadable21/Release/1_0_reloadable21.o + set -o pipefail; (rm -f "1_0_reloadable21/Release/1_0_reloadable21.##") 2>&1 |& tee -a 1_0_reloadable21/1_0_reloadable21.log 1_0_reloadable21/timestamped_log/1_0_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable21/scripts/1_0_reloadable21.prx) 2>&1 |& tee -a 1_0_reloadable21/1_0_reloadable21.log 1_0_reloadable21/timestamped_log/1_0_reloadable21-${ts}.log + +1_1_reloadable21: 0_0_reloadable21 + mkdir -p 1_1_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 1_1_reloadable21/Release/1_1_reloadable21.o + set -o pipefail; (rm -f "1_1_reloadable21/Release/1_1_reloadable21.##") 2>&1 |& tee -a 1_1_reloadable21/1_1_reloadable21.log 1_1_reloadable21/timestamped_log/1_1_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable21/scripts/1_1_reloadable21.prx) 2>&1 |& tee -a 1_1_reloadable21/1_1_reloadable21.log 1_1_reloadable21/timestamped_log/1_1_reloadable21-${ts}.log + +1_2_reloadable21: 0_0_reloadable21 + mkdir -p 1_2_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 1_2_reloadable21/Release/1_2_reloadable21.o + set -o pipefail; (rm -f "1_2_reloadable21/Release/1_2_reloadable21.##") 2>&1 |& tee -a 1_2_reloadable21/1_2_reloadable21.log 1_2_reloadable21/timestamped_log/1_2_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable21/scripts/1_2_reloadable21.prx) 2>&1 |& tee -a 1_2_reloadable21/1_2_reloadable21.log 1_2_reloadable21/timestamped_log/1_2_reloadable21-${ts}.log + +1_3_reloadable21: 0_0_reloadable21 + mkdir -p 1_3_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 1_3_reloadable21/Release/1_3_reloadable21.o + set -o pipefail; (rm -f "1_3_reloadable21/Release/1_3_reloadable21.##") 2>&1 |& tee -a 1_3_reloadable21/1_3_reloadable21.log 1_3_reloadable21/timestamped_log/1_3_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable21/scripts/1_3_reloadable21.prx) 2>&1 |& tee -a 1_3_reloadable21/1_3_reloadable21.log 1_3_reloadable21/timestamped_log/1_3_reloadable21-${ts}.log + +2_0_reloadable21: 0_0_reloadable21 + mkdir -p 2_0_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 2_0_reloadable21/Release/2_0_reloadable21.o + set -o pipefail; (rm -f "2_0_reloadable21/Release/2_0_reloadable21.##") 2>&1 |& tee -a 2_0_reloadable21/2_0_reloadable21.log 2_0_reloadable21/timestamped_log/2_0_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable21/scripts/2_0_reloadable21.prx) 2>&1 |& tee -a 2_0_reloadable21/2_0_reloadable21.log 2_0_reloadable21/timestamped_log/2_0_reloadable21-${ts}.log + +2_1_reloadable21: 0_0_reloadable21 + mkdir -p 2_1_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 2_1_reloadable21/Release/2_1_reloadable21.o + set -o pipefail; (rm -f "2_1_reloadable21/Release/2_1_reloadable21.##") 2>&1 |& tee -a 2_1_reloadable21/2_1_reloadable21.log 2_1_reloadable21/timestamped_log/2_1_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable21/scripts/2_1_reloadable21.prx) 2>&1 |& tee -a 2_1_reloadable21/2_1_reloadable21.log 2_1_reloadable21/timestamped_log/2_1_reloadable21-${ts}.log + +2_2_reloadable21: 0_0_reloadable21 + mkdir -p 2_2_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 2_2_reloadable21/Release/2_2_reloadable21.o + set -o pipefail; (rm -f "2_2_reloadable21/Release/2_2_reloadable21.##") 2>&1 |& tee -a 2_2_reloadable21/2_2_reloadable21.log 2_2_reloadable21/timestamped_log/2_2_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable21/scripts/2_2_reloadable21.prx) 2>&1 |& tee -a 2_2_reloadable21/2_2_reloadable21.log 2_2_reloadable21/timestamped_log/2_2_reloadable21-${ts}.log + +2_3_reloadable21: 0_0_reloadable21 + mkdir -p 2_3_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 2_3_reloadable21/Release/2_3_reloadable21.o + set -o pipefail; (rm -f "2_3_reloadable21/Release/2_3_reloadable21.##") 2>&1 |& tee -a 2_3_reloadable21/2_3_reloadable21.log 2_3_reloadable21/timestamped_log/2_3_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable21/scripts/2_3_reloadable21.prx) 2>&1 |& tee -a 2_3_reloadable21/2_3_reloadable21.log 2_3_reloadable21/timestamped_log/2_3_reloadable21-${ts}.log + +3_0_reloadable21: 0_0_reloadable21 + mkdir -p 3_0_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 3_0_reloadable21/Release/3_0_reloadable21.o + set -o pipefail; (rm -f "3_0_reloadable21/Release/3_0_reloadable21.##") 2>&1 |& tee -a 3_0_reloadable21/3_0_reloadable21.log 3_0_reloadable21/timestamped_log/3_0_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable21/scripts/3_0_reloadable21.prx) 2>&1 |& tee -a 3_0_reloadable21/3_0_reloadable21.log 3_0_reloadable21/timestamped_log/3_0_reloadable21-${ts}.log + +3_1_reloadable21: 0_0_reloadable21 + mkdir -p 3_1_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 3_1_reloadable21/Release/3_1_reloadable21.o + set -o pipefail; (rm -f "3_1_reloadable21/Release/3_1_reloadable21.##") 2>&1 |& tee -a 3_1_reloadable21/3_1_reloadable21.log 3_1_reloadable21/timestamped_log/3_1_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable21/scripts/3_1_reloadable21.prx) 2>&1 |& tee -a 3_1_reloadable21/3_1_reloadable21.log 3_1_reloadable21/timestamped_log/3_1_reloadable21-${ts}.log + +3_2_reloadable21: 0_0_reloadable21 + mkdir -p 3_2_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 3_2_reloadable21/Release/3_2_reloadable21.o + set -o pipefail; (rm -f "3_2_reloadable21/Release/3_2_reloadable21.##") 2>&1 |& tee -a 3_2_reloadable21/3_2_reloadable21.log 3_2_reloadable21/timestamped_log/3_2_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable21/scripts/3_2_reloadable21.prx) 2>&1 |& tee -a 3_2_reloadable21/3_2_reloadable21.log 3_2_reloadable21/timestamped_log/3_2_reloadable21-${ts}.log + +3_3_reloadable21: 0_0_reloadable21 + mkdir -p 3_3_reloadable21/Release + cp 0_0_reloadable21/Release/0_0_reloadable21.o 3_3_reloadable21/Release/3_3_reloadable21.o + set -o pipefail; (rm -f "3_3_reloadable21/Release/3_3_reloadable21.##") 2>&1 |& tee -a 3_3_reloadable21/3_3_reloadable21.log 3_3_reloadable21/timestamped_log/3_3_reloadable21-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable21/scripts/3_3_reloadable21.prx) 2>&1 |& tee -a 3_3_reloadable21/3_3_reloadable21.log 3_3_reloadable21/timestamped_log/3_3_reloadable21-${ts}.log + +0_1_reloadable22: 0_0_reloadable22 + mkdir -p 0_1_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 0_1_reloadable22/Release/0_1_reloadable22.o + set -o pipefail; (rm -f "0_1_reloadable22/Release/0_1_reloadable22.##") 2>&1 |& tee -a 0_1_reloadable22/0_1_reloadable22.log 0_1_reloadable22/timestamped_log/0_1_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable22/scripts/0_1_reloadable22.prx) 2>&1 |& tee -a 0_1_reloadable22/0_1_reloadable22.log 0_1_reloadable22/timestamped_log/0_1_reloadable22-${ts}.log + +0_2_reloadable22: 0_0_reloadable22 + mkdir -p 0_2_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 0_2_reloadable22/Release/0_2_reloadable22.o + set -o pipefail; (rm -f "0_2_reloadable22/Release/0_2_reloadable22.##") 2>&1 |& tee -a 0_2_reloadable22/0_2_reloadable22.log 0_2_reloadable22/timestamped_log/0_2_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable22/scripts/0_2_reloadable22.prx) 2>&1 |& tee -a 0_2_reloadable22/0_2_reloadable22.log 0_2_reloadable22/timestamped_log/0_2_reloadable22-${ts}.log + +0_3_reloadable22: 0_0_reloadable22 + mkdir -p 0_3_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 0_3_reloadable22/Release/0_3_reloadable22.o + set -o pipefail; (rm -f "0_3_reloadable22/Release/0_3_reloadable22.##") 2>&1 |& tee -a 0_3_reloadable22/0_3_reloadable22.log 0_3_reloadable22/timestamped_log/0_3_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable22/scripts/0_3_reloadable22.prx) 2>&1 |& tee -a 0_3_reloadable22/0_3_reloadable22.log 0_3_reloadable22/timestamped_log/0_3_reloadable22-${ts}.log + +1_0_reloadable22: 0_0_reloadable22 + mkdir -p 1_0_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 1_0_reloadable22/Release/1_0_reloadable22.o + set -o pipefail; (rm -f "1_0_reloadable22/Release/1_0_reloadable22.##") 2>&1 |& tee -a 1_0_reloadable22/1_0_reloadable22.log 1_0_reloadable22/timestamped_log/1_0_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable22/scripts/1_0_reloadable22.prx) 2>&1 |& tee -a 1_0_reloadable22/1_0_reloadable22.log 1_0_reloadable22/timestamped_log/1_0_reloadable22-${ts}.log + +1_1_reloadable22: 0_0_reloadable22 + mkdir -p 1_1_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 1_1_reloadable22/Release/1_1_reloadable22.o + set -o pipefail; (rm -f "1_1_reloadable22/Release/1_1_reloadable22.##") 2>&1 |& tee -a 1_1_reloadable22/1_1_reloadable22.log 1_1_reloadable22/timestamped_log/1_1_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable22/scripts/1_1_reloadable22.prx) 2>&1 |& tee -a 1_1_reloadable22/1_1_reloadable22.log 1_1_reloadable22/timestamped_log/1_1_reloadable22-${ts}.log + +1_2_reloadable22: 0_0_reloadable22 + mkdir -p 1_2_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 1_2_reloadable22/Release/1_2_reloadable22.o + set -o pipefail; (rm -f "1_2_reloadable22/Release/1_2_reloadable22.##") 2>&1 |& tee -a 1_2_reloadable22/1_2_reloadable22.log 1_2_reloadable22/timestamped_log/1_2_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable22/scripts/1_2_reloadable22.prx) 2>&1 |& tee -a 1_2_reloadable22/1_2_reloadable22.log 1_2_reloadable22/timestamped_log/1_2_reloadable22-${ts}.log + +1_3_reloadable22: 0_0_reloadable22 + mkdir -p 1_3_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 1_3_reloadable22/Release/1_3_reloadable22.o + set -o pipefail; (rm -f "1_3_reloadable22/Release/1_3_reloadable22.##") 2>&1 |& tee -a 1_3_reloadable22/1_3_reloadable22.log 1_3_reloadable22/timestamped_log/1_3_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable22/scripts/1_3_reloadable22.prx) 2>&1 |& tee -a 1_3_reloadable22/1_3_reloadable22.log 1_3_reloadable22/timestamped_log/1_3_reloadable22-${ts}.log + +2_0_reloadable22: 0_0_reloadable22 + mkdir -p 2_0_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 2_0_reloadable22/Release/2_0_reloadable22.o + set -o pipefail; (rm -f "2_0_reloadable22/Release/2_0_reloadable22.##") 2>&1 |& tee -a 2_0_reloadable22/2_0_reloadable22.log 2_0_reloadable22/timestamped_log/2_0_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable22/scripts/2_0_reloadable22.prx) 2>&1 |& tee -a 2_0_reloadable22/2_0_reloadable22.log 2_0_reloadable22/timestamped_log/2_0_reloadable22-${ts}.log + +2_1_reloadable22: 0_0_reloadable22 + mkdir -p 2_1_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 2_1_reloadable22/Release/2_1_reloadable22.o + set -o pipefail; (rm -f "2_1_reloadable22/Release/2_1_reloadable22.##") 2>&1 |& tee -a 2_1_reloadable22/2_1_reloadable22.log 2_1_reloadable22/timestamped_log/2_1_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable22/scripts/2_1_reloadable22.prx) 2>&1 |& tee -a 2_1_reloadable22/2_1_reloadable22.log 2_1_reloadable22/timestamped_log/2_1_reloadable22-${ts}.log + +2_2_reloadable22: 0_0_reloadable22 + mkdir -p 2_2_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 2_2_reloadable22/Release/2_2_reloadable22.o + set -o pipefail; (rm -f "2_2_reloadable22/Release/2_2_reloadable22.##") 2>&1 |& tee -a 2_2_reloadable22/2_2_reloadable22.log 2_2_reloadable22/timestamped_log/2_2_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable22/scripts/2_2_reloadable22.prx) 2>&1 |& tee -a 2_2_reloadable22/2_2_reloadable22.log 2_2_reloadable22/timestamped_log/2_2_reloadable22-${ts}.log + +2_3_reloadable22: 0_0_reloadable22 + mkdir -p 2_3_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 2_3_reloadable22/Release/2_3_reloadable22.o + set -o pipefail; (rm -f "2_3_reloadable22/Release/2_3_reloadable22.##") 2>&1 |& tee -a 2_3_reloadable22/2_3_reloadable22.log 2_3_reloadable22/timestamped_log/2_3_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable22/scripts/2_3_reloadable22.prx) 2>&1 |& tee -a 2_3_reloadable22/2_3_reloadable22.log 2_3_reloadable22/timestamped_log/2_3_reloadable22-${ts}.log + +3_0_reloadable22: 0_0_reloadable22 + mkdir -p 3_0_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 3_0_reloadable22/Release/3_0_reloadable22.o + set -o pipefail; (rm -f "3_0_reloadable22/Release/3_0_reloadable22.##") 2>&1 |& tee -a 3_0_reloadable22/3_0_reloadable22.log 3_0_reloadable22/timestamped_log/3_0_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable22/scripts/3_0_reloadable22.prx) 2>&1 |& tee -a 3_0_reloadable22/3_0_reloadable22.log 3_0_reloadable22/timestamped_log/3_0_reloadable22-${ts}.log + +3_1_reloadable22: 0_0_reloadable22 + mkdir -p 3_1_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 3_1_reloadable22/Release/3_1_reloadable22.o + set -o pipefail; (rm -f "3_1_reloadable22/Release/3_1_reloadable22.##") 2>&1 |& tee -a 3_1_reloadable22/3_1_reloadable22.log 3_1_reloadable22/timestamped_log/3_1_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable22/scripts/3_1_reloadable22.prx) 2>&1 |& tee -a 3_1_reloadable22/3_1_reloadable22.log 3_1_reloadable22/timestamped_log/3_1_reloadable22-${ts}.log + +3_2_reloadable22: 0_0_reloadable22 + mkdir -p 3_2_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 3_2_reloadable22/Release/3_2_reloadable22.o + set -o pipefail; (rm -f "3_2_reloadable22/Release/3_2_reloadable22.##") 2>&1 |& tee -a 3_2_reloadable22/3_2_reloadable22.log 3_2_reloadable22/timestamped_log/3_2_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable22/scripts/3_2_reloadable22.prx) 2>&1 |& tee -a 3_2_reloadable22/3_2_reloadable22.log 3_2_reloadable22/timestamped_log/3_2_reloadable22-${ts}.log + +3_3_reloadable22: 0_0_reloadable22 + mkdir -p 3_3_reloadable22/Release + cp 0_0_reloadable22/Release/0_0_reloadable22.o 3_3_reloadable22/Release/3_3_reloadable22.o + set -o pipefail; (rm -f "3_3_reloadable22/Release/3_3_reloadable22.##") 2>&1 |& tee -a 3_3_reloadable22/3_3_reloadable22.log 3_3_reloadable22/timestamped_log/3_3_reloadable22-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable22/scripts/3_3_reloadable22.prx) 2>&1 |& tee -a 3_3_reloadable22/3_3_reloadable22.log 3_3_reloadable22/timestamped_log/3_3_reloadable22-${ts}.log + +0_1_reloadable23: 0_0_reloadable23 + mkdir -p 0_1_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 0_1_reloadable23/Release/0_1_reloadable23.o + set -o pipefail; (rm -f "0_1_reloadable23/Release/0_1_reloadable23.##") 2>&1 |& tee -a 0_1_reloadable23/0_1_reloadable23.log 0_1_reloadable23/timestamped_log/0_1_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable23/scripts/0_1_reloadable23.prx) 2>&1 |& tee -a 0_1_reloadable23/0_1_reloadable23.log 0_1_reloadable23/timestamped_log/0_1_reloadable23-${ts}.log + +0_2_reloadable23: 0_0_reloadable23 + mkdir -p 0_2_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 0_2_reloadable23/Release/0_2_reloadable23.o + set -o pipefail; (rm -f "0_2_reloadable23/Release/0_2_reloadable23.##") 2>&1 |& tee -a 0_2_reloadable23/0_2_reloadable23.log 0_2_reloadable23/timestamped_log/0_2_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable23/scripts/0_2_reloadable23.prx) 2>&1 |& tee -a 0_2_reloadable23/0_2_reloadable23.log 0_2_reloadable23/timestamped_log/0_2_reloadable23-${ts}.log + +0_3_reloadable23: 0_0_reloadable23 + mkdir -p 0_3_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 0_3_reloadable23/Release/0_3_reloadable23.o + set -o pipefail; (rm -f "0_3_reloadable23/Release/0_3_reloadable23.##") 2>&1 |& tee -a 0_3_reloadable23/0_3_reloadable23.log 0_3_reloadable23/timestamped_log/0_3_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable23/scripts/0_3_reloadable23.prx) 2>&1 |& tee -a 0_3_reloadable23/0_3_reloadable23.log 0_3_reloadable23/timestamped_log/0_3_reloadable23-${ts}.log + +1_0_reloadable23: 0_0_reloadable23 + mkdir -p 1_0_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 1_0_reloadable23/Release/1_0_reloadable23.o + set -o pipefail; (rm -f "1_0_reloadable23/Release/1_0_reloadable23.##") 2>&1 |& tee -a 1_0_reloadable23/1_0_reloadable23.log 1_0_reloadable23/timestamped_log/1_0_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable23/scripts/1_0_reloadable23.prx) 2>&1 |& tee -a 1_0_reloadable23/1_0_reloadable23.log 1_0_reloadable23/timestamped_log/1_0_reloadable23-${ts}.log + +1_1_reloadable23: 0_0_reloadable23 + mkdir -p 1_1_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 1_1_reloadable23/Release/1_1_reloadable23.o + set -o pipefail; (rm -f "1_1_reloadable23/Release/1_1_reloadable23.##") 2>&1 |& tee -a 1_1_reloadable23/1_1_reloadable23.log 1_1_reloadable23/timestamped_log/1_1_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable23/scripts/1_1_reloadable23.prx) 2>&1 |& tee -a 1_1_reloadable23/1_1_reloadable23.log 1_1_reloadable23/timestamped_log/1_1_reloadable23-${ts}.log + +1_2_reloadable23: 0_0_reloadable23 + mkdir -p 1_2_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 1_2_reloadable23/Release/1_2_reloadable23.o + set -o pipefail; (rm -f "1_2_reloadable23/Release/1_2_reloadable23.##") 2>&1 |& tee -a 1_2_reloadable23/1_2_reloadable23.log 1_2_reloadable23/timestamped_log/1_2_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable23/scripts/1_2_reloadable23.prx) 2>&1 |& tee -a 1_2_reloadable23/1_2_reloadable23.log 1_2_reloadable23/timestamped_log/1_2_reloadable23-${ts}.log + +1_3_reloadable23: 0_0_reloadable23 + mkdir -p 1_3_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 1_3_reloadable23/Release/1_3_reloadable23.o + set -o pipefail; (rm -f "1_3_reloadable23/Release/1_3_reloadable23.##") 2>&1 |& tee -a 1_3_reloadable23/1_3_reloadable23.log 1_3_reloadable23/timestamped_log/1_3_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable23/scripts/1_3_reloadable23.prx) 2>&1 |& tee -a 1_3_reloadable23/1_3_reloadable23.log 1_3_reloadable23/timestamped_log/1_3_reloadable23-${ts}.log + +2_0_reloadable23: 0_0_reloadable23 + mkdir -p 2_0_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 2_0_reloadable23/Release/2_0_reloadable23.o + set -o pipefail; (rm -f "2_0_reloadable23/Release/2_0_reloadable23.##") 2>&1 |& tee -a 2_0_reloadable23/2_0_reloadable23.log 2_0_reloadable23/timestamped_log/2_0_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable23/scripts/2_0_reloadable23.prx) 2>&1 |& tee -a 2_0_reloadable23/2_0_reloadable23.log 2_0_reloadable23/timestamped_log/2_0_reloadable23-${ts}.log + +2_1_reloadable23: 0_0_reloadable23 + mkdir -p 2_1_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 2_1_reloadable23/Release/2_1_reloadable23.o + set -o pipefail; (rm -f "2_1_reloadable23/Release/2_1_reloadable23.##") 2>&1 |& tee -a 2_1_reloadable23/2_1_reloadable23.log 2_1_reloadable23/timestamped_log/2_1_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable23/scripts/2_1_reloadable23.prx) 2>&1 |& tee -a 2_1_reloadable23/2_1_reloadable23.log 2_1_reloadable23/timestamped_log/2_1_reloadable23-${ts}.log + +2_2_reloadable23: 0_0_reloadable23 + mkdir -p 2_2_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 2_2_reloadable23/Release/2_2_reloadable23.o + set -o pipefail; (rm -f "2_2_reloadable23/Release/2_2_reloadable23.##") 2>&1 |& tee -a 2_2_reloadable23/2_2_reloadable23.log 2_2_reloadable23/timestamped_log/2_2_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable23/scripts/2_2_reloadable23.prx) 2>&1 |& tee -a 2_2_reloadable23/2_2_reloadable23.log 2_2_reloadable23/timestamped_log/2_2_reloadable23-${ts}.log + +2_3_reloadable23: 0_0_reloadable23 + mkdir -p 2_3_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 2_3_reloadable23/Release/2_3_reloadable23.o + set -o pipefail; (rm -f "2_3_reloadable23/Release/2_3_reloadable23.##") 2>&1 |& tee -a 2_3_reloadable23/2_3_reloadable23.log 2_3_reloadable23/timestamped_log/2_3_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable23/scripts/2_3_reloadable23.prx) 2>&1 |& tee -a 2_3_reloadable23/2_3_reloadable23.log 2_3_reloadable23/timestamped_log/2_3_reloadable23-${ts}.log + +3_0_reloadable23: 0_0_reloadable23 + mkdir -p 3_0_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 3_0_reloadable23/Release/3_0_reloadable23.o + set -o pipefail; (rm -f "3_0_reloadable23/Release/3_0_reloadable23.##") 2>&1 |& tee -a 3_0_reloadable23/3_0_reloadable23.log 3_0_reloadable23/timestamped_log/3_0_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable23/scripts/3_0_reloadable23.prx) 2>&1 |& tee -a 3_0_reloadable23/3_0_reloadable23.log 3_0_reloadable23/timestamped_log/3_0_reloadable23-${ts}.log + +3_1_reloadable23: 0_0_reloadable23 + mkdir -p 3_1_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 3_1_reloadable23/Release/3_1_reloadable23.o + set -o pipefail; (rm -f "3_1_reloadable23/Release/3_1_reloadable23.##") 2>&1 |& tee -a 3_1_reloadable23/3_1_reloadable23.log 3_1_reloadable23/timestamped_log/3_1_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable23/scripts/3_1_reloadable23.prx) 2>&1 |& tee -a 3_1_reloadable23/3_1_reloadable23.log 3_1_reloadable23/timestamped_log/3_1_reloadable23-${ts}.log + +3_2_reloadable23: 0_0_reloadable23 + mkdir -p 3_2_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 3_2_reloadable23/Release/3_2_reloadable23.o + set -o pipefail; (rm -f "3_2_reloadable23/Release/3_2_reloadable23.##") 2>&1 |& tee -a 3_2_reloadable23/3_2_reloadable23.log 3_2_reloadable23/timestamped_log/3_2_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable23/scripts/3_2_reloadable23.prx) 2>&1 |& tee -a 3_2_reloadable23/3_2_reloadable23.log 3_2_reloadable23/timestamped_log/3_2_reloadable23-${ts}.log + +3_3_reloadable23: 0_0_reloadable23 + mkdir -p 3_3_reloadable23/Release + cp 0_0_reloadable23/Release/0_0_reloadable23.o 3_3_reloadable23/Release/3_3_reloadable23.o + set -o pipefail; (rm -f "3_3_reloadable23/Release/3_3_reloadable23.##") 2>&1 |& tee -a 3_3_reloadable23/3_3_reloadable23.log 3_3_reloadable23/timestamped_log/3_3_reloadable23-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable23/scripts/3_3_reloadable23.prx) 2>&1 |& tee -a 3_3_reloadable23/3_3_reloadable23.log 3_3_reloadable23/timestamped_log/3_3_reloadable23-${ts}.log + +0_1_reloadable24: 0_0_reloadable24 + mkdir -p 0_1_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 0_1_reloadable24/Release/0_1_reloadable24.o + set -o pipefail; (rm -f "0_1_reloadable24/Release/0_1_reloadable24.##") 2>&1 |& tee -a 0_1_reloadable24/0_1_reloadable24.log 0_1_reloadable24/timestamped_log/0_1_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable24/scripts/0_1_reloadable24.prx) 2>&1 |& tee -a 0_1_reloadable24/0_1_reloadable24.log 0_1_reloadable24/timestamped_log/0_1_reloadable24-${ts}.log + +0_2_reloadable24: 0_0_reloadable24 + mkdir -p 0_2_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 0_2_reloadable24/Release/0_2_reloadable24.o + set -o pipefail; (rm -f "0_2_reloadable24/Release/0_2_reloadable24.##") 2>&1 |& tee -a 0_2_reloadable24/0_2_reloadable24.log 0_2_reloadable24/timestamped_log/0_2_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable24/scripts/0_2_reloadable24.prx) 2>&1 |& tee -a 0_2_reloadable24/0_2_reloadable24.log 0_2_reloadable24/timestamped_log/0_2_reloadable24-${ts}.log + +0_3_reloadable24: 0_0_reloadable24 + mkdir -p 0_3_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 0_3_reloadable24/Release/0_3_reloadable24.o + set -o pipefail; (rm -f "0_3_reloadable24/Release/0_3_reloadable24.##") 2>&1 |& tee -a 0_3_reloadable24/0_3_reloadable24.log 0_3_reloadable24/timestamped_log/0_3_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable24/scripts/0_3_reloadable24.prx) 2>&1 |& tee -a 0_3_reloadable24/0_3_reloadable24.log 0_3_reloadable24/timestamped_log/0_3_reloadable24-${ts}.log + +1_0_reloadable24: 0_0_reloadable24 + mkdir -p 1_0_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 1_0_reloadable24/Release/1_0_reloadable24.o + set -o pipefail; (rm -f "1_0_reloadable24/Release/1_0_reloadable24.##") 2>&1 |& tee -a 1_0_reloadable24/1_0_reloadable24.log 1_0_reloadable24/timestamped_log/1_0_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable24/scripts/1_0_reloadable24.prx) 2>&1 |& tee -a 1_0_reloadable24/1_0_reloadable24.log 1_0_reloadable24/timestamped_log/1_0_reloadable24-${ts}.log + +1_1_reloadable24: 0_0_reloadable24 + mkdir -p 1_1_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 1_1_reloadable24/Release/1_1_reloadable24.o + set -o pipefail; (rm -f "1_1_reloadable24/Release/1_1_reloadable24.##") 2>&1 |& tee -a 1_1_reloadable24/1_1_reloadable24.log 1_1_reloadable24/timestamped_log/1_1_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable24/scripts/1_1_reloadable24.prx) 2>&1 |& tee -a 1_1_reloadable24/1_1_reloadable24.log 1_1_reloadable24/timestamped_log/1_1_reloadable24-${ts}.log + +1_2_reloadable24: 0_0_reloadable24 + mkdir -p 1_2_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 1_2_reloadable24/Release/1_2_reloadable24.o + set -o pipefail; (rm -f "1_2_reloadable24/Release/1_2_reloadable24.##") 2>&1 |& tee -a 1_2_reloadable24/1_2_reloadable24.log 1_2_reloadable24/timestamped_log/1_2_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable24/scripts/1_2_reloadable24.prx) 2>&1 |& tee -a 1_2_reloadable24/1_2_reloadable24.log 1_2_reloadable24/timestamped_log/1_2_reloadable24-${ts}.log + +1_3_reloadable24: 0_0_reloadable24 + mkdir -p 1_3_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 1_3_reloadable24/Release/1_3_reloadable24.o + set -o pipefail; (rm -f "1_3_reloadable24/Release/1_3_reloadable24.##") 2>&1 |& tee -a 1_3_reloadable24/1_3_reloadable24.log 1_3_reloadable24/timestamped_log/1_3_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable24/scripts/1_3_reloadable24.prx) 2>&1 |& tee -a 1_3_reloadable24/1_3_reloadable24.log 1_3_reloadable24/timestamped_log/1_3_reloadable24-${ts}.log + +2_0_reloadable24: 0_0_reloadable24 + mkdir -p 2_0_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 2_0_reloadable24/Release/2_0_reloadable24.o + set -o pipefail; (rm -f "2_0_reloadable24/Release/2_0_reloadable24.##") 2>&1 |& tee -a 2_0_reloadable24/2_0_reloadable24.log 2_0_reloadable24/timestamped_log/2_0_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable24/scripts/2_0_reloadable24.prx) 2>&1 |& tee -a 2_0_reloadable24/2_0_reloadable24.log 2_0_reloadable24/timestamped_log/2_0_reloadable24-${ts}.log + +2_1_reloadable24: 0_0_reloadable24 + mkdir -p 2_1_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 2_1_reloadable24/Release/2_1_reloadable24.o + set -o pipefail; (rm -f "2_1_reloadable24/Release/2_1_reloadable24.##") 2>&1 |& tee -a 2_1_reloadable24/2_1_reloadable24.log 2_1_reloadable24/timestamped_log/2_1_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable24/scripts/2_1_reloadable24.prx) 2>&1 |& tee -a 2_1_reloadable24/2_1_reloadable24.log 2_1_reloadable24/timestamped_log/2_1_reloadable24-${ts}.log + +2_2_reloadable24: 0_0_reloadable24 + mkdir -p 2_2_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 2_2_reloadable24/Release/2_2_reloadable24.o + set -o pipefail; (rm -f "2_2_reloadable24/Release/2_2_reloadable24.##") 2>&1 |& tee -a 2_2_reloadable24/2_2_reloadable24.log 2_2_reloadable24/timestamped_log/2_2_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable24/scripts/2_2_reloadable24.prx) 2>&1 |& tee -a 2_2_reloadable24/2_2_reloadable24.log 2_2_reloadable24/timestamped_log/2_2_reloadable24-${ts}.log + +2_3_reloadable24: 0_0_reloadable24 + mkdir -p 2_3_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 2_3_reloadable24/Release/2_3_reloadable24.o + set -o pipefail; (rm -f "2_3_reloadable24/Release/2_3_reloadable24.##") 2>&1 |& tee -a 2_3_reloadable24/2_3_reloadable24.log 2_3_reloadable24/timestamped_log/2_3_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable24/scripts/2_3_reloadable24.prx) 2>&1 |& tee -a 2_3_reloadable24/2_3_reloadable24.log 2_3_reloadable24/timestamped_log/2_3_reloadable24-${ts}.log + +3_0_reloadable24: 0_0_reloadable24 + mkdir -p 3_0_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 3_0_reloadable24/Release/3_0_reloadable24.o + set -o pipefail; (rm -f "3_0_reloadable24/Release/3_0_reloadable24.##") 2>&1 |& tee -a 3_0_reloadable24/3_0_reloadable24.log 3_0_reloadable24/timestamped_log/3_0_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable24/scripts/3_0_reloadable24.prx) 2>&1 |& tee -a 3_0_reloadable24/3_0_reloadable24.log 3_0_reloadable24/timestamped_log/3_0_reloadable24-${ts}.log + +3_1_reloadable24: 0_0_reloadable24 + mkdir -p 3_1_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 3_1_reloadable24/Release/3_1_reloadable24.o + set -o pipefail; (rm -f "3_1_reloadable24/Release/3_1_reloadable24.##") 2>&1 |& tee -a 3_1_reloadable24/3_1_reloadable24.log 3_1_reloadable24/timestamped_log/3_1_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable24/scripts/3_1_reloadable24.prx) 2>&1 |& tee -a 3_1_reloadable24/3_1_reloadable24.log 3_1_reloadable24/timestamped_log/3_1_reloadable24-${ts}.log + +3_2_reloadable24: 0_0_reloadable24 + mkdir -p 3_2_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 3_2_reloadable24/Release/3_2_reloadable24.o + set -o pipefail; (rm -f "3_2_reloadable24/Release/3_2_reloadable24.##") 2>&1 |& tee -a 3_2_reloadable24/3_2_reloadable24.log 3_2_reloadable24/timestamped_log/3_2_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable24/scripts/3_2_reloadable24.prx) 2>&1 |& tee -a 3_2_reloadable24/3_2_reloadable24.log 3_2_reloadable24/timestamped_log/3_2_reloadable24-${ts}.log + +3_3_reloadable24: 0_0_reloadable24 + mkdir -p 3_3_reloadable24/Release + cp 0_0_reloadable24/Release/0_0_reloadable24.o 3_3_reloadable24/Release/3_3_reloadable24.o + set -o pipefail; (rm -f "3_3_reloadable24/Release/3_3_reloadable24.##") 2>&1 |& tee -a 3_3_reloadable24/3_3_reloadable24.log 3_3_reloadable24/timestamped_log/3_3_reloadable24-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable24/scripts/3_3_reloadable24.prx) 2>&1 |& tee -a 3_3_reloadable24/3_3_reloadable24.log 3_3_reloadable24/timestamped_log/3_3_reloadable24-${ts}.log + +0_1_reloadable25: 0_0_reloadable25 + mkdir -p 0_1_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 0_1_reloadable25/Release/0_1_reloadable25.o + set -o pipefail; (rm -f "0_1_reloadable25/Release/0_1_reloadable25.##") 2>&1 |& tee -a 0_1_reloadable25/0_1_reloadable25.log 0_1_reloadable25/timestamped_log/0_1_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_1_reloadable25/scripts/0_1_reloadable25.prx) 2>&1 |& tee -a 0_1_reloadable25/0_1_reloadable25.log 0_1_reloadable25/timestamped_log/0_1_reloadable25-${ts}.log + +0_2_reloadable25: 0_0_reloadable25 + mkdir -p 0_2_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 0_2_reloadable25/Release/0_2_reloadable25.o + set -o pipefail; (rm -f "0_2_reloadable25/Release/0_2_reloadable25.##") 2>&1 |& tee -a 0_2_reloadable25/0_2_reloadable25.log 0_2_reloadable25/timestamped_log/0_2_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_2_reloadable25/scripts/0_2_reloadable25.prx) 2>&1 |& tee -a 0_2_reloadable25/0_2_reloadable25.log 0_2_reloadable25/timestamped_log/0_2_reloadable25-${ts}.log + +0_3_reloadable25: 0_0_reloadable25 + mkdir -p 0_3_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 0_3_reloadable25/Release/0_3_reloadable25.o + set -o pipefail; (rm -f "0_3_reloadable25/Release/0_3_reloadable25.##") 2>&1 |& tee -a 0_3_reloadable25/0_3_reloadable25.log 0_3_reloadable25/timestamped_log/0_3_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 0_3_reloadable25/scripts/0_3_reloadable25.prx) 2>&1 |& tee -a 0_3_reloadable25/0_3_reloadable25.log 0_3_reloadable25/timestamped_log/0_3_reloadable25-${ts}.log + +1_0_reloadable25: 0_0_reloadable25 + mkdir -p 1_0_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 1_0_reloadable25/Release/1_0_reloadable25.o + set -o pipefail; (rm -f "1_0_reloadable25/Release/1_0_reloadable25.##") 2>&1 |& tee -a 1_0_reloadable25/1_0_reloadable25.log 1_0_reloadable25/timestamped_log/1_0_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_0_reloadable25/scripts/1_0_reloadable25.prx) 2>&1 |& tee -a 1_0_reloadable25/1_0_reloadable25.log 1_0_reloadable25/timestamped_log/1_0_reloadable25-${ts}.log + +1_1_reloadable25: 0_0_reloadable25 + mkdir -p 1_1_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 1_1_reloadable25/Release/1_1_reloadable25.o + set -o pipefail; (rm -f "1_1_reloadable25/Release/1_1_reloadable25.##") 2>&1 |& tee -a 1_1_reloadable25/1_1_reloadable25.log 1_1_reloadable25/timestamped_log/1_1_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_1_reloadable25/scripts/1_1_reloadable25.prx) 2>&1 |& tee -a 1_1_reloadable25/1_1_reloadable25.log 1_1_reloadable25/timestamped_log/1_1_reloadable25-${ts}.log + +1_2_reloadable25: 0_0_reloadable25 + mkdir -p 1_2_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 1_2_reloadable25/Release/1_2_reloadable25.o + set -o pipefail; (rm -f "1_2_reloadable25/Release/1_2_reloadable25.##") 2>&1 |& tee -a 1_2_reloadable25/1_2_reloadable25.log 1_2_reloadable25/timestamped_log/1_2_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_2_reloadable25/scripts/1_2_reloadable25.prx) 2>&1 |& tee -a 1_2_reloadable25/1_2_reloadable25.log 1_2_reloadable25/timestamped_log/1_2_reloadable25-${ts}.log + +1_3_reloadable25: 0_0_reloadable25 + mkdir -p 1_3_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 1_3_reloadable25/Release/1_3_reloadable25.o + set -o pipefail; (rm -f "1_3_reloadable25/Release/1_3_reloadable25.##") 2>&1 |& tee -a 1_3_reloadable25/1_3_reloadable25.log 1_3_reloadable25/timestamped_log/1_3_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 1_3_reloadable25/scripts/1_3_reloadable25.prx) 2>&1 |& tee -a 1_3_reloadable25/1_3_reloadable25.log 1_3_reloadable25/timestamped_log/1_3_reloadable25-${ts}.log + +2_0_reloadable25: 0_0_reloadable25 + mkdir -p 2_0_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 2_0_reloadable25/Release/2_0_reloadable25.o + set -o pipefail; (rm -f "2_0_reloadable25/Release/2_0_reloadable25.##") 2>&1 |& tee -a 2_0_reloadable25/2_0_reloadable25.log 2_0_reloadable25/timestamped_log/2_0_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_0_reloadable25/scripts/2_0_reloadable25.prx) 2>&1 |& tee -a 2_0_reloadable25/2_0_reloadable25.log 2_0_reloadable25/timestamped_log/2_0_reloadable25-${ts}.log + +2_1_reloadable25: 0_0_reloadable25 + mkdir -p 2_1_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 2_1_reloadable25/Release/2_1_reloadable25.o + set -o pipefail; (rm -f "2_1_reloadable25/Release/2_1_reloadable25.##") 2>&1 |& tee -a 2_1_reloadable25/2_1_reloadable25.log 2_1_reloadable25/timestamped_log/2_1_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_1_reloadable25/scripts/2_1_reloadable25.prx) 2>&1 |& tee -a 2_1_reloadable25/2_1_reloadable25.log 2_1_reloadable25/timestamped_log/2_1_reloadable25-${ts}.log + +2_2_reloadable25: 0_0_reloadable25 + mkdir -p 2_2_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 2_2_reloadable25/Release/2_2_reloadable25.o + set -o pipefail; (rm -f "2_2_reloadable25/Release/2_2_reloadable25.##") 2>&1 |& tee -a 2_2_reloadable25/2_2_reloadable25.log 2_2_reloadable25/timestamped_log/2_2_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_2_reloadable25/scripts/2_2_reloadable25.prx) 2>&1 |& tee -a 2_2_reloadable25/2_2_reloadable25.log 2_2_reloadable25/timestamped_log/2_2_reloadable25-${ts}.log + +2_3_reloadable25: 0_0_reloadable25 + mkdir -p 2_3_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 2_3_reloadable25/Release/2_3_reloadable25.o + set -o pipefail; (rm -f "2_3_reloadable25/Release/2_3_reloadable25.##") 2>&1 |& tee -a 2_3_reloadable25/2_3_reloadable25.log 2_3_reloadable25/timestamped_log/2_3_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 2_3_reloadable25/scripts/2_3_reloadable25.prx) 2>&1 |& tee -a 2_3_reloadable25/2_3_reloadable25.log 2_3_reloadable25/timestamped_log/2_3_reloadable25-${ts}.log + +3_0_reloadable25: 0_0_reloadable25 + mkdir -p 3_0_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 3_0_reloadable25/Release/3_0_reloadable25.o + set -o pipefail; (rm -f "3_0_reloadable25/Release/3_0_reloadable25.##") 2>&1 |& tee -a 3_0_reloadable25/3_0_reloadable25.log 3_0_reloadable25/timestamped_log/3_0_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_0_reloadable25/scripts/3_0_reloadable25.prx) 2>&1 |& tee -a 3_0_reloadable25/3_0_reloadable25.log 3_0_reloadable25/timestamped_log/3_0_reloadable25-${ts}.log + +3_1_reloadable25: 0_0_reloadable25 + mkdir -p 3_1_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 3_1_reloadable25/Release/3_1_reloadable25.o + set -o pipefail; (rm -f "3_1_reloadable25/Release/3_1_reloadable25.##") 2>&1 |& tee -a 3_1_reloadable25/3_1_reloadable25.log 3_1_reloadable25/timestamped_log/3_1_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_1_reloadable25/scripts/3_1_reloadable25.prx) 2>&1 |& tee -a 3_1_reloadable25/3_1_reloadable25.log 3_1_reloadable25/timestamped_log/3_1_reloadable25-${ts}.log + +3_2_reloadable25: 0_0_reloadable25 + mkdir -p 3_2_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 3_2_reloadable25/Release/3_2_reloadable25.o + set -o pipefail; (rm -f "3_2_reloadable25/Release/3_2_reloadable25.##") 2>&1 |& tee -a 3_2_reloadable25/3_2_reloadable25.log 3_2_reloadable25/timestamped_log/3_2_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_2_reloadable25/scripts/3_2_reloadable25.prx) 2>&1 |& tee -a 3_2_reloadable25/3_2_reloadable25.log 3_2_reloadable25/timestamped_log/3_2_reloadable25-${ts}.log + +3_3_reloadable25: 0_0_reloadable25 + mkdir -p 3_3_reloadable25/Release + cp 0_0_reloadable25/Release/0_0_reloadable25.o 3_3_reloadable25/Release/3_3_reloadable25.o + set -o pipefail; (rm -f "3_3_reloadable25/Release/3_3_reloadable25.##") 2>&1 |& tee -a 3_3_reloadable25/3_3_reloadable25.log 3_3_reloadable25/timestamped_log/3_3_reloadable25-${ts}.log + set -o pipefail; (${XCHESSMK} -C Release_LLVM -P ${CARDANO_AIE_ARCH_MODEL_DIR} +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +o ../Release 3_3_reloadable25/scripts/3_3_reloadable25.prx) 2>&1 |& tee -a 3_3_reloadable25/3_3_reloadable25.log 3_3_reloadable25/timestamped_log/3_3_reloadable25-${ts}.log + +clean: + (rm -rf 0_0/Release) + (rm -rf 0_0_reloadable0/Release) + (rm -rf 0_0_reloadable1/Release) + (rm -rf 0_0_reloadable2/Release) + (rm -rf 0_0_reloadable3/Release) + (rm -rf 0_0_reloadable4/Release) + (rm -rf 0_0_reloadable5/Release) + (rm -rf 0_0_reloadable6/Release) + (rm -rf 0_0_reloadable7/Release) + (rm -rf 0_0_reloadable8/Release) + (rm -rf 0_0_reloadable9/Release) + (rm -rf 0_0_reloadable10/Release) + (rm -rf 0_0_reloadable11/Release) + (rm -rf 0_0_reloadable12/Release) + (rm -rf 0_0_reloadable13/Release) + (rm -rf 0_0_reloadable14/Release) + (rm -rf 0_0_reloadable15/Release) + (rm -rf 0_0_reloadable16/Release) + (rm -rf 0_0_reloadable17/Release) + (rm -rf 0_0_reloadable18/Release) + (rm -rf 0_0_reloadable19/Release) + (rm -rf 0_0_reloadable20/Release) + (rm -rf 0_0_reloadable21/Release) + (rm -rf 0_0_reloadable22/Release) + (rm -rf 0_0_reloadable23/Release) + (rm -rf 0_0_reloadable24/Release) + (rm -rf 0_0_reloadable25/Release) + (rm -rf 0_1/Release) + (rm -rf 0_1_reloadable0/Release) + (rm -rf 0_1_reloadable1/Release) + (rm -rf 0_1_reloadable2/Release) + (rm -rf 0_1_reloadable3/Release) + (rm -rf 0_1_reloadable4/Release) + (rm -rf 0_1_reloadable5/Release) + (rm -rf 0_1_reloadable6/Release) + (rm -rf 0_1_reloadable7/Release) + (rm -rf 0_1_reloadable8/Release) + (rm -rf 0_1_reloadable9/Release) + (rm -rf 0_1_reloadable10/Release) + (rm -rf 0_1_reloadable11/Release) + (rm -rf 0_1_reloadable12/Release) + (rm -rf 0_1_reloadable13/Release) + (rm -rf 0_1_reloadable14/Release) + (rm -rf 0_1_reloadable15/Release) + (rm -rf 0_1_reloadable16/Release) + (rm -rf 0_1_reloadable17/Release) + (rm -rf 0_1_reloadable18/Release) + (rm -rf 0_1_reloadable19/Release) + (rm -rf 0_1_reloadable20/Release) + (rm -rf 0_1_reloadable21/Release) + (rm -rf 0_1_reloadable22/Release) + (rm -rf 0_1_reloadable23/Release) + (rm -rf 0_1_reloadable24/Release) + (rm -rf 0_1_reloadable25/Release) + (rm -rf 0_2/Release) + (rm -rf 0_2_reloadable0/Release) + (rm -rf 0_2_reloadable1/Release) + (rm -rf 0_2_reloadable2/Release) + (rm -rf 0_2_reloadable3/Release) + (rm -rf 0_2_reloadable4/Release) + (rm -rf 0_2_reloadable5/Release) + (rm -rf 0_2_reloadable6/Release) + (rm -rf 0_2_reloadable7/Release) + (rm -rf 0_2_reloadable8/Release) + (rm -rf 0_2_reloadable9/Release) + (rm -rf 0_2_reloadable10/Release) + (rm -rf 0_2_reloadable11/Release) + (rm -rf 0_2_reloadable12/Release) + (rm -rf 0_2_reloadable13/Release) + (rm -rf 0_2_reloadable14/Release) + (rm -rf 0_2_reloadable15/Release) + (rm -rf 0_2_reloadable16/Release) + (rm -rf 0_2_reloadable17/Release) + (rm -rf 0_2_reloadable18/Release) + (rm -rf 0_2_reloadable19/Release) + (rm -rf 0_2_reloadable20/Release) + (rm -rf 0_2_reloadable21/Release) + (rm -rf 0_2_reloadable22/Release) + (rm -rf 0_2_reloadable23/Release) + (rm -rf 0_2_reloadable24/Release) + (rm -rf 0_2_reloadable25/Release) + (rm -rf 0_3/Release) + (rm -rf 0_3_reloadable0/Release) + (rm -rf 0_3_reloadable1/Release) + (rm -rf 0_3_reloadable2/Release) + (rm -rf 0_3_reloadable3/Release) + (rm -rf 0_3_reloadable4/Release) + (rm -rf 0_3_reloadable5/Release) + (rm -rf 0_3_reloadable6/Release) + (rm -rf 0_3_reloadable7/Release) + (rm -rf 0_3_reloadable8/Release) + (rm -rf 0_3_reloadable9/Release) + (rm -rf 0_3_reloadable10/Release) + (rm -rf 0_3_reloadable11/Release) + (rm -rf 0_3_reloadable12/Release) + (rm -rf 0_3_reloadable13/Release) + (rm -rf 0_3_reloadable14/Release) + (rm -rf 0_3_reloadable15/Release) + (rm -rf 0_3_reloadable16/Release) + (rm -rf 0_3_reloadable17/Release) + (rm -rf 0_3_reloadable18/Release) + (rm -rf 0_3_reloadable19/Release) + (rm -rf 0_3_reloadable20/Release) + (rm -rf 0_3_reloadable21/Release) + (rm -rf 0_3_reloadable22/Release) + (rm -rf 0_3_reloadable23/Release) + (rm -rf 0_3_reloadable24/Release) + (rm -rf 0_3_reloadable25/Release) + (rm -rf 1_0/Release) + (rm -rf 1_0_reloadable0/Release) + (rm -rf 1_0_reloadable1/Release) + (rm -rf 1_0_reloadable2/Release) + (rm -rf 1_0_reloadable3/Release) + (rm -rf 1_0_reloadable4/Release) + (rm -rf 1_0_reloadable5/Release) + (rm -rf 1_0_reloadable6/Release) + (rm -rf 1_0_reloadable7/Release) + (rm -rf 1_0_reloadable8/Release) + (rm -rf 1_0_reloadable9/Release) + (rm -rf 1_0_reloadable10/Release) + (rm -rf 1_0_reloadable11/Release) + (rm -rf 1_0_reloadable12/Release) + (rm -rf 1_0_reloadable13/Release) + (rm -rf 1_0_reloadable14/Release) + (rm -rf 1_0_reloadable15/Release) + (rm -rf 1_0_reloadable16/Release) + (rm -rf 1_0_reloadable17/Release) + (rm -rf 1_0_reloadable18/Release) + (rm -rf 1_0_reloadable19/Release) + (rm -rf 1_0_reloadable20/Release) + (rm -rf 1_0_reloadable21/Release) + (rm -rf 1_0_reloadable22/Release) + (rm -rf 1_0_reloadable23/Release) + (rm -rf 1_0_reloadable24/Release) + (rm -rf 1_0_reloadable25/Release) + (rm -rf 1_1/Release) + (rm -rf 1_1_reloadable0/Release) + (rm -rf 1_1_reloadable1/Release) + (rm -rf 1_1_reloadable2/Release) + (rm -rf 1_1_reloadable3/Release) + (rm -rf 1_1_reloadable4/Release) + (rm -rf 1_1_reloadable5/Release) + (rm -rf 1_1_reloadable6/Release) + (rm -rf 1_1_reloadable7/Release) + (rm -rf 1_1_reloadable8/Release) + (rm -rf 1_1_reloadable9/Release) + (rm -rf 1_1_reloadable10/Release) + (rm -rf 1_1_reloadable11/Release) + (rm -rf 1_1_reloadable12/Release) + (rm -rf 1_1_reloadable13/Release) + (rm -rf 1_1_reloadable14/Release) + (rm -rf 1_1_reloadable15/Release) + (rm -rf 1_1_reloadable16/Release) + (rm -rf 1_1_reloadable17/Release) + (rm -rf 1_1_reloadable18/Release) + (rm -rf 1_1_reloadable19/Release) + (rm -rf 1_1_reloadable20/Release) + (rm -rf 1_1_reloadable21/Release) + (rm -rf 1_1_reloadable22/Release) + (rm -rf 1_1_reloadable23/Release) + (rm -rf 1_1_reloadable24/Release) + (rm -rf 1_1_reloadable25/Release) + (rm -rf 1_2/Release) + (rm -rf 1_2_reloadable0/Release) + (rm -rf 1_2_reloadable1/Release) + (rm -rf 1_2_reloadable2/Release) + (rm -rf 1_2_reloadable3/Release) + (rm -rf 1_2_reloadable4/Release) + (rm -rf 1_2_reloadable5/Release) + (rm -rf 1_2_reloadable6/Release) + (rm -rf 1_2_reloadable7/Release) + (rm -rf 1_2_reloadable8/Release) + (rm -rf 1_2_reloadable9/Release) + (rm -rf 1_2_reloadable10/Release) + (rm -rf 1_2_reloadable11/Release) + (rm -rf 1_2_reloadable12/Release) + (rm -rf 1_2_reloadable13/Release) + (rm -rf 1_2_reloadable14/Release) + (rm -rf 1_2_reloadable15/Release) + (rm -rf 1_2_reloadable16/Release) + (rm -rf 1_2_reloadable17/Release) + (rm -rf 1_2_reloadable18/Release) + (rm -rf 1_2_reloadable19/Release) + (rm -rf 1_2_reloadable20/Release) + (rm -rf 1_2_reloadable21/Release) + (rm -rf 1_2_reloadable22/Release) + (rm -rf 1_2_reloadable23/Release) + (rm -rf 1_2_reloadable24/Release) + (rm -rf 1_2_reloadable25/Release) + (rm -rf 1_3/Release) + (rm -rf 1_3_reloadable0/Release) + (rm -rf 1_3_reloadable1/Release) + (rm -rf 1_3_reloadable2/Release) + (rm -rf 1_3_reloadable3/Release) + (rm -rf 1_3_reloadable4/Release) + (rm -rf 1_3_reloadable5/Release) + (rm -rf 1_3_reloadable6/Release) + (rm -rf 1_3_reloadable7/Release) + (rm -rf 1_3_reloadable8/Release) + (rm -rf 1_3_reloadable9/Release) + (rm -rf 1_3_reloadable10/Release) + (rm -rf 1_3_reloadable11/Release) + (rm -rf 1_3_reloadable12/Release) + (rm -rf 1_3_reloadable13/Release) + (rm -rf 1_3_reloadable14/Release) + (rm -rf 1_3_reloadable15/Release) + (rm -rf 1_3_reloadable16/Release) + (rm -rf 1_3_reloadable17/Release) + (rm -rf 1_3_reloadable18/Release) + (rm -rf 1_3_reloadable19/Release) + (rm -rf 1_3_reloadable20/Release) + (rm -rf 1_3_reloadable21/Release) + (rm -rf 1_3_reloadable22/Release) + (rm -rf 1_3_reloadable23/Release) + (rm -rf 1_3_reloadable24/Release) + (rm -rf 1_3_reloadable25/Release) + (rm -rf 2_0/Release) + (rm -rf 2_0_reloadable0/Release) + (rm -rf 2_0_reloadable1/Release) + (rm -rf 2_0_reloadable2/Release) + (rm -rf 2_0_reloadable3/Release) + (rm -rf 2_0_reloadable4/Release) + (rm -rf 2_0_reloadable5/Release) + (rm -rf 2_0_reloadable6/Release) + (rm -rf 2_0_reloadable7/Release) + (rm -rf 2_0_reloadable8/Release) + (rm -rf 2_0_reloadable9/Release) + (rm -rf 2_0_reloadable10/Release) + (rm -rf 2_0_reloadable11/Release) + (rm -rf 2_0_reloadable12/Release) + (rm -rf 2_0_reloadable13/Release) + (rm -rf 2_0_reloadable14/Release) + (rm -rf 2_0_reloadable15/Release) + (rm -rf 2_0_reloadable16/Release) + (rm -rf 2_0_reloadable17/Release) + (rm -rf 2_0_reloadable18/Release) + (rm -rf 2_0_reloadable19/Release) + (rm -rf 2_0_reloadable20/Release) + (rm -rf 2_0_reloadable21/Release) + (rm -rf 2_0_reloadable22/Release) + (rm -rf 2_0_reloadable23/Release) + (rm -rf 2_0_reloadable24/Release) + (rm -rf 2_0_reloadable25/Release) + (rm -rf 2_1/Release) + (rm -rf 2_1_reloadable0/Release) + (rm -rf 2_1_reloadable1/Release) + (rm -rf 2_1_reloadable2/Release) + (rm -rf 2_1_reloadable3/Release) + (rm -rf 2_1_reloadable4/Release) + (rm -rf 2_1_reloadable5/Release) + (rm -rf 2_1_reloadable6/Release) + (rm -rf 2_1_reloadable7/Release) + (rm -rf 2_1_reloadable8/Release) + (rm -rf 2_1_reloadable9/Release) + (rm -rf 2_1_reloadable10/Release) + (rm -rf 2_1_reloadable11/Release) + (rm -rf 2_1_reloadable12/Release) + (rm -rf 2_1_reloadable13/Release) + (rm -rf 2_1_reloadable14/Release) + (rm -rf 2_1_reloadable15/Release) + (rm -rf 2_1_reloadable16/Release) + (rm -rf 2_1_reloadable17/Release) + (rm -rf 2_1_reloadable18/Release) + (rm -rf 2_1_reloadable19/Release) + (rm -rf 2_1_reloadable20/Release) + (rm -rf 2_1_reloadable21/Release) + (rm -rf 2_1_reloadable22/Release) + (rm -rf 2_1_reloadable23/Release) + (rm -rf 2_1_reloadable24/Release) + (rm -rf 2_1_reloadable25/Release) + (rm -rf 2_2/Release) + (rm -rf 2_2_reloadable0/Release) + (rm -rf 2_2_reloadable1/Release) + (rm -rf 2_2_reloadable2/Release) + (rm -rf 2_2_reloadable3/Release) + (rm -rf 2_2_reloadable4/Release) + (rm -rf 2_2_reloadable5/Release) + (rm -rf 2_2_reloadable6/Release) + (rm -rf 2_2_reloadable7/Release) + (rm -rf 2_2_reloadable8/Release) + (rm -rf 2_2_reloadable9/Release) + (rm -rf 2_2_reloadable10/Release) + (rm -rf 2_2_reloadable11/Release) + (rm -rf 2_2_reloadable12/Release) + (rm -rf 2_2_reloadable13/Release) + (rm -rf 2_2_reloadable14/Release) + (rm -rf 2_2_reloadable15/Release) + (rm -rf 2_2_reloadable16/Release) + (rm -rf 2_2_reloadable17/Release) + (rm -rf 2_2_reloadable18/Release) + (rm -rf 2_2_reloadable19/Release) + (rm -rf 2_2_reloadable20/Release) + (rm -rf 2_2_reloadable21/Release) + (rm -rf 2_2_reloadable22/Release) + (rm -rf 2_2_reloadable23/Release) + (rm -rf 2_2_reloadable24/Release) + (rm -rf 2_2_reloadable25/Release) + (rm -rf 2_3/Release) + (rm -rf 2_3_reloadable0/Release) + (rm -rf 2_3_reloadable1/Release) + (rm -rf 2_3_reloadable2/Release) + (rm -rf 2_3_reloadable3/Release) + (rm -rf 2_3_reloadable4/Release) + (rm -rf 2_3_reloadable5/Release) + (rm -rf 2_3_reloadable6/Release) + (rm -rf 2_3_reloadable7/Release) + (rm -rf 2_3_reloadable8/Release) + (rm -rf 2_3_reloadable9/Release) + (rm -rf 2_3_reloadable10/Release) + (rm -rf 2_3_reloadable11/Release) + (rm -rf 2_3_reloadable12/Release) + (rm -rf 2_3_reloadable13/Release) + (rm -rf 2_3_reloadable14/Release) + (rm -rf 2_3_reloadable15/Release) + (rm -rf 2_3_reloadable16/Release) + (rm -rf 2_3_reloadable17/Release) + (rm -rf 2_3_reloadable18/Release) + (rm -rf 2_3_reloadable19/Release) + (rm -rf 2_3_reloadable20/Release) + (rm -rf 2_3_reloadable21/Release) + (rm -rf 2_3_reloadable22/Release) + (rm -rf 2_3_reloadable23/Release) + (rm -rf 2_3_reloadable24/Release) + (rm -rf 2_3_reloadable25/Release) + (rm -rf 3_0/Release) + (rm -rf 3_0_reloadable0/Release) + (rm -rf 3_0_reloadable1/Release) + (rm -rf 3_0_reloadable2/Release) + (rm -rf 3_0_reloadable3/Release) + (rm -rf 3_0_reloadable4/Release) + (rm -rf 3_0_reloadable5/Release) + (rm -rf 3_0_reloadable6/Release) + (rm -rf 3_0_reloadable7/Release) + (rm -rf 3_0_reloadable8/Release) + (rm -rf 3_0_reloadable9/Release) + (rm -rf 3_0_reloadable10/Release) + (rm -rf 3_0_reloadable11/Release) + (rm -rf 3_0_reloadable12/Release) + (rm -rf 3_0_reloadable13/Release) + (rm -rf 3_0_reloadable14/Release) + (rm -rf 3_0_reloadable15/Release) + (rm -rf 3_0_reloadable16/Release) + (rm -rf 3_0_reloadable17/Release) + (rm -rf 3_0_reloadable18/Release) + (rm -rf 3_0_reloadable19/Release) + (rm -rf 3_0_reloadable20/Release) + (rm -rf 3_0_reloadable21/Release) + (rm -rf 3_0_reloadable22/Release) + (rm -rf 3_0_reloadable23/Release) + (rm -rf 3_0_reloadable24/Release) + (rm -rf 3_0_reloadable25/Release) + (rm -rf 3_1/Release) + (rm -rf 3_1_reloadable0/Release) + (rm -rf 3_1_reloadable1/Release) + (rm -rf 3_1_reloadable2/Release) + (rm -rf 3_1_reloadable3/Release) + (rm -rf 3_1_reloadable4/Release) + (rm -rf 3_1_reloadable5/Release) + (rm -rf 3_1_reloadable6/Release) + (rm -rf 3_1_reloadable7/Release) + (rm -rf 3_1_reloadable8/Release) + (rm -rf 3_1_reloadable9/Release) + (rm -rf 3_1_reloadable10/Release) + (rm -rf 3_1_reloadable11/Release) + (rm -rf 3_1_reloadable12/Release) + (rm -rf 3_1_reloadable13/Release) + (rm -rf 3_1_reloadable14/Release) + (rm -rf 3_1_reloadable15/Release) + (rm -rf 3_1_reloadable16/Release) + (rm -rf 3_1_reloadable17/Release) + (rm -rf 3_1_reloadable18/Release) + (rm -rf 3_1_reloadable19/Release) + (rm -rf 3_1_reloadable20/Release) + (rm -rf 3_1_reloadable21/Release) + (rm -rf 3_1_reloadable22/Release) + (rm -rf 3_1_reloadable23/Release) + (rm -rf 3_1_reloadable24/Release) + (rm -rf 3_1_reloadable25/Release) + (rm -rf 3_2/Release) + (rm -rf 3_2_reloadable0/Release) + (rm -rf 3_2_reloadable1/Release) + (rm -rf 3_2_reloadable2/Release) + (rm -rf 3_2_reloadable3/Release) + (rm -rf 3_2_reloadable4/Release) + (rm -rf 3_2_reloadable5/Release) + (rm -rf 3_2_reloadable6/Release) + (rm -rf 3_2_reloadable7/Release) + (rm -rf 3_2_reloadable8/Release) + (rm -rf 3_2_reloadable9/Release) + (rm -rf 3_2_reloadable10/Release) + (rm -rf 3_2_reloadable11/Release) + (rm -rf 3_2_reloadable12/Release) + (rm -rf 3_2_reloadable13/Release) + (rm -rf 3_2_reloadable14/Release) + (rm -rf 3_2_reloadable15/Release) + (rm -rf 3_2_reloadable16/Release) + (rm -rf 3_2_reloadable17/Release) + (rm -rf 3_2_reloadable18/Release) + (rm -rf 3_2_reloadable19/Release) + (rm -rf 3_2_reloadable20/Release) + (rm -rf 3_2_reloadable21/Release) + (rm -rf 3_2_reloadable22/Release) + (rm -rf 3_2_reloadable23/Release) + (rm -rf 3_2_reloadable24/Release) + (rm -rf 3_2_reloadable25/Release) + (rm -rf 3_3/Release) + (rm -rf 3_3_reloadable0/Release) + (rm -rf 3_3_reloadable1/Release) + (rm -rf 3_3_reloadable2/Release) + (rm -rf 3_3_reloadable3/Release) + (rm -rf 3_3_reloadable4/Release) + (rm -rf 3_3_reloadable5/Release) + (rm -rf 3_3_reloadable6/Release) + (rm -rf 3_3_reloadable7/Release) + (rm -rf 3_3_reloadable8/Release) + (rm -rf 3_3_reloadable9/Release) + (rm -rf 3_3_reloadable10/Release) + (rm -rf 3_3_reloadable11/Release) + (rm -rf 3_3_reloadable12/Release) + (rm -rf 3_3_reloadable13/Release) + (rm -rf 3_3_reloadable14/Release) + (rm -rf 3_3_reloadable15/Release) + (rm -rf 3_3_reloadable16/Release) + (rm -rf 3_3_reloadable17/Release) + (rm -rf 3_3_reloadable18/Release) + (rm -rf 3_3_reloadable19/Release) + (rm -rf 3_3_reloadable20/Release) + (rm -rf 3_3_reloadable21/Release) + (rm -rf 3_3_reloadable22/Release) + (rm -rf 3_3_reloadable23/Release) + (rm -rf 3_3_reloadable24/Release) + (rm -rf 3_3_reloadable25/Release) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/layer_control_parameters.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/layer_control_parameters.json new file mode 100644 index 0000000000000000000000000000000000000000..dce0cbd366be4ef3e23c3db0b9de7daaf4a978d8 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/layer_control_parameters.json @@ -0,0 +1,359522 @@ +{ + "0_0": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1073741824, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1073741824, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "0_1": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074266112, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074266112, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "0_2": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074790400, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074790400, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "0_3": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1075314688, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1075314688, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "1_0": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1073741824, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1073741824, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "1_1": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074266112, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074266112, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "1_2": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074790400, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074790400, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "1_3": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1075314688, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1075314688, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "2_0": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1073741824, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1073741824, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "2_1": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074266112, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074266112, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "2_2": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074790400, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074790400, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "2_3": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1075314688, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1075314688, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "3_0": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1073741824, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1073741824, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1073741824, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "3_1": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074266112, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074266112, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074266112, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "3_2": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074790400, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1074790400, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1074790400, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ], + "3_3": [ + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 998244353, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 16, + 1, + 0, + 1, + 0, + 1, + 466944, + 48, + 483328, + 49, + 458752, + 50, + 475136, + 51, + 5, + 1, + 4, + 8, + 1, + 1, + 12, + 126976, + 2, + 126992, + 0, + 118784, + 33554448, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 16711680, + 122372, + 9895936, + 2, + 12, + 127008, + 2, + 127024, + 0, + 118816, + 16, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 16711681, + 122388, + 9895937, + 920, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521920, + 52, + 502080, + 53, + 458752, + 50, + 475136, + 51, + 16, + 526914, + 50528264, + 16912640, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220368, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 177471792, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 1310720, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 1, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 54, + 499712, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194848, + 16842760, + 16908288, + 16256, + 327744, + 65548, + 0, + 0, + 512, + 0, + 3932160, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33555456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 3866624, + 127072, + 2, + 127088, + 0, + 118880, + 167772432, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 3866627, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 60, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 52, + 500736, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4196616, + 16842768, + 16842752, + 16777216, + 327692, + 65546, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3211264, + 127040, + 2, + 127056, + 0, + 118848, + 171967008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3211266, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3211265, + 50, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080282, + 66307, + 1344, + 84, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5439488, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5439490, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 96, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5439489, + 84, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4195344, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167772704, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303570, + 66307, + 1280, + 30, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1900544, + 127040, + 2, + 127056, + 0, + 118848, + 176160832, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1900546, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 0, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 520576, + 54, + 500736, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4719632, + 16842768, + 16842752, + 16256, + 327680, + 65539, + 0, + 0, + 1024, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 26, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 134220032, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 917504, + 127072, + 2, + 127088, + 0, + 118880, + 171967072, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 33882086, + 122380, + 917507, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842776, + 16908288, + 16777216, + 196672, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 167772976, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 18, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34080788, + 66821, + 1280, + 45, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2883584, + 127040, + 2, + 127056, + 0, + 118848, + 176160944, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2883586, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2883585, + 45, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 72, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513664, + 52, + 493824, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4718864, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218304, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 143655520, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 327680, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518272, + 52, + 498432, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4720136, + 16842768, + 16842752, + 16256, + 327692, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219456, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 127040, + 2, + 127056, + 0, + 118848, + 162529888, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 262146, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16777216, + 131136, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304076, + 66821, + 960, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 165675184, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 1, + 32, + 64, + 2, + 1, + 1, + 15, + 0, + 14990, + 0, + 15, + 920, + 1, + 120, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073680, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864592, + 16842768, + 16908288, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 149947360, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512384, + 52, + 492544, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194568, + 16842784, + 16842752, + 16256, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134217984, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 0, + 127040, + 2, + 127056, + 0, + 118848, + 138413120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 54, + 496640, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 4194600, + 16842768, + 33882112, + 16256, + 65548, + 65542, + 0, + 0, + 640, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219008, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483648, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555712, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 118912, + 155189792, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 720900, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 320, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 12, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194848, + 16842784, + 16908288, + 16256, + 131136, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 720896, + 127040, + 2, + 127056, + 0, + 118848, + 167773248, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 720898, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 720897, + 12, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 524288, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1048576, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 983040, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 983041, + 16, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519040, + 52, + 499200, + 53, + 458752, + 50, + 475136, + 51, + 16, + 34081290, + 771, + 960, + 40, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219648, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2555904, + 127040, + 2, + 127056, + 0, + 118848, + 165675072, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2555906, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 2555905, + 40, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842776, + 50462720, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190256, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 9, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842784, + 16908288, + 16256, + 65600, + 131075, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 155190592, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 7, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 393217, + 7, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 131072, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518016, + 54, + 498176, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6816032, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219392, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215484032, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 161482000, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515200, + 52, + 495360, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243656, + 16842800, + 16842752, + 16256, + 196620, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 131072, + 127040, + 2, + 127056, + 0, + 118848, + 149948384, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 131074, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 576, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 6, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 327680, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 327682, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 6, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 131073, + 3, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842776, + 33685504, + 16256, + 65600, + 65539, + 0, + 0, + 768, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 32, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 118912, + 159384752, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 537439, + 118932, + 33882086, + 122380, + 327684, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 6, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 16908288, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 262144, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 3, + 5, + 0, + 15241, + 0, + 15, + 240, + 1, + 480, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 131073, + 15, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 67174400, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 4, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 513280, + 52, + 493440, + 53, + 458752, + 50, + 475136, + 51, + 16, + 7864584, + 16842784, + 16842752, + 16256, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218208, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 196608, + 127040, + 2, + 127056, + 0, + 118848, + 142084032, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 196610, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291744, + 16842784, + 84017152, + 16256, + 65536, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218688, + 118788, + 0, + 118792, + 1040384, + 118796, + 25821184, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 159385152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 522112, + 52, + 502272, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17303066, + 771, + 1344, + 21, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220416, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1310720, + 127040, + 2, + 127056, + 0, + 118848, + 178257984, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1310722, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1310721, + 21, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 117571584, + 16256, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 0, + 62, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219264, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219264, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219264, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119072, + 159385152, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 33882086, + 122380, + 1310729, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 21, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 515456, + 52, + 495616, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194592, + 16842808, + 33685504, + 16256, + 65600, + 196611, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218368, + 118788, + 0, + 118792, + 1040384, + 118796, + 17432576, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1114112, + 127040, + 2, + 127056, + 0, + 118848, + 150996848, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1114114, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 896, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 524289, + 18, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 126, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 8192000, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 8192002, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 8192001, + 126, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 11, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503552, + 50, + 511360, + 51, + 16, + 1, + 56, + 32, + 2, + 1, + 3, + 8, + 0, + 15241, + 0, + 24, + 240, + 1, + 672, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3584, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1507328, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 183500828, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10143, + 118836, + 33841123, + 122388, + 131073, + 24, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512896, + 52, + 493056, + 53, + 458752, + 50, + 475136, + 51, + 16, + 6291720, + 16842800, + 117506048, + 16777216, + 65548, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218112, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 393216, + 127040, + 2, + 127056, + 0, + 118848, + 140511584, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 393218, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 7, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516736, + 52, + 496896, + 53, + 458752, + 50, + 475136, + 51, + 16, + 11010320, + 16842768, + 16908288, + 16256, + 65548, + 720897, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219072, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 655360, + 127040, + 2, + 127056, + 0, + 118848, + 156239200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 655362, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 720896, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 655360, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 655361, + 11, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 151126016, + 16256, + 65600, + 65539, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1703936, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1703938, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 27, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 521600, + 52, + 501760, + 53, + 458752, + 50, + 475136, + 51, + 16, + 17304080, + 2313, + 1280, + 180, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 11730944, + 127040, + 2, + 127056, + 0, + 118848, + 176161216, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 11730946, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 64, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16777216, + 65548, + 262145, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 512640, + 52, + 492800, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243144, + 16842800, + 50397184, + 16256, + 65548, + 327681, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 139462624, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503680, + 50, + 511360, + 51, + 16, + 384, + 0, + 65536, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 768, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184025280, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10111, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503936, + 50, + 511360, + 51, + 16, + 256, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 512, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 185073792, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10047, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 65536, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 983040, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 4, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518528, + 54, + 498688, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 7340320, + 16842776, + 151126016, + 16256, + 65600, + 131075, + 0, + 0, + 768, + 0, + 393216, + 0, + 0, + 0, + 0, + 0, + 0, + 74, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219520, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 503594976, + 118880, + 134219520, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 537439, + 118900, + 637812704, + 118912, + 134219520, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 13151, + 118932, + 772030432, + 118944, + 134219520, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 537439, + 118964, + 906248160, + 118976, + 134219520, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 13151, + 118996, + 1040465888, + 119008, + 134219520, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 537439, + 119028, + 1174683616, + 119040, + 134219520, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 13151, + 119060, + 1308901344, + 119072, + 134219520, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 537439, + 119092, + 1443119072, + 119104, + 134219520, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 13151, + 119124, + 369377248, + 118848, + 33555968, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 327680, + 127072, + 2, + 127088, + 0, + 119136, + 163579248, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 537439, + 119156, + 33882086, + 122380, + 3473419, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 384, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 327681, + 54, + 0, + 5, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516480, + 52, + 496640, + 53, + 458752, + 50, + 475136, + 51, + 16, + 5243168, + 16842792, + 33685504, + 16256, + 65600, + 393219, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218528, + 118788, + 0, + 118792, + 1040384, + 118796, + 21626880, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2293760, + 127040, + 2, + 127056, + 0, + 118848, + 155190928, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2293762, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1114113, + 36, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1077936129, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 9, + 0, + 0, + 0, + 0, + 0, + 0, + 16576, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 589824, + 1042939905, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1310720, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503808, + 50, + 511360, + 51, + 16, + 1, + 40, + 48, + 2, + 1, + 6, + 5, + 0, + 15241, + 0, + 30, + 240, + 1, + 960, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 184549396, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10079, + 118836, + 33841123, + 122388, + 327681, + 30, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 52, + 497664, + 53, + 458752, + 50, + 475136, + 51, + 16, + 12583184, + 16842768, + 84017152, + 16256, + 65548, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 159385120, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 128, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 503424, + 50, + 511360, + 51, + 16, + 512, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1024, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 182976768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10175, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 2, + 1, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517504, + 54, + 497664, + 55, + 466944, + 52, + 483328, + 53, + 458752, + 50, + 475136, + 51, + 17, + 6291744, + 16842784, + 167903232, + 16777216, + 65536, + 65539, + 0, + 0, + 1024, + 0, + 196608, + 0, + 0, + 0, + 0, + 0, + 1, + 80, + 126976, + 2, + 126992, + 0, + 127040, + 2, + 127056, + 0, + 118784, + 134219264, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 4959, + 118804, + 503594976, + 118880, + 215483904, + 118884, + 0, + 118888, + 0, + 118892, + 0, + 118896, + 4959, + 118900, + 637812704, + 118912, + 134219264, + 118916, + 0, + 118920, + 0, + 118924, + 0, + 118928, + 4959, + 118932, + 772030432, + 118944, + 215483904, + 118948, + 0, + 118952, + 0, + 118956, + 0, + 118960, + 4959, + 118964, + 906248160, + 118976, + 134219264, + 118980, + 0, + 118984, + 0, + 118988, + 0, + 118992, + 4959, + 118996, + 1040465888, + 119008, + 215483904, + 119012, + 0, + 119016, + 0, + 119020, + 0, + 119024, + 4959, + 119028, + 1174683616, + 119040, + 134219264, + 119044, + 0, + 119048, + 0, + 119052, + 0, + 119056, + 4959, + 119060, + 1308901344, + 119072, + 215483904, + 119076, + 0, + 119080, + 0, + 119084, + 0, + 119088, + 4959, + 119092, + 1443119072, + 119104, + 134219264, + 119108, + 0, + 119112, + 0, + 119116, + 0, + 119120, + 4959, + 119124, + 1577336800, + 119136, + 215483904, + 119140, + 0, + 119144, + 0, + 119148, + 0, + 119152, + 4959, + 119156, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122372, + 131072, + 127072, + 2, + 127088, + 0, + 119168, + 159385152, + 119172, + 0, + 119176, + 0, + 119180, + 0, + 119184, + 537439, + 119188, + 33882086, + 122380, + 1900556, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 30, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 131073, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 983040, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 983042, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 16, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 516800, + 52, + 496960, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049890, + 50528272, + 134348800, + 16256, + 65536, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134218608, + 118788, + 0, + 118792, + 1105920, + 118796, + 21692416, + 118800, + 13151, + 118804, + 33832928, + 122372, + 458752, + 127040, + 2, + 127056, + 0, + 118848, + 156501152, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 458754, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 768, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1, + 8, + 0, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 1, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 65536, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 0, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 1, + 1, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 12, + 20, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 17563928, + 19398913, + 16843028, + 17563936, + 35651841, + 16909073, + 0, + 0, + 2, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67112128, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 65537, + 2, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 524880, + 258, + 0, + 0, + 83886080, + 96, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220288, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6225920, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 160, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 6225921, + 96, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 20, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 458752, + 50, + 475136, + 51, + 16, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 527376, + 258, + 0, + 0, + 100663296, + 5, + 1048576000, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 134220800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 262144, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 192, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 5, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 184745984, + 16777216, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 44, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 131074, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 1245184, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 1245186, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 196609, + 20, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 3, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 131072, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 131073, + 3, + 1, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519360, + 52, + 499520, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1049906, + 50528272, + 84082688, + 16256, + 65536, + 65538, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219408, + 118788, + 0, + 118792, + 1630208, + 118796, + 22347776, + 118800, + 13151, + 118804, + 33832928, + 122372, + 589824, + 127040, + 2, + 127056, + 0, + 118848, + 166986912, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 589826, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1152, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 65537, + 10, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 2, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 65536, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 65537, + 2, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 262144, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 475136, + 48, + 475136, + 49, + 458752, + 50, + 458752, + 51, + 16, + 23, + 40, + 1056964608, + 1056964608, + 3196059648, + 3196059648, + 4, + 18284846, + 36176129, + 16913173, + 101777952, + 35651842, + 16912146, + 0, + 0, + 8, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 67113760, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 4096, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 117506048, + 16777216, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 2228224, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 2228226, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 35, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519424, + 52, + 499584, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1052178, + 50528272, + 50397184, + 16256, + 327680, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219744, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 167249056, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 262145, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 458753, + 8, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 20, + 40, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 458752, + 49, + 483328, + 50, + 483328, + 51, + 7, + 40, + 3, + 80, + 0, + 20, + 0, + 9600, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 4800, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 0, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 100666176, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 0, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 1, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 527906, + 50528264, + 84017152, + 16256, + 196672, + 65537, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 917504, + 127040, + 2, + 127056, + 0, + 118848, + 165413168, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 917506, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1536, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 131073, + 15, + 0, + 5, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500608, + 50, + 511360, + 51, + 16, + 1920, + 0, + 4, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3840, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 196608, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 171443136, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10879, + 118836, + 33841123, + 122388, + 196609, + 4, + 0, + 6, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 482816, + 52, + 466432, + 53, + 502528, + 50, + 511360, + 51, + 16, + 960, + 0, + 524288, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 31459200, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 458752, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 179306976, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10399, + 118836, + 33841123, + 122388, + 458753, + 8, + 0, + 7, + 0 + ], + [ + 1, + 2, + 0, + 1, + 0, + 1, + 458752, + 48, + 475136, + 49, + 466944, + 52, + 483328, + 53, + 491520, + 50, + 516096, + 51, + 8, + 24, + 24, + 40, + 25, + 20, + 20, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 1200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 524288, + 127040, + 2, + 127056, + 0, + 118848, + 33555632, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 12287, + 118868, + 33865700, + 122380, + 524290, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 134218228, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 14335, + 118836, + 33841123, + 122388, + 524289, + 9, + 0, + 4, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 48, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 3080192, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1075314688, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 3080193, + 48, + 1, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 67239936, + 16777216, + 327744, + 65541, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 6488064, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 6488066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1572865, + 100, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 500352, + 50, + 511360, + 51, + 16, + 2048, + 0, + 15, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 4096, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 917504, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 170394624, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10943, + 118836, + 33841123, + 122388, + 917505, + 15, + 0, + 2, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 3, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517888, + 52, + 498048, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050146, + 50528264, + 33685504, + 16256, + 327744, + 65542, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219360, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 3866624, + 127040, + 2, + 127056, + 0, + 118848, + 160957008, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 3866626, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 512, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 1900545, + 60, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501504, + 50, + 511360, + 51, + 16, + 1472, + 0, + 20, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 2944, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 1245184, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 175112928, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10655, + 118836, + 33841123, + 122388, + 1245185, + 20, + 0, + 5, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 4, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 483328, + 52, + 466944, + 53, + 502400, + 50, + 511360, + 51, + 16, + 1024, + 0, + 1966080, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 2048, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 33556480, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 1900544, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 178782720, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10431, + 118836, + 33841123, + 122388, + 1900545, + 30, + 0, + 6, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 479232, + 48, + 479232, + 49, + 487424, + 50, + 487424, + 51, + 16, + 32, + 8, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 69, + 9, + 126976, + 1, + 126992, + 0, + 118784, + 84608000, + 118788, + 0, + 118792, + 319488, + 118796, + 67371008, + 118800, + 0, + 118804, + 33832928, + 122372, + 4456448, + 2, + 9, + 127008, + 1, + 127024, + 0, + 118816, + 118686720, + 118820, + 1075314688, + 118824, + 581632, + 118828, + 34078720, + 118832, + 0, + 118836, + 33841123, + 122388, + 4456449, + 69, + 0, + 0, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 517696, + 52, + 497856, + 53, + 458752, + 50, + 475136, + 51, + 16, + 525890, + 50528264, + 84148224, + 16777216, + 327744, + 65548, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 20, + 126976, + 2, + 126992, + 0, + 118784, + 134219312, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 16711680, + 122372, + 2818048, + 127040, + 2, + 127056, + 0, + 118848, + 160170288, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 16711682, + 122380, + 2818050, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 1024, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 3866625, + 300, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 518976, + 52, + 499136, + 53, + 458752, + 50, + 475136, + 51, + 16, + 1050402, + 50528264, + 16908288, + 16777216, + 655424, + 65545, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219632, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 5832704, + 127040, + 2, + 127056, + 0, + 118848, + 165413456, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 5832706, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 640, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 5832705, + 90, + 0, + 1, + 0 + ], + [ + 1, + 2, + 0, + 0, + 1, + 1, + 491520, + 48, + 511360, + 49, + 519552, + 52, + 499712, + 53, + 458752, + 50, + 475136, + 51, + 16, + 4194624, + 16842760, + 17039360, + 16256, + 327744, + 65581, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 18, + 126976, + 2, + 126992, + 0, + 118784, + 134219776, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 13151, + 118804, + 33832928, + 122372, + 14680064, + 127040, + 2, + 127056, + 0, + 118848, + 167772432, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 537439, + 118868, + 33865700, + 122380, + 14680066, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 12287, + 118836, + 33841123, + 122388, + 14680065, + 225, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 3, + 4, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 1, + 0, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 0 + ], + [ + 1, + 1, + 0, + 1, + 0, + 1, + 458752, + 48, + 466944, + 49, + 475136, + 50, + 483328, + 51, + 7, + 8, + 3, + 40, + 0, + 3, + 1, + 3840, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 1920, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 10239, + 118804, + 33832928, + 122372, + 2031616, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 67109344, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 2031617, + 32, + 0, + 0, + 0 + ], + [ + 1, + 1, + 1, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 480256, + 52, + 463872, + 53, + 503168, + 50, + 511360, + 51, + 16, + 640, + 0, + 11796480, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 0, + 17, + 126976, + 2, + 127040, + 2, + 126992, + 0, + 127056, + 0, + 118784, + 1280, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 369377248, + 118848, + 20972800, + 118852, + 0, + 118856, + 0, + 118860, + 0, + 118864, + 536575, + 118868, + 33865700, + 122372, + 11730944, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 181928256, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10239, + 118836, + 33841123, + 122388, + 11730945, + 180, + 0, + 2, + 0 + ], + [ + 1, + 1, + 0, + 0, + 1, + 1, + 458752, + 48, + 475136, + 49, + 501248, + 50, + 511360, + 51, + 16, + 1600, + 0, + 72, + 0, + 0, + 0, + 0, + 0, + 0, + 16256, + 0, + 0, + 0, + 0, + 0, + 0, + 9, + 126976, + 2, + 126992, + 0, + 118784, + 3200, + 118788, + 0, + 118792, + 0, + 118796, + 0, + 118800, + 12287, + 118804, + 33832928, + 122372, + 4653056, + 2, + 9, + 127008, + 2, + 127024, + 0, + 118816, + 174064416, + 118820, + 1075314688, + 118824, + 0, + 118828, + 0, + 118832, + 10719, + 118836, + 33841123, + 122388, + 4653057, + 72, + 0, + 1, + 1 + ] + ] +} \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/top.aiecompile_summary b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/top.aiecompile_summary new file mode 100644 index 0000000000000000000000000000000000000000..0789535129966802abbf47ce089b3a0c84462b69 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/top.aiecompile_summary @@ -0,0 +1,18513 @@ + +{ + "thisFile": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/top.aiecompile_summary", + "connectId": "", + "serverToken": "", + "timestamp": "0" +} + + +{ + "type": "ET_CmdStep", + "dateTimestamp": "Fri Mar 21 03:30:57 2025", + "timestampMillis": "1742527857400", + "buildStep": { + "cmdId": "dcd70920-0818-4355-83df-cd68be811a1d", + "name": "aiecompiler", + "logFile": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/AIECompiler.log", + "commandLine": "/usr/local/lib/python3.10/dist-packages/bin/unwrapped/lnx64.o/aiecompiler /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/top.cpp --part=xc10AIE2P_ML-die-0x-e-S-es1 --nodot-graph --runtime-opt=1 --disable-multirate-analysis --enable-core-processor-bus --enable-multi-layer --heapsize=1792 --stacksize=1400 --max-layer-ctrl-param-size=256 --compile-for-aiesim=false --workdir=Work --multi-layer-ctrl-pkt --aie2ipu-base-addr=0 -enable-light-cdo --Xelfgen=-j4 --multi-layer-pipelining --multi-layer-opt=3 --Xpreproc=-D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ --multi-layer-ext-buf-file=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/flexmlrt-hsi.json --enable-partition=0:4 --multi-layer-ctrl-pkt-column-span=4 --multi-layer-prebuilt-archive=/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras/data/ryzen-ai/stx/unified-overlay-4x4.json --multi-layer-prebuilt-archive-enable-elf-gen --multi-layer-init-core-elf-ctrl-pkt --multi-layer-pm-id 29006 --Xpreproc=-DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 --include=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend --include=/usr/local/lib/python3.10/site-packages/include/aie_api --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf --output-archive libadf.a --adf-api-log-level=0 --multi-layer-pm-reloading=1 --Xelfgen=-j1 ", + "args": [ + "/usr/local/lib/python3.10/dist-packages/bin/unwrapped/lnx64.o/aiecompiler", + "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/top.cpp", + "--part=xc10AIE2P_ML-die-0x-e-S-es1", + "--nodot-graph", + "--runtime-opt=1", + "--disable-multirate-analysis", + "--enable-core-processor-bus", + "--enable-multi-layer", + "--heapsize=1792", + "--stacksize=1400", + "--max-layer-ctrl-param-size=256", + "--compile-for-aiesim=false", + "--workdir=Work", + "--multi-layer-ctrl-pkt", + "--aie2ipu-base-addr=0", + "-enable-light-cdo", + "--Xelfgen=-j4", + "--multi-layer-pipelining", + "--multi-layer-opt=3", + "--Xpreproc=-D__IO_BUFFER_FORCE_LIGHT_WEIGHT__", + "--multi-layer-ext-buf-file=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/flexmlrt-hsi.json", + "--enable-partition=0:4", + "--multi-layer-ctrl-pkt-column-span=4", + "--multi-layer-prebuilt-archive=/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras/data/ryzen-ai/stx/unified-overlay-4x4.json", + "--multi-layer-prebuilt-archive-enable-elf-gen", + "--multi-layer-init-core-elf-ctrl-pkt", + "--multi-layer-pm-id", + "29006", + "--Xpreproc=-DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1", + "--include=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend", + "--include=/usr/local/lib/python3.10/site-packages/include/aie_api", + "--include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common", + "--include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common", + "--include=/usr/local/lib/python3.10/dist-packages/vitis_mllib", + "--include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc", + "--include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf", + "--output-archive", + "libadf.a", + "--adf-api-log-level=0", + "--multi-layer-pm-reloading=1", + "--Xelfgen=-j1" + ], + "iniFiles": [], + "cwd": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler" + } +} + + +{ + "type": "ET_Status", + "dateTimestamp": "Fri Mar 21 03:30:57 2025", + "timestampMillis": "1742527857401", + "status": { + "cmdId": "dcd70920-0818-4355-83df-cd68be811a1d", + "state": "CS_RUNNING" + } +} + + +{ + "type": "ET_FlowMetaData", + "dateTimestamp": "Fri Mar 21 03:30:57 2025", + "timestampMillis": "1742527857401", + "buildSummary": { + "hardwarePlatform": "", + "hardwareDsa": "", + "platformDirectory": "", + "runtime": "", + "systemConfig": "", + "flow": "BF_UNKNOWN", + "target": "TT_UNKNOWN", + "kernels": [], + "toolVersion": "Vitis AI Engine Compiler Release 2024.2.0. SW Build 6074767 on 2025-03-20-05:22:32" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:30:57 2025", + "timestampMillis": "1742527857402", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work", + "name": "", + "fileType": "DIR", + "reportType": "AIE_COMPILER_WORKDIR", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:30:57 2025", + "timestampMillis": "1742527857402", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/reports/guidance.html", + "name": "", + "fileType": "HTML", + "reportType": "GLOBAL_RULECHECK_GUIDANCE", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:30:57 2025", + "timestampMillis": "1742527857402", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/reports/guidance.pb3", + "name": "", + "fileType": "BINARY_PROTOBUF", + "reportType": "GLOBAL_RULECHECK_GUIDANCE", + "cmdId": "" + } +} + + +{ + "type": "ET_CmdStep", + "dateTimestamp": "Fri Mar 21 03:31:54 2025", + "timestampMillis": "1742527914334", + "buildStep": { + "cmdId": "31edb987-7a66-4abb-b94a-830cd7d309ef", + "name": "aieir_be", + "logFile": "", + "commandLine": "/usr/local/lib/python3.10/dist-packages/bin/unwrapped/lnx64.o/aieir_be --time-passes=0 --Xelfgen=-j4 --Xelfgen=-j1 --compile-for-aiesim=false --multi-layer-pipelining=true --multi-layer-ctrl-pkt-column-span=4 --multi-layer-ctrl-pkt=true --heapsize=1792 --enable-light-cdo=true --Xpreproc=-D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ --Xpreproc=-DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 --target=hw --enable-core-processor-bus=true --include=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend --include=/usr/local/lib/python3.10/site-packages/include/aie_api --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf --log-level=1 --compute-heapsize=false --stacksize=1400 --multi-layer-prebuilt-archive-enable-elf-gen=true --multi-layer-ext-buf-file=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/flexmlrt-hsi.json --multi-layer-prebuilt-archive=/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras/data/ryzen-ai/stx/unified-overlay-4x4.json --multi-layer-pm-reloading=1 --multi-layer-init-core-elf-ctrl-pkt=true --enable-partition=0:4 --multi-layer-opt=3 --float-accuracy=fast --multi-layer-pm-id=29006 --disable-multirate-analysis=true --nodot-graph=true --aie2ipu-base-addr=0 --enable-multi-layer=true --adf-api-log-level=0 --event-trace-port=gmio --part=xc10AIE2P_ML-die-0x-e-S-es1 --json=Work/temp/top.json --aiearch aie2p --frontend-warning=0 --frontend-critical-warning=0 --frontend-summary-name=Work/top.aiecompile_summary --frontend-files-to-report=Guidance:Work/reports/guidance.html --sdf-graph=/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/top.cpp ", + "args": [ + "/usr/local/lib/python3.10/dist-packages/bin/unwrapped/lnx64.o/aieir_be", + "--time-passes=0", + "--Xelfgen=-j4", + "--Xelfgen=-j1", + "--compile-for-aiesim=false", + "--multi-layer-pipelining=true", + "--multi-layer-ctrl-pkt-column-span=4", + "--multi-layer-ctrl-pkt=true", + "--heapsize=1792", + "--enable-light-cdo=true", + "--Xpreproc=-D__IO_BUFFER_FORCE_LIGHT_WEIGHT__", + "--Xpreproc=-DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1", + "--target=hw", + "--enable-core-processor-bus=true", + "--include=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend", + "--include=/usr/local/lib/python3.10/site-packages/include/aie_api", + "--include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common", + "--include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common", + "--include=/usr/local/lib/python3.10/dist-packages/vitis_mllib", + "--include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc", + "--include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf", + "--log-level=1", + "--compute-heapsize=false", + "--stacksize=1400", + "--multi-layer-prebuilt-archive-enable-elf-gen=true", + "--multi-layer-ext-buf-file=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/flexmlrt-hsi.json", + "--multi-layer-prebuilt-archive=/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras/data/ryzen-ai/stx/unified-overlay-4x4.json", + "--multi-layer-pm-reloading=1", + "--multi-layer-init-core-elf-ctrl-pkt=true", + "--enable-partition=0:4", + "--multi-layer-opt=3", + "--float-accuracy=fast", + "--multi-layer-pm-id=29006", + "--disable-multirate-analysis=true", + "--nodot-graph=true", + "--aie2ipu-base-addr=0", + "--enable-multi-layer=true", + "--adf-api-log-level=0", + "--event-trace-port=gmio", + "--part=xc10AIE2P_ML-die-0x-e-S-es1", + "--json=Work/temp/top.json", + "--aiearch", + "aie2p", + "--frontend-warning=0", + "--frontend-critical-warning=0", + "--frontend-summary-name=Work/top.aiecompile_summary", + "--frontend-files-to-report=Guidance:Work/reports/guidance.html", + "--sdf-graph=/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/top.cpp" + ], + "iniFiles": [], + "cwd": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler" + } +} + + +{ + "type": "ET_Status", + "dateTimestamp": "Fri Mar 21 03:31:54 2025", + "timestampMillis": "1742527914335", + "status": { + "cmdId": "31edb987-7a66-4abb-b94a-830cd7d309ef", + "state": "CS_RUNNING" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:31:54 2025", + "timestampMillis": "1742527914336", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work", + "name": "", + "fileType": "DIR", + "reportType": "AIE_COMPILER_WORKDIR", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:31:54 2025", + "timestampMillis": "1742527914416", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/temp/debug_config.json", + "name": "", + "fileType": "JSON", + "reportType": "AIE_DEBUG_CONFIG", + "cmdId": "" + } +} + + +{ + "type": "ET_FlowMetaData", + "dateTimestamp": "Fri Mar 21 03:31:54 2025", + "timestampMillis": "1742527914416", + "buildSummary": { + "hardwarePlatform": "xc10AIE2P_ML-die-0x-e-S-es1", + "hardwareDsa": "aie2p_8x4_device", + "platformDirectory": "", + "runtime": "", + "systemConfig": "", + "flow": "BF_COMPILE", + "target": "TT_HW", + "kernels": [ + { + "base": { + "type": "KERNEL", + "name": "", + "file": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/temp/top.json", + "reports": [], + "uuid": "" + }, + "sources": [ + "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/top.cpp" + ], + "psSources": [], + "cuNames": [], + "type": "AIE", + "frequency": 1000, + "freqUnits": "" + } + ], + "toolVersion": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:31:54 2025", + "timestampMillis": "1742527914419", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/temp/top.json", + "name": "", + "fileType": "JSON", + "reportType": "AIE_GRAPH", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:33:40 2025", + "timestampMillis": "1742528020980", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/temp/top_partition.json", + "name": "", + "fileType": "JSON", + "reportType": "AIE_GRAPH_PARTITION", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:33:40 2025", + "timestampMillis": "1742528020984", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/scripts/pm_reload_analysis0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:33:40 2025", + "timestampMillis": "1742528020985", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:33:40 2025", + "timestampMillis": "1742528020987", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/pm_reload_analysis0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276592", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/scripts/0_0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276592", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/src/0_0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276593", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1/scripts/0_1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276594", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1/src/0_1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276595", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2/scripts/0_2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276595", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2/src/0_2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276596", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3/scripts/0_3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276596", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3/src/0_3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276597", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0/scripts/1_0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276597", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0/src/1_0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276598", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1/scripts/1_1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276599", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1/src/1_1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276599", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2/scripts/1_2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276600", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2/src/1_2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276601", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3/scripts/1_3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276601", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3/src/1_3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276602", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0/scripts/2_0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276602", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0/src/2_0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276603", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1/scripts/2_1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276603", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1/src/2_1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276604", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2/scripts/2_2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276605", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2/src/2_2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276605", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3/scripts/2_3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276606", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3/src/2_3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276607", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0/scripts/3_0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276607", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0/src/3_0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276608", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1/scripts/3_1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276608", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1/src/3_1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276610", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2/scripts/3_2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276610", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2/src/3_2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276611", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3/scripts/3_3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276611", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3/src/3_3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276612", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable0/scripts/0_0_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276612", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable0/src/0_0_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276613", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable1/scripts/0_0_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276614", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable1/src/0_0_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276614", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable2/scripts/0_0_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276615", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable2/src/0_0_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276615", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable3/scripts/0_0_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276616", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable3/src/0_0_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276617", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable4/scripts/0_0_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276617", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable4/src/0_0_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276618", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable5/scripts/0_0_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276618", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable5/src/0_0_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276619", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable6/scripts/0_0_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276619", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable6/src/0_0_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276620", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable7/scripts/0_0_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276620", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable7/src/0_0_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276621", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable8/scripts/0_0_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276621", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable8/src/0_0_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276622", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable9/scripts/0_0_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276623", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable9/src/0_0_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276623", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable10/scripts/0_0_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276624", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable10/src/0_0_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276625", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable11/scripts/0_0_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276625", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable11/src/0_0_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276626", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable12/scripts/0_0_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276626", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable12/src/0_0_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276627", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable13/scripts/0_0_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276627", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable13/src/0_0_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276628", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable14/scripts/0_0_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276628", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable14/src/0_0_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276629", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable15/scripts/0_0_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276629", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable15/src/0_0_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276630", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable16/scripts/0_0_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276631", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable16/src/0_0_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276631", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable17/scripts/0_0_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276632", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable17/src/0_0_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276633", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable18/scripts/0_0_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276633", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable18/src/0_0_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276634", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable19/scripts/0_0_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276634", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable19/src/0_0_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276635", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable20/scripts/0_0_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276635", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable20/src/0_0_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276636", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable21/scripts/0_0_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276636", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable21/src/0_0_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276637", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/scripts/0_0_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276637", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276638", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/scripts/0_0_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276638", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276639", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable24/scripts/0_0_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276639", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable24/src/0_0_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276640", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable25/scripts/0_0_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276641", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable25/src/0_0_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276642", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable0/scripts/0_1_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276642", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable0/src/0_1_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276643", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable1/scripts/0_1_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276643", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable1/src/0_1_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276644", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable2/scripts/0_1_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276644", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable2/src/0_1_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276645", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable3/scripts/0_1_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276645", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable3/src/0_1_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276646", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable4/scripts/0_1_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276646", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable4/src/0_1_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276647", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable5/scripts/0_1_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276647", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable5/src/0_1_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276648", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable6/scripts/0_1_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276648", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable6/src/0_1_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276649", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable7/scripts/0_1_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276650", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable7/src/0_1_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276651", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable8/scripts/0_1_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276651", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable8/src/0_1_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276652", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable9/scripts/0_1_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276652", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable9/src/0_1_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276653", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable10/scripts/0_1_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276653", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable10/src/0_1_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276654", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable11/scripts/0_1_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276654", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable11/src/0_1_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276655", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable12/scripts/0_1_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276655", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable12/src/0_1_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276656", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable13/scripts/0_1_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276656", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable13/src/0_1_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276657", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable14/scripts/0_1_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276657", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable14/src/0_1_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276658", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable15/scripts/0_1_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276659", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable15/src/0_1_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276659", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable16/scripts/0_1_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276660", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable16/src/0_1_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276660", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable17/scripts/0_1_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276661", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable17/src/0_1_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276662", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable18/scripts/0_1_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276662", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable18/src/0_1_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276663", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable19/scripts/0_1_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276663", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable19/src/0_1_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276664", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable20/scripts/0_1_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276664", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable20/src/0_1_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276665", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable21/scripts/0_1_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276665", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable21/src/0_1_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276666", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable22/scripts/0_1_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276666", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable22/src/0_1_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276667", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable23/scripts/0_1_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276667", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable23/src/0_1_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276668", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable24/scripts/0_1_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276669", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable24/src/0_1_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276669", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable25/scripts/0_1_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276670", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable25/src/0_1_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276671", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable0/scripts/0_2_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276671", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable0/src/0_2_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276672", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable1/scripts/0_2_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276672", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable1/src/0_2_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276673", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable2/scripts/0_2_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276673", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable2/src/0_2_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276674", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable3/scripts/0_2_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276674", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable3/src/0_2_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276675", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable4/scripts/0_2_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276675", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable4/src/0_2_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276676", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable5/scripts/0_2_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276676", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable5/src/0_2_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276677", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable6/scripts/0_2_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276678", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable6/src/0_2_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276678", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable7/scripts/0_2_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276679", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable7/src/0_2_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276680", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable8/scripts/0_2_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276680", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable8/src/0_2_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276681", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable9/scripts/0_2_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276681", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable9/src/0_2_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276682", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable10/scripts/0_2_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276682", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable10/src/0_2_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276683", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable11/scripts/0_2_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276683", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable11/src/0_2_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276684", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable12/scripts/0_2_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276684", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable12/src/0_2_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276685", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable13/scripts/0_2_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276685", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable13/src/0_2_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276686", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable14/scripts/0_2_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276687", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable14/src/0_2_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276687", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable15/scripts/0_2_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276688", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable15/src/0_2_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276688", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable16/scripts/0_2_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276689", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable16/src/0_2_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276690", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable17/scripts/0_2_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276690", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable17/src/0_2_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276691", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable18/scripts/0_2_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276691", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable18/src/0_2_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276692", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable19/scripts/0_2_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276692", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable19/src/0_2_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276693", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable20/scripts/0_2_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276693", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable20/src/0_2_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276694", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable21/scripts/0_2_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276694", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable21/src/0_2_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276695", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable22/scripts/0_2_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276695", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable22/src/0_2_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276696", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable23/scripts/0_2_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276697", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable23/src/0_2_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276697", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable24/scripts/0_2_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276698", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable24/src/0_2_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276698", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable25/scripts/0_2_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276699", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable25/src/0_2_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276700", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable0/scripts/0_3_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276700", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable0/src/0_3_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276701", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable1/scripts/0_3_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276701", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable1/src/0_3_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276702", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable2/scripts/0_3_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276702", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable2/src/0_3_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276703", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable3/scripts/0_3_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276704", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable3/src/0_3_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276704", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable4/scripts/0_3_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276705", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable4/src/0_3_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276706", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable5/scripts/0_3_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276706", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable5/src/0_3_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276707", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable6/scripts/0_3_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276707", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable6/src/0_3_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276708", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable7/scripts/0_3_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276708", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable7/src/0_3_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276709", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable8/scripts/0_3_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276709", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable8/src/0_3_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276710", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable9/scripts/0_3_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276710", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable9/src/0_3_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276711", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable10/scripts/0_3_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276711", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable10/src/0_3_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276712", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable11/scripts/0_3_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276712", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable11/src/0_3_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276713", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable12/scripts/0_3_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276713", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable12/src/0_3_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276714", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable13/scripts/0_3_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276714", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable13/src/0_3_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276715", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable14/scripts/0_3_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276715", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable14/src/0_3_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276716", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable15/scripts/0_3_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276716", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable15/src/0_3_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276717", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable16/scripts/0_3_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276717", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable16/src/0_3_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276718", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable17/scripts/0_3_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276718", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable17/src/0_3_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276719", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable18/scripts/0_3_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276719", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable18/src/0_3_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276720", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable19/scripts/0_3_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276720", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable19/src/0_3_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276721", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable20/scripts/0_3_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276721", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable20/src/0_3_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276722", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable21/scripts/0_3_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276722", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable21/src/0_3_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276723", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable22/scripts/0_3_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276723", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable22/src/0_3_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276724", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable23/scripts/0_3_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276725", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable23/src/0_3_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276725", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable24/scripts/0_3_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276726", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable24/src/0_3_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276726", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable25/scripts/0_3_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276727", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable25/src/0_3_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276727", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable0/scripts/1_0_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276728", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable0/src/1_0_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276728", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable1/scripts/1_0_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276729", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable1/src/1_0_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276729", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable2/scripts/1_0_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276730", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable2/src/1_0_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276730", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable3/scripts/1_0_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276731", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable3/src/1_0_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276731", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable4/scripts/1_0_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276732", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable4/src/1_0_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276732", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable5/scripts/1_0_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276733", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable5/src/1_0_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276733", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable6/scripts/1_0_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276734", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable6/src/1_0_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276734", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable7/scripts/1_0_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276735", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable7/src/1_0_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276736", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable8/scripts/1_0_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276736", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable8/src/1_0_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276737", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable9/scripts/1_0_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276737", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable9/src/1_0_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276738", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable10/scripts/1_0_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276738", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable10/src/1_0_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276739", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable11/scripts/1_0_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276739", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable11/src/1_0_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276740", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable12/scripts/1_0_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276741", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable12/src/1_0_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276741", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable13/scripts/1_0_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276742", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable13/src/1_0_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276742", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable14/scripts/1_0_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276743", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable14/src/1_0_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276743", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable15/scripts/1_0_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276744", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable15/src/1_0_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276744", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable16/scripts/1_0_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276745", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable16/src/1_0_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276745", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable17/scripts/1_0_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276746", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable17/src/1_0_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276746", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable18/scripts/1_0_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276747", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable18/src/1_0_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276747", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable19/scripts/1_0_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276748", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable19/src/1_0_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276748", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable20/scripts/1_0_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276749", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable20/src/1_0_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276749", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable21/scripts/1_0_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276750", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable21/src/1_0_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276750", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable22/scripts/1_0_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276751", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable22/src/1_0_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276751", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable23/scripts/1_0_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276752", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable23/src/1_0_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276752", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable24/scripts/1_0_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276753", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable24/src/1_0_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276753", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable25/scripts/1_0_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276754", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable25/src/1_0_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276754", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable0/scripts/1_1_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276755", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable0/src/1_1_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276755", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable1/scripts/1_1_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276756", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable1/src/1_1_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276756", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable2/scripts/1_1_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276757", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable2/src/1_1_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276757", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable3/scripts/1_1_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276758", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable3/src/1_1_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276758", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable4/scripts/1_1_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276759", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable4/src/1_1_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276759", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable5/scripts/1_1_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276760", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable5/src/1_1_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276760", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable6/scripts/1_1_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276761", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable6/src/1_1_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276762", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable7/scripts/1_1_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276762", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable7/src/1_1_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276763", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable8/scripts/1_1_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276763", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable8/src/1_1_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276764", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable9/scripts/1_1_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276764", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable9/src/1_1_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276765", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable10/scripts/1_1_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276765", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable10/src/1_1_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276766", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable11/scripts/1_1_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276766", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable11/src/1_1_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276767", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable12/scripts/1_1_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276767", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable12/src/1_1_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276768", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable13/scripts/1_1_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276768", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable13/src/1_1_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276769", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable14/scripts/1_1_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276769", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable14/src/1_1_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276770", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable15/scripts/1_1_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276770", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable15/src/1_1_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276771", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable16/scripts/1_1_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276771", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable16/src/1_1_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276772", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable17/scripts/1_1_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276772", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable17/src/1_1_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276773", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable18/scripts/1_1_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276773", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable18/src/1_1_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276774", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable19/scripts/1_1_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276774", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable19/src/1_1_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276775", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable20/scripts/1_1_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276775", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable20/src/1_1_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276776", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable21/scripts/1_1_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276776", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable21/src/1_1_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276777", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable22/scripts/1_1_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276777", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable22/src/1_1_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276778", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable23/scripts/1_1_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276778", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable23/src/1_1_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276779", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable24/scripts/1_1_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276779", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable24/src/1_1_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276780", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable25/scripts/1_1_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276780", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable25/src/1_1_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276781", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable0/scripts/1_2_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276781", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable0/src/1_2_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276782", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable1/scripts/1_2_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276782", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable1/src/1_2_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276783", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable2/scripts/1_2_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276783", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable2/src/1_2_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276784", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable3/scripts/1_2_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276784", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable3/src/1_2_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276785", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable4/scripts/1_2_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276785", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable4/src/1_2_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276786", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable5/scripts/1_2_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276786", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable5/src/1_2_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276787", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable6/scripts/1_2_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276787", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable6/src/1_2_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276788", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable7/scripts/1_2_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276788", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable7/src/1_2_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276789", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable8/scripts/1_2_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276790", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable8/src/1_2_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276790", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable9/scripts/1_2_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276791", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable9/src/1_2_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276791", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable10/scripts/1_2_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276792", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable10/src/1_2_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276792", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable11/scripts/1_2_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276793", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable11/src/1_2_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276793", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable12/scripts/1_2_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276794", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable12/src/1_2_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276794", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable13/scripts/1_2_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276795", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable13/src/1_2_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276795", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable14/scripts/1_2_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276796", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable14/src/1_2_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276797", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable15/scripts/1_2_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276797", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable15/src/1_2_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276798", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable16/scripts/1_2_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276798", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable16/src/1_2_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276799", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable17/scripts/1_2_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276799", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable17/src/1_2_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276800", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable18/scripts/1_2_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276800", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable18/src/1_2_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276801", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable19/scripts/1_2_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276801", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable19/src/1_2_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276802", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable20/scripts/1_2_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276803", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable20/src/1_2_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276803", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable21/scripts/1_2_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276804", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable21/src/1_2_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276804", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable22/scripts/1_2_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276805", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable22/src/1_2_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276805", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable23/scripts/1_2_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276806", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable23/src/1_2_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276806", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable24/scripts/1_2_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276807", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable24/src/1_2_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276807", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable25/scripts/1_2_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276808", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable25/src/1_2_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276808", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable0/scripts/1_3_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276809", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable0/src/1_3_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276809", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable1/scripts/1_3_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276810", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable1/src/1_3_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276810", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable2/scripts/1_3_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276811", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable2/src/1_3_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276811", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable3/scripts/1_3_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276812", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable3/src/1_3_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276812", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable4/scripts/1_3_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276813", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable4/src/1_3_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276813", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable5/scripts/1_3_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276814", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable5/src/1_3_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276814", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable6/scripts/1_3_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276815", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable6/src/1_3_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276815", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable7/scripts/1_3_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276816", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable7/src/1_3_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276817", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable8/scripts/1_3_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276817", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable8/src/1_3_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276818", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable9/scripts/1_3_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276818", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable9/src/1_3_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276819", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable10/scripts/1_3_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276819", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable10/src/1_3_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276820", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable11/scripts/1_3_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276820", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable11/src/1_3_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276821", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable12/scripts/1_3_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276821", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable12/src/1_3_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276822", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable13/scripts/1_3_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276822", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable13/src/1_3_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276823", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable14/scripts/1_3_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276823", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable14/src/1_3_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276824", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable15/scripts/1_3_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276824", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable15/src/1_3_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276825", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable16/scripts/1_3_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276826", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable16/src/1_3_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276826", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable17/scripts/1_3_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276827", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable17/src/1_3_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276827", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable18/scripts/1_3_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276827", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable18/src/1_3_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276828", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable19/scripts/1_3_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276829", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable19/src/1_3_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276829", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable20/scripts/1_3_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276830", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable20/src/1_3_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276830", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable21/scripts/1_3_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276831", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable21/src/1_3_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276831", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable22/scripts/1_3_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276832", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable22/src/1_3_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276832", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable23/scripts/1_3_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276833", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable23/src/1_3_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276833", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable24/scripts/1_3_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276834", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable24/src/1_3_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276834", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable25/scripts/1_3_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276835", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable25/src/1_3_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276835", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable0/scripts/2_0_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276836", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable0/src/2_0_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276836", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable1/scripts/2_0_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276837", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable1/src/2_0_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276837", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable2/scripts/2_0_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276838", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable2/src/2_0_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276838", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable3/scripts/2_0_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276839", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable3/src/2_0_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276839", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable4/scripts/2_0_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276840", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable4/src/2_0_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276840", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable5/scripts/2_0_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276841", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable5/src/2_0_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276841", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable6/scripts/2_0_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276842", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable6/src/2_0_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276842", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable7/scripts/2_0_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276843", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable7/src/2_0_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276844", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable8/scripts/2_0_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276844", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable8/src/2_0_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276845", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable9/scripts/2_0_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276845", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable9/src/2_0_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276846", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable10/scripts/2_0_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276846", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable10/src/2_0_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276847", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable11/scripts/2_0_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276847", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable11/src/2_0_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276848", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable12/scripts/2_0_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276848", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable12/src/2_0_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276849", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable13/scripts/2_0_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276849", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable13/src/2_0_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276850", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable14/scripts/2_0_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276850", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable14/src/2_0_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276851", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable15/scripts/2_0_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276851", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable15/src/2_0_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276852", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable16/scripts/2_0_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276852", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable16/src/2_0_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276853", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable17/scripts/2_0_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276853", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable17/src/2_0_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276854", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable18/scripts/2_0_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276854", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable18/src/2_0_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276855", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable19/scripts/2_0_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276855", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable19/src/2_0_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276856", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable20/scripts/2_0_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276856", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable20/src/2_0_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276857", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable21/scripts/2_0_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276857", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable21/src/2_0_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276858", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable22/scripts/2_0_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276858", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable22/src/2_0_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276859", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable23/scripts/2_0_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276859", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable23/src/2_0_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276860", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable24/scripts/2_0_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276860", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable24/src/2_0_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276861", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable25/scripts/2_0_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276861", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable25/src/2_0_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276862", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable0/scripts/2_1_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276862", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable0/src/2_1_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276863", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable1/scripts/2_1_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276863", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable1/src/2_1_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276864", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable2/scripts/2_1_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276864", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable2/src/2_1_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276865", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable3/scripts/2_1_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276865", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable3/src/2_1_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276866", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable4/scripts/2_1_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276866", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable4/src/2_1_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276867", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable5/scripts/2_1_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276867", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable5/src/2_1_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276868", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable6/scripts/2_1_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276868", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable6/src/2_1_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276869", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable7/scripts/2_1_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276869", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable7/src/2_1_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276870", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable8/scripts/2_1_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276870", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable8/src/2_1_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276871", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable9/scripts/2_1_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276871", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable9/src/2_1_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276872", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable10/scripts/2_1_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276872", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable10/src/2_1_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276873", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable11/scripts/2_1_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276873", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable11/src/2_1_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276874", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable12/scripts/2_1_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276874", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable12/src/2_1_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276875", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable13/scripts/2_1_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276875", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable13/src/2_1_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276876", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable14/scripts/2_1_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276877", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable14/src/2_1_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276877", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable15/scripts/2_1_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276878", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable15/src/2_1_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276878", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable16/scripts/2_1_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276879", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable16/src/2_1_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276879", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable17/scripts/2_1_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276880", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable17/src/2_1_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276880", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable18/scripts/2_1_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276881", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable18/src/2_1_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276881", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable19/scripts/2_1_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276882", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable19/src/2_1_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276882", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable20/scripts/2_1_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276883", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable20/src/2_1_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276883", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable21/scripts/2_1_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276884", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable21/src/2_1_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276884", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable22/scripts/2_1_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276885", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable22/src/2_1_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276885", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable23/scripts/2_1_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276886", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable23/src/2_1_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276887", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable24/scripts/2_1_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276887", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable24/src/2_1_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276888", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable25/scripts/2_1_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276888", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable25/src/2_1_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276889", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable0/scripts/2_2_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276889", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable0/src/2_2_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276890", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable1/scripts/2_2_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276890", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable1/src/2_2_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276891", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable2/scripts/2_2_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276891", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable2/src/2_2_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276892", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable3/scripts/2_2_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276892", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable3/src/2_2_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276893", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable4/scripts/2_2_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276893", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable4/src/2_2_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276894", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable5/scripts/2_2_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276894", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable5/src/2_2_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276894", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable6/scripts/2_2_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276895", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable6/src/2_2_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276895", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable7/scripts/2_2_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276896", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable7/src/2_2_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276897", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable8/scripts/2_2_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276897", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable8/src/2_2_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276898", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable9/scripts/2_2_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276898", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable9/src/2_2_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276899", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable10/scripts/2_2_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276899", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable10/src/2_2_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276900", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable11/scripts/2_2_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276900", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable11/src/2_2_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276901", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable12/scripts/2_2_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276901", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable12/src/2_2_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276902", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable13/scripts/2_2_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276902", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable13/src/2_2_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276903", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable14/scripts/2_2_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276903", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable14/src/2_2_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276904", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable15/scripts/2_2_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276904", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable15/src/2_2_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276905", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable16/scripts/2_2_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276905", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable16/src/2_2_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276906", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable17/scripts/2_2_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276906", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable17/src/2_2_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276907", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable18/scripts/2_2_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276907", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable18/src/2_2_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276908", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable19/scripts/2_2_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276908", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable19/src/2_2_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276909", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable20/scripts/2_2_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276909", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable20/src/2_2_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276910", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable21/scripts/2_2_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276910", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable21/src/2_2_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276911", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable22/scripts/2_2_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276911", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable22/src/2_2_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276912", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable23/scripts/2_2_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276912", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable23/src/2_2_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276913", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable24/scripts/2_2_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276913", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable24/src/2_2_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276914", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable25/scripts/2_2_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276914", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable25/src/2_2_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276915", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable0/scripts/2_3_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276915", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable0/src/2_3_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276916", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable1/scripts/2_3_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276916", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable1/src/2_3_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276917", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable2/scripts/2_3_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276917", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable2/src/2_3_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276918", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable3/scripts/2_3_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276918", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable3/src/2_3_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276919", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable4/scripts/2_3_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276919", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable4/src/2_3_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276920", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable5/scripts/2_3_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276920", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable5/src/2_3_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276921", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable6/scripts/2_3_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276921", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable6/src/2_3_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276922", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable7/scripts/2_3_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276922", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable7/src/2_3_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276923", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable8/scripts/2_3_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276923", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable8/src/2_3_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276924", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable9/scripts/2_3_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276925", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable9/src/2_3_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276925", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable10/scripts/2_3_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276925", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable10/src/2_3_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276926", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable11/scripts/2_3_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276926", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable11/src/2_3_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276927", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable12/scripts/2_3_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276927", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable12/src/2_3_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276928", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable13/scripts/2_3_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276929", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable13/src/2_3_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276929", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable14/scripts/2_3_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276930", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable14/src/2_3_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276930", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable15/scripts/2_3_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276931", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable15/src/2_3_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276931", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable16/scripts/2_3_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276932", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable16/src/2_3_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276932", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable17/scripts/2_3_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276933", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable17/src/2_3_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276933", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable18/scripts/2_3_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276934", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable18/src/2_3_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276934", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable19/scripts/2_3_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276935", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable19/src/2_3_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276935", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable20/scripts/2_3_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276936", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable20/src/2_3_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276936", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable21/scripts/2_3_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276937", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable21/src/2_3_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276937", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable22/scripts/2_3_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276938", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable22/src/2_3_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276938", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable23/scripts/2_3_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276939", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable23/src/2_3_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276939", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable24/scripts/2_3_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276940", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable24/src/2_3_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276940", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable25/scripts/2_3_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276941", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable25/src/2_3_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276941", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable0/scripts/3_0_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276942", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable0/src/3_0_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276942", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable1/scripts/3_0_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276943", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable1/src/3_0_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276944", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable2/scripts/3_0_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276944", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable2/src/3_0_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276945", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable3/scripts/3_0_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276945", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable3/src/3_0_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276946", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable4/scripts/3_0_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276946", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable4/src/3_0_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276947", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable5/scripts/3_0_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276947", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable5/src/3_0_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276948", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable6/scripts/3_0_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276948", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable6/src/3_0_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276949", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable7/scripts/3_0_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276949", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable7/src/3_0_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276950", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable8/scripts/3_0_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276950", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable8/src/3_0_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276951", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable9/scripts/3_0_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276951", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable9/src/3_0_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276952", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable10/scripts/3_0_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276952", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable10/src/3_0_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276953", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable11/scripts/3_0_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276953", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable11/src/3_0_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276954", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable12/scripts/3_0_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276954", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable12/src/3_0_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276955", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable13/scripts/3_0_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276955", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable13/src/3_0_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276956", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable14/scripts/3_0_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276956", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable14/src/3_0_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276957", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable15/scripts/3_0_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276957", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable15/src/3_0_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276958", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable16/scripts/3_0_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276958", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable16/src/3_0_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276959", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable17/scripts/3_0_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276959", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable17/src/3_0_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276960", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable18/scripts/3_0_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276960", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable18/src/3_0_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276961", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable19/scripts/3_0_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276961", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable19/src/3_0_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276962", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable20/scripts/3_0_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276962", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable20/src/3_0_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276963", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable21/scripts/3_0_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276963", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable21/src/3_0_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276964", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable22/scripts/3_0_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276964", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable22/src/3_0_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276965", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable23/scripts/3_0_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276965", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable23/src/3_0_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276966", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable24/scripts/3_0_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276966", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable24/src/3_0_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276967", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable25/scripts/3_0_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276967", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable25/src/3_0_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276968", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable0/scripts/3_1_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276968", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable0/src/3_1_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276969", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable1/scripts/3_1_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276969", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable1/src/3_1_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276970", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable2/scripts/3_1_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276970", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable2/src/3_1_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276971", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable3/scripts/3_1_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276971", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable3/src/3_1_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276972", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable4/scripts/3_1_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276972", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable4/src/3_1_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276973", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable5/scripts/3_1_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276973", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable5/src/3_1_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276974", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable6/scripts/3_1_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276974", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable6/src/3_1_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276975", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable7/scripts/3_1_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276975", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable7/src/3_1_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276976", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable8/scripts/3_1_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276977", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable8/src/3_1_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276977", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable9/scripts/3_1_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276978", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable9/src/3_1_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276978", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable10/scripts/3_1_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276979", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable10/src/3_1_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276979", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable11/scripts/3_1_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276980", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable11/src/3_1_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276980", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable12/scripts/3_1_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276981", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable12/src/3_1_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276981", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable13/scripts/3_1_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276982", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable13/src/3_1_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276982", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable14/scripts/3_1_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276983", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable14/src/3_1_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276983", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable15/scripts/3_1_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276984", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable15/src/3_1_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276984", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable16/scripts/3_1_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276985", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable16/src/3_1_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276985", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable17/scripts/3_1_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276986", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable17/src/3_1_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276986", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable18/scripts/3_1_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276987", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable18/src/3_1_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276988", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable19/scripts/3_1_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276988", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable19/src/3_1_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276989", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable20/scripts/3_1_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276989", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable20/src/3_1_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276990", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable21/scripts/3_1_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276990", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable21/src/3_1_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276991", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable22/scripts/3_1_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276991", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable22/src/3_1_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276992", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable23/scripts/3_1_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276992", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable23/src/3_1_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276993", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable24/scripts/3_1_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276993", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable24/src/3_1_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276994", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable25/scripts/3_1_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276994", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable25/src/3_1_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276995", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable0/scripts/3_2_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276995", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable0/src/3_2_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276996", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable1/scripts/3_2_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276996", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable1/src/3_2_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276997", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable2/scripts/3_2_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276997", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable2/src/3_2_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276998", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable3/scripts/3_2_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276998", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable3/src/3_2_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276999", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable4/scripts/3_2_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:56 2025", + "timestampMillis": "1742528276999", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable4/src/3_2_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277000", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable5/scripts/3_2_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277000", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable5/src/3_2_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277001", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable6/scripts/3_2_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277001", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable6/src/3_2_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277002", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable7/scripts/3_2_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277002", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable7/src/3_2_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277003", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable8/scripts/3_2_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277003", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable8/src/3_2_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277004", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable9/scripts/3_2_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277004", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable9/src/3_2_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277005", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable10/scripts/3_2_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277005", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable10/src/3_2_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277006", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable11/scripts/3_2_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277006", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable11/src/3_2_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277007", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable12/scripts/3_2_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277007", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable12/src/3_2_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277008", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable13/scripts/3_2_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277009", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable13/src/3_2_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277009", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable14/scripts/3_2_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277010", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable14/src/3_2_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277010", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable15/scripts/3_2_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277011", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable15/src/3_2_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277011", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable16/scripts/3_2_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277012", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable16/src/3_2_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277012", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable17/scripts/3_2_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277013", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable17/src/3_2_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277013", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable18/scripts/3_2_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277014", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable18/src/3_2_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277014", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable19/scripts/3_2_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277015", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable19/src/3_2_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277015", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable20/scripts/3_2_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277016", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable20/src/3_2_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277016", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable21/scripts/3_2_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277017", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable21/src/3_2_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277017", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable22/scripts/3_2_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277018", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable22/src/3_2_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277018", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable23/scripts/3_2_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277019", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable23/src/3_2_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277019", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable24/scripts/3_2_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277020", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable24/src/3_2_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277020", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable25/scripts/3_2_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277021", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable25/src/3_2_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277021", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable0/scripts/3_3_reloadable0.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277022", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable0/src/3_3_reloadable0.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277022", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable1/scripts/3_3_reloadable1.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277023", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable1/src/3_3_reloadable1.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277023", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable2/scripts/3_3_reloadable2.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277024", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable2/src/3_3_reloadable2.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277024", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable3/scripts/3_3_reloadable3.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277025", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable3/src/3_3_reloadable3.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277025", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable4/scripts/3_3_reloadable4.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277026", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable4/src/3_3_reloadable4.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277026", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable5/scripts/3_3_reloadable5.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277027", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable5/src/3_3_reloadable5.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277027", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable6/scripts/3_3_reloadable6.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277028", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable6/src/3_3_reloadable6.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277028", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable7/scripts/3_3_reloadable7.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277029", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable7/src/3_3_reloadable7.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277030", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable8/scripts/3_3_reloadable8.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277030", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable8/src/3_3_reloadable8.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277031", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable9/scripts/3_3_reloadable9.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277031", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable9/src/3_3_reloadable9.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277032", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable10/scripts/3_3_reloadable10.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277032", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable10/src/3_3_reloadable10.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277033", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable11/scripts/3_3_reloadable11.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277033", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable11/src/3_3_reloadable11.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277034", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable12/scripts/3_3_reloadable12.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277034", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable12/src/3_3_reloadable12.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277035", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable13/scripts/3_3_reloadable13.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277035", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable13/src/3_3_reloadable13.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277036", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable14/scripts/3_3_reloadable14.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277036", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable14/src/3_3_reloadable14.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277037", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable15/scripts/3_3_reloadable15.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277037", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable15/src/3_3_reloadable15.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277038", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable16/scripts/3_3_reloadable16.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277038", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable16/src/3_3_reloadable16.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277039", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable17/scripts/3_3_reloadable17.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277039", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable17/src/3_3_reloadable17.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277040", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable18/scripts/3_3_reloadable18.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277040", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable18/src/3_3_reloadable18.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277041", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable19/scripts/3_3_reloadable19.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277041", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable19/src/3_3_reloadable19.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277042", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable20/scripts/3_3_reloadable20.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277042", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable20/src/3_3_reloadable20.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277043", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable21/scripts/3_3_reloadable21.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277043", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable21/src/3_3_reloadable21.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277044", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable22/scripts/3_3_reloadable22.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277044", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable22/src/3_3_reloadable22.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277045", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable23/scripts/3_3_reloadable23.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277045", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable23/src/3_3_reloadable23.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277046", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable24/scripts/3_3_reloadable24.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277046", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable24/src/3_3_reloadable24.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277047", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable25/scripts/3_3_reloadable25.prx", + "name": "", + "fileType": "XML", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277047", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable25/src/3_3_reloadable25.cc", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_PROJECT", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277057", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/0_0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277059", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable0/0_0_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277061", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable1/0_0_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277069", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable2/0_0_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277073", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable3/0_0_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277074", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable4/0_0_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277078", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable5/0_0_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277079", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable6/0_0_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277080", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable7/0_0_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277082", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable8/0_0_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277083", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable9/0_0_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277084", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable10/0_0_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277085", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable11/0_0_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277086", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable12/0_0_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277087", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable13/0_0_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277088", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable14/0_0_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277089", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable15/0_0_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277091", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable16/0_0_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277092", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable17/0_0_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277094", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable18/0_0_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277095", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable19/0_0_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277097", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable20/0_0_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277099", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable21/0_0_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277101", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/0_0_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277103", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/0_0_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277104", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable24/0_0_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277106", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable25/0_0_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277107", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1/0_1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277108", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable0/0_1_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277109", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable1/0_1_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277110", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable2/0_1_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277112", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable3/0_1_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277113", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable4/0_1_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277114", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable5/0_1_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277115", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable6/0_1_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277116", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable7/0_1_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277117", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable8/0_1_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277118", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable9/0_1_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277119", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable10/0_1_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277120", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable11/0_1_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277121", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable12/0_1_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277123", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable13/0_1_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277124", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable14/0_1_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277125", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable15/0_1_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277126", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable16/0_1_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277127", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable17/0_1_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277128", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable18/0_1_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277129", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable19/0_1_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277130", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable20/0_1_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277131", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable21/0_1_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277132", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable22/0_1_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277133", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable23/0_1_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277134", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable24/0_1_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277136", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_1_reloadable25/0_1_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277137", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2/0_2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277138", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable0/0_2_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277139", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable1/0_2_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277140", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable2/0_2_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277141", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable3/0_2_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277142", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable4/0_2_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277143", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable5/0_2_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277144", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable6/0_2_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277145", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable7/0_2_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277146", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable8/0_2_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277147", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable9/0_2_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277148", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable10/0_2_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277149", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable11/0_2_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277150", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable12/0_2_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277152", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable13/0_2_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277153", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable14/0_2_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277154", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable15/0_2_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277155", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable16/0_2_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277156", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable17/0_2_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277157", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable18/0_2_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277158", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable19/0_2_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277159", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable20/0_2_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277160", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable21/0_2_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277161", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable22/0_2_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277162", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable23/0_2_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277163", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable24/0_2_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277164", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_2_reloadable25/0_2_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277166", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3/0_3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277167", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable0/0_3_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277168", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable1/0_3_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277169", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable2/0_3_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277170", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable3/0_3_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277171", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable4/0_3_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277172", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable5/0_3_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277174", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable6/0_3_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277175", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable7/0_3_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277176", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable8/0_3_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277177", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable9/0_3_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277178", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable10/0_3_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277179", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable11/0_3_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277180", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable12/0_3_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277182", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable13/0_3_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277183", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable14/0_3_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277184", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable15/0_3_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277185", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable16/0_3_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277186", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable17/0_3_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277187", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable18/0_3_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277188", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable19/0_3_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277189", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable20/0_3_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277191", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable21/0_3_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277192", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable22/0_3_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277193", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable23/0_3_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277194", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable24/0_3_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277195", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_3_reloadable25/0_3_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277196", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0/1_0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277197", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable0/1_0_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277198", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable1/1_0_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277199", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable2/1_0_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277200", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable3/1_0_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277202", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable4/1_0_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277203", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable5/1_0_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277204", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable6/1_0_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277205", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable7/1_0_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277206", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable8/1_0_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277207", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable9/1_0_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277208", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable10/1_0_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277209", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable11/1_0_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277210", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable12/1_0_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277212", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable13/1_0_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277213", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable14/1_0_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277214", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable15/1_0_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277215", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable16/1_0_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277216", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable17/1_0_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277217", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable18/1_0_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277219", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable19/1_0_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277220", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable20/1_0_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277221", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable21/1_0_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277222", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable22/1_0_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277223", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable23/1_0_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277224", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable24/1_0_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277225", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_0_reloadable25/1_0_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277226", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1/1_1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277228", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable0/1_1_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277229", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable1/1_1_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277230", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable2/1_1_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277231", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable3/1_1_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277232", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable4/1_1_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277233", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable5/1_1_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277234", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable6/1_1_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277235", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable7/1_1_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277236", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable8/1_1_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277237", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable9/1_1_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277239", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable10/1_1_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277240", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable11/1_1_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277241", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable12/1_1_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277242", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable13/1_1_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277243", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable14/1_1_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277244", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable15/1_1_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277245", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable16/1_1_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277247", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable17/1_1_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277248", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable18/1_1_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277249", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable19/1_1_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277250", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable20/1_1_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277251", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable21/1_1_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277252", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable22/1_1_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277254", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable23/1_1_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277255", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable24/1_1_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277256", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_1_reloadable25/1_1_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277257", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2/1_2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277259", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable0/1_2_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277260", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable1/1_2_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277261", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable2/1_2_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277262", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable3/1_2_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277263", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable4/1_2_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277264", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable5/1_2_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277265", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable6/1_2_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277267", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable7/1_2_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277268", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable8/1_2_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277269", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable9/1_2_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277270", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable10/1_2_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277271", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable11/1_2_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277272", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable12/1_2_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277273", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable13/1_2_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277274", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable14/1_2_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277276", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable15/1_2_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277277", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable16/1_2_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277278", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable17/1_2_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277279", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable18/1_2_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277280", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable19/1_2_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277281", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable20/1_2_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277282", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable21/1_2_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277283", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable22/1_2_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277284", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable23/1_2_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277285", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable24/1_2_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277287", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_2_reloadable25/1_2_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277288", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3/1_3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277289", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable0/1_3_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277290", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable1/1_3_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277292", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable2/1_3_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277293", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable3/1_3_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277294", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable4/1_3_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277295", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable5/1_3_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277296", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable6/1_3_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277298", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable7/1_3_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277299", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable8/1_3_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277300", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable9/1_3_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277301", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable10/1_3_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277302", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable11/1_3_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277303", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable12/1_3_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277304", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable13/1_3_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277306", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable14/1_3_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277307", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable15/1_3_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277308", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable16/1_3_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277309", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable17/1_3_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277310", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable18/1_3_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277311", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable19/1_3_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277312", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable20/1_3_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277313", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable21/1_3_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277314", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable22/1_3_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277315", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable23/1_3_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277317", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable24/1_3_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277318", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/1_3_reloadable25/1_3_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277319", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0/2_0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277320", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable0/2_0_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277321", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable1/2_0_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277322", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable2/2_0_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277323", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable3/2_0_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277325", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable4/2_0_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277326", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable5/2_0_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277327", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable6/2_0_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277328", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable7/2_0_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277329", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable8/2_0_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277330", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable9/2_0_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277331", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable10/2_0_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277332", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable11/2_0_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277333", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable12/2_0_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277334", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable13/2_0_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277335", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable14/2_0_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277336", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable15/2_0_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277338", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable16/2_0_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277339", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable17/2_0_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277340", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable18/2_0_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277341", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable19/2_0_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277342", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable20/2_0_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277343", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable21/2_0_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277344", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable22/2_0_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277345", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable23/2_0_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277346", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable24/2_0_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277347", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_0_reloadable25/2_0_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277349", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1/2_1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277350", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable0/2_1_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277351", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable1/2_1_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277352", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable2/2_1_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277353", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable3/2_1_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277354", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable4/2_1_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277355", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable5/2_1_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277356", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable6/2_1_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277357", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable7/2_1_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277358", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable8/2_1_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277360", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable9/2_1_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277361", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable10/2_1_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277362", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable11/2_1_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277363", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable12/2_1_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277364", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable13/2_1_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277365", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable14/2_1_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277366", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable15/2_1_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277367", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable16/2_1_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277369", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable17/2_1_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277370", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable18/2_1_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277371", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable19/2_1_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277372", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable20/2_1_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277373", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable21/2_1_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277374", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable22/2_1_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277375", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable23/2_1_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277377", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable24/2_1_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277378", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_1_reloadable25/2_1_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277379", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2/2_2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277380", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable0/2_2_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277381", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable1/2_2_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277382", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable2/2_2_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277383", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable3/2_2_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277384", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable4/2_2_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277385", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable5/2_2_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277386", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable6/2_2_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277388", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable7/2_2_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277389", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable8/2_2_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277390", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable9/2_2_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277391", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable10/2_2_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277392", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable11/2_2_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277394", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable12/2_2_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277395", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable13/2_2_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277396", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable14/2_2_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277397", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable15/2_2_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277398", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable16/2_2_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277399", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable17/2_2_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277400", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable18/2_2_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277401", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable19/2_2_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277403", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable20/2_2_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277404", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable21/2_2_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277405", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable22/2_2_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277406", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable23/2_2_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277407", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable24/2_2_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277408", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_2_reloadable25/2_2_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277409", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3/2_3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277410", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable0/2_3_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277411", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable1/2_3_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277413", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable2/2_3_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277414", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable3/2_3_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277415", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable4/2_3_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277416", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable5/2_3_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277417", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable6/2_3_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277418", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable7/2_3_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277419", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable8/2_3_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277420", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable9/2_3_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277421", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable10/2_3_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277423", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable11/2_3_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277424", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable12/2_3_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277425", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable13/2_3_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277426", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable14/2_3_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277427", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable15/2_3_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277428", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable16/2_3_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277429", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable17/2_3_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277430", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable18/2_3_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277431", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable19/2_3_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277432", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable20/2_3_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277434", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable21/2_3_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277435", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable22/2_3_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277436", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable23/2_3_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277437", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable24/2_3_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277438", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/2_3_reloadable25/2_3_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277439", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0/3_0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277440", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable0/3_0_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277441", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable1/3_0_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277442", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable2/3_0_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277444", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable3/3_0_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277445", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable4/3_0_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277446", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable5/3_0_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277447", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable6/3_0_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277448", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable7/3_0_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277449", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable8/3_0_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277450", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable9/3_0_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277451", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable10/3_0_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277452", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable11/3_0_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277453", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable12/3_0_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277454", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable13/3_0_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277455", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable14/3_0_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277457", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable15/3_0_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277458", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable16/3_0_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277459", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable17/3_0_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277460", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable18/3_0_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277461", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable19/3_0_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277462", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable20/3_0_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277463", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable21/3_0_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277464", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable22/3_0_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277465", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable23/3_0_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277466", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable24/3_0_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277467", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_0_reloadable25/3_0_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277469", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1/3_1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277470", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable0/3_1_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277471", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable1/3_1_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277472", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable2/3_1_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277473", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable3/3_1_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277474", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable4/3_1_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277475", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable5/3_1_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277476", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable6/3_1_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277477", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable7/3_1_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277478", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable8/3_1_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277479", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable9/3_1_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277480", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable10/3_1_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277481", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable11/3_1_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277483", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable12/3_1_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277484", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable13/3_1_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277485", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable14/3_1_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277486", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable15/3_1_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277487", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable16/3_1_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277488", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable17/3_1_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277489", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable18/3_1_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277490", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable19/3_1_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277491", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable20/3_1_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277492", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable21/3_1_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277494", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable22/3_1_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277495", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable23/3_1_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277496", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable24/3_1_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277497", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_1_reloadable25/3_1_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277498", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2/3_2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277499", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable0/3_2_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277500", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable1/3_2_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277501", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable2/3_2_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277503", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable3/3_2_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277504", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable4/3_2_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277505", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable5/3_2_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277506", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable6/3_2_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277507", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable7/3_2_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277508", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable8/3_2_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277509", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable9/3_2_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277510", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable10/3_2_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277511", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable11/3_2_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277512", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable12/3_2_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277514", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable13/3_2_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277515", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable14/3_2_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277516", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable15/3_2_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277517", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable16/3_2_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277519", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable17/3_2_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277520", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable18/3_2_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277521", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable19/3_2_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277522", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable20/3_2_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277523", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable21/3_2_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277524", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable22/3_2_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277525", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable23/3_2_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277526", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable24/3_2_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277527", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_2_reloadable25/3_2_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277528", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3/3_3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277529", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable0/3_3_reloadable0.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277531", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable1/3_3_reloadable1.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277532", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable2/3_3_reloadable2.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277533", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable3/3_3_reloadable3.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277534", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable4/3_3_reloadable4.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277535", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable5/3_3_reloadable5.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277536", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable6/3_3_reloadable6.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277537", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable7/3_3_reloadable7.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277538", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable8/3_3_reloadable8.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277539", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable9/3_3_reloadable9.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277541", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable10/3_3_reloadable10.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277542", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable11/3_3_reloadable11.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277543", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable12/3_3_reloadable12.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277544", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable13/3_3_reloadable13.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277545", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable14/3_3_reloadable14.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277546", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable15/3_3_reloadable15.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277547", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable16/3_3_reloadable16.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277549", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable17/3_3_reloadable17.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277550", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable18/3_3_reloadable18.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277551", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable19/3_3_reloadable19.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277552", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable20/3_3_reloadable20.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277553", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable21/3_3_reloadable21.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277554", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable22/3_3_reloadable22.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277555", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable23/3_3_reloadable23.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277556", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable24/3_3_reloadable24.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Report", + "dateTimestamp": "Fri Mar 21 03:37:57 2025", + "timestampMillis": "1742528277557", + "report": { + "path": "/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/3_3_reloadable25/3_3_reloadable25.log", + "name": "", + "fileType": "TEXT", + "reportType": "AIE_CORE_COMPILE_LOG", + "cmdId": "" + } +} + + +{ + "type": "ET_Status", + "dateTimestamp": "Fri Mar 21 04:04:52 2025", + "timestampMillis": "1742529892428", + "status": { + "cmdId": "31edb987-7a66-4abb-b94a-830cd7d309ef", + "state": "CS_PASSED" + } +} + + +{ + "type": "ET_Status", + "dateTimestamp": "Fri Mar 21 04:04:52 2025", + "timestampMillis": "1742529892529", + "status": { + "cmdId": "dcd70920-0818-4355-83df-cd68be811a1d", + "state": "CS_PASSED" + } +} + diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/aie.mk b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/aie.mk new file mode 100644 index 0000000000000000000000000000000000000000..0d5d5ec962d01f629498f85cb7f8151a2841b0d2 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/aie.mk @@ -0,0 +1,2 @@ +aiecompile: + aiecompiler /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/top.cpp --part=xc10AIE2P_ML-die-0x-e-S-es1 --nodot-graph --runtime-opt=1 --disable-multirate-analysis --enable-core-processor-bus --enable-multi-layer --heapsize=1792 --stacksize=1400 --max-layer-ctrl-param-size=256 --compile-for-aiesim=false --workdir=Work --multi-layer-ctrl-pkt --aie2ipu-base-addr=0 -enable-light-cdo --Xelfgen=-j4 --multi-layer-pipelining --multi-layer-opt=3 --Xpreproc=-D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ --multi-layer-ext-buf-file=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/flexmlrt-hsi.json --enable-partition=0:4 --multi-layer-ctrl-pkt-column-span=4 --multi-layer-prebuilt-archive=/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras/data/ryzen-ai/stx/unified-overlay-4x4.json --multi-layer-prebuilt-archive-enable-elf-gen --multi-layer-init-core-elf-ctrl-pkt --multi-layer-pm-id 29006 --Xpreproc="-DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1" --include=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend --include=/usr/local/lib/python3.10/site-packages/include/aie_api --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf --output-archive libadf.a --adf-api-log-level=0 --multi-layer-pm-reloading=1 --Xelfgen=-j1 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/aiecompiler-flexml.log b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/aiecompiler-flexml.log new file mode 100644 index 0000000000000000000000000000000000000000..0cc8fa796a5e3b225d89e00008828f3764a179b6 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/aiecompiler-flexml.log @@ -0,0 +1,44420 @@ +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. + AI Engine Compiler + Version 2024.2.0 (linux64-bit) + SW Build 6074767 on 2025-03-20-05:22:32 + Copyright 1986-2022 Xilinx, Inc. All Rights Reserved. + Copyright 2022-2025 Advanced Micro Devices, Inc. All Rights Reserved. +INFO: ML part stats configuration file found in /usr/local/lib/python3.10/dist-packages/data +INFO: [aiecompiler 77-297] Cmd Line : /usr/local/lib/python3.10/dist-packages/bin/unwrapped/lnx64.o/aiecompiler /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/top.cpp --part=xc10AIE2P_ML-die-0x-e-S-es1 --nodot-graph --runtime-opt=1 --disable-multirate-analysis --enable-core-processor-bus --enable-multi-layer --heapsize=1792 --stacksize=1400 --max-layer-ctrl-param-size=256 --compile-for-aiesim=false --workdir=Work --multi-layer-ctrl-pkt --aie2ipu-base-addr=0 -enable-light-cdo --Xelfgen=-j4 --multi-layer-pipelining --multi-layer-opt=3 --Xpreproc=-D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ --multi-layer-ext-buf-file=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/flexmlrt-hsi.json --enable-partition=0:4 --multi-layer-ctrl-pkt-column-span=4 --multi-layer-prebuilt-archive=/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras/data/ryzen-ai/stx/unified-overlay-4x4.json --multi-layer-prebuilt-archive-enable-elf-gen --multi-layer-init-core-elf-ctrl-pkt --multi-layer-pm-id 29006 --Xpreproc=-DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 --include=/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend --include=/usr/local/lib/python3.10/site-packages/include/aie_api --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc --include=/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf --output-archive libadf.a --adf-api-log-level=0 --multi-layer-pm-reloading=1 --Xelfgen=-j1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +INFO: [aiecompiler 77-404] Executing Cmd: /usr/local/lib/python3.10/dist-packages/bin/../lnx64.o/tools/clang/bin/clang++ -I/usr/local/lib/python3.10/dist-packages/include -I/usr/local/lib/python3.10/dist-packages/tps/lnx64/gcc/include/c++/8.3.0 -I/usr/local/lib/python3.10/dist-packages/tps/lnx64/gcc/include/c++/8.3.0/x86_64-pc-linux-gnu -I/usr/local/lib/python3.10/dist-packages/tps/lnx64/gcc/include -Wno-error=reserved-user-defined-literal -E -std=c++17 -D__ADF_FRONTEND__ -D__AIE_ARCH__=21 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -I/usr/local/lib/python3.10/dist-packages/include -I . -I /app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/site-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/top.cpp > Work/temp/top.ii +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +INFO: [aiecompiler 77-404] Executing Cmd: graph_preprocessor Work/temp/top.ii -o Work/temp/top.processed.ii -report-core-dump -- -std=c++17 -ftemplate-depth=2048 -Wno-return-stack-address -Wno-missing-declarations -Wno-parentheses-equality -Wno-shift-negative-value +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +INFO: [aiecompiler 77-404] Executing Cmd: /usr/local/lib/python3.10/dist-packages/bin/../lnx64.o/tools/clang/bin/clang++ --gcc-install-dir=/usr/local/lib/python3.10/dist-packages/tps/lnx64/gcc/lib/gcc/x86_64-pc-linux-gnu/8.3.0 -stdlib=libstdc++ -Wno-error=reserved-user-defined-literal -std=c++17 -I . Work/temp/top.processed.ii -o Work/temp/top.out -L /usr/local/lib/python3.10/dist-packages/lib/lnx64.o -g -O0 -Wl,--unresolved-symbols=ignore-all -Wno-return-stack-address -Wno-missing-declarations -lmeir_frontend -ladf_api_frontend +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +INFO: [aiecompiler 77-404] Executing Cmd: Work/temp/top.out -I /app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/site-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf --workdir=Work --part=xc10AIE2P_ML-die-0x-e-S-es1 --aiearch=aie2p --log-level=1 --pl-axi-lite=0 --enable-multi-layer=1 --multi-layer-opt=3 --disable-dma-autostart=0 --enable-light-cdo=1 --large-program-memory=0 --swfifo-threshold=40 --enable-dma-fifo=1 --target=hw +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 4, 23, 320, +mt shape:8, 4, 23, 178, +mt stride:8, 4, 23, 142, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 4, 23, 320, }, + .tiling_dimension = {8, 3, 23, 89, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 142, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 4, 23, 320, }, + .tiling_dimension = {8, 3, 23, 89, }, + .offset = {0, 0, 0, 89, }, + .tiling_traversing ={ + {.dim = 3, .stride = 142, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {98256, }, + .tiling_dimension = {49128, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {98256, }, + .tiling_dimension = {49128, }, + .offset = {49128, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {98256, }, + .tiling_dimension = {49128, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {98256, }, + .tiling_dimension = {49128, }, + .offset = {49128, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 3, 23, 320, }, + .tiling_dimension = {8, 3, 23, 89, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 142, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 3, 23, 320, }, + .tiling_dimension = {8, 3, 23, 89, }, + .offset = {0, 0, 0, 89, }, + .tiling_traversing ={ + {.dim = 3, .stride = 142, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 2 + .wts_mem_rep = 0 + .ofm_mem_rep = 2 +[bufferpad] using by passing mode +[DEBUG] Found +[INFO] bufferunpad bypass kernel +[INFO] bufferunpad bypass kernel +[INFO] bufferunpad bypass kernel +search dm seg, num = 72 +dm_seg = 72 +Construct tile layer is done +search dm seg, num = 2880 +dm_seg = 2880 +Construct tile layer is done +search dm seg, num = 2880 +dm_seg = 2880 +Construct tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +g0 print + + + graph tiling: +ifm_mem_rep: 40 +ifm_mem_sz: 72 +ifm_dm_rep: 40 +ifm_dm_sz: 72 +ifm_dm_pad_sz: 96 + +MSG! read_ifm_ddr +buff dimension: +72 +tiling dimension: +72 +tile_traversal: +dim = 0, stride = 72, wrap = 1 +tiling rep = 40 + +MSG! write_ifm_mem +buff dimension: +72 +tiling dimension: +72 +tile_traversal: +dim = 0, stride = 72, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +72 +tiling dimension: +72 +tile_traversal: +dim = 0, stride = 72, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +320 9 +tiling dimension: +8 9 +tile_traversal: +dim = 1, stride = 9, wrap = 1 +dim = 0, stride = 8, wrap = 40 +tiling rep = 1 +g1 print + + + graph tiling: +ifm_mem_rep: 1 +ifm_mem_sz: 2880 +ifm_dm_rep: 1 +ifm_dm_sz: 2880 +ifm_dm_pad_sz: 2880 + +MSG! read_ifm_ddr +buff dimension: +2880 +tiling dimension: +2880 +tile_traversal: +dim = 0, stride = 2880, wrap = 1 +tiling rep = 1 + +MSG! write_ifm_mem +buff dimension: +2880 +tiling dimension: +2880 +tile_traversal: +dim = 0, stride = 2880, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +2880 +tiling dimension: +2880 +tile_traversal: +dim = 0, stride = 2880, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +2880 1 +tiling dimension: +2880 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 2880, wrap = 1 +tiling rep = 1 +g2 print + + + graph tiling: +ifm_mem_rep: 23 +ifm_mem_sz: 2880 +ifm_dm_rep: 23 +ifm_dm_sz: 2880 +ifm_dm_pad_sz: 2880 + +MSG! read_ifm_ddr +buff dimension: +2880 +tiling dimension: +2880 +tile_traversal: +dim = 0, stride = 2880, wrap = 1 +tiling rep = 23 + +MSG! write_ifm_mem +buff dimension: +2880 +tiling dimension: +2880 +tile_traversal: +dim = 0, stride = 2880, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +2880 +tiling dimension: +2880 +tile_traversal: +dim = 0, stride = 2880, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +66240 1 +tiling dimension: +2880 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 2880, wrap = 23 +tiling rep = 1 +search dm seg, num = 120 +dm_seg = 120 +Construct tile layer is done +search dm seg, num = 4800 +dm_seg = 4800 +Construct tile layer is done +search dm seg, num = 4800 +dm_seg = 4800 +Construct tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +g0 print + + + graph tiling: +ifm_mem_rep: 40 +ifm_mem_sz: 120 +ifm_dm_rep: 40 +ifm_dm_sz: 120 +ifm_dm_pad_sz: 128 + +MSG! read_ifm_ddr +buff dimension: +120 +tiling dimension: +120 +tile_traversal: +dim = 0, stride = 120, wrap = 1 +tiling rep = 40 + +MSG! write_ifm_mem +buff dimension: +120 +tiling dimension: +120 +tile_traversal: +dim = 0, stride = 120, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +120 +tiling dimension: +120 +tile_traversal: +dim = 0, stride = 120, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +320 15 +tiling dimension: +8 15 +tile_traversal: +dim = 1, stride = 15, wrap = 1 +dim = 0, stride = 8, wrap = 40 +tiling rep = 1 +g1 print + + + graph tiling: +ifm_mem_rep: 1 +ifm_mem_sz: 4800 +ifm_dm_rep: 1 +ifm_dm_sz: 4800 +ifm_dm_pad_sz: 4800 + +MSG! read_ifm_ddr +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 + +MSG! write_ifm_mem +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +4800 1 +tiling dimension: +4800 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 +g2 print + + + graph tiling: +ifm_mem_rep: 23 +ifm_mem_sz: 4800 +ifm_dm_rep: 23 +ifm_dm_sz: 4800 +ifm_dm_pad_sz: 4800 + +MSG! read_ifm_ddr +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 23 + +MSG! write_ifm_mem +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +110400 1 +tiling dimension: +4800 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 4800, wrap = 23 +tiling rep = 1 +search dm seg, num = 120 +dm_seg = 120 +Construct tile layer is done +search dm seg, num = 4800 +dm_seg = 4800 +Construct tile layer is done +search dm seg, num = 4800 +dm_seg = 4800 +Construct tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +g0 print + + + graph tiling: +ifm_mem_rep: 40 +ifm_mem_sz: 120 +ifm_dm_rep: 40 +ifm_dm_sz: 120 +ifm_dm_pad_sz: 128 + +MSG! read_ifm_ddr +buff dimension: +120 +tiling dimension: +120 +tile_traversal: +dim = 0, stride = 120, wrap = 1 +tiling rep = 40 + +MSG! write_ifm_mem +buff dimension: +120 +tiling dimension: +120 +tile_traversal: +dim = 0, stride = 120, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +120 +tiling dimension: +120 +tile_traversal: +dim = 0, stride = 120, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +320 15 +tiling dimension: +8 15 +tile_traversal: +dim = 1, stride = 15, wrap = 1 +dim = 0, stride = 8, wrap = 40 +tiling rep = 1 +g1 print + + + graph tiling: +ifm_mem_rep: 1 +ifm_mem_sz: 4800 +ifm_dm_rep: 1 +ifm_dm_sz: 4800 +ifm_dm_pad_sz: 4800 + +MSG! read_ifm_ddr +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 + +MSG! write_ifm_mem +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +4800 1 +tiling dimension: +4800 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 +g2 print + + + graph tiling: +ifm_mem_rep: 23 +ifm_mem_sz: 4800 +ifm_dm_rep: 23 +ifm_dm_sz: 4800 +ifm_dm_pad_sz: 4800 + +MSG! read_ifm_ddr +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 23 + +MSG! write_ifm_mem +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +4800 +tiling dimension: +4800 +tile_traversal: +dim = 0, stride = 4800, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +110400 1 +tiling dimension: +4800 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 4800, wrap = 23 +tiling rep = 1 +[INFO] bufferunpad bypass kernel +search dm seg, num = 480 +dm_seg = 480 +Construct tile layer is done +search dm seg, num = 9600 +dm_seg = 4800 +Construct tile layer is done +search dm seg, num = 9600 +dm_seg = 4800 +Construct tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +g0 print + + + graph tiling: +ifm_mem_rep: 20 +ifm_mem_sz: 480 +ifm_dm_rep: 20 +ifm_dm_sz: 480 +ifm_dm_pad_sz: 480 + +MSG! read_ifm_ddr +buff dimension: +480 +tiling dimension: +480 +tile_traversal: +dim = 0, stride = 480, wrap = 1 +tiling rep = 20 + +MSG! write_ifm_mem +buff dimension: +480 +tiling dimension: +480 +tile_traversal: +dim = 0, stride = 480, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +480 +tiling dimension: +480 +tile_traversal: +dim = 0, stride = 480, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +160 60 +tiling dimension: +8 60 +tile_traversal: +dim = 1, stride = 60, wrap = 1 +dim = 0, stride = 8, wrap = 20 +tiling rep = 1 +g1 print + + + graph tiling: +ifm_mem_rep: 1 +ifm_mem_sz: 9600 +ifm_dm_rep: 2 +ifm_dm_sz: 4800 +ifm_dm_pad_sz: 4800 + +MSG! read_ifm_ddr +buff dimension: +9600 +tiling dimension: +9600 +tile_traversal: +dim = 0, stride = 9600, wrap = 1 +tiling rep = 1 + +MSG! write_ifm_mem +buff dimension: +9600 +tiling dimension: +9600 +tile_traversal: +dim = 0, stride = 9600, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +9600 +tiling dimension: +9600 +tile_traversal: +dim = 0, stride = 9600, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +9600 1 +tiling dimension: +9600 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 9600, wrap = 1 +tiling rep = 1 +g2 print + + + graph tiling: +ifm_mem_rep: 12 +ifm_mem_sz: 9600 +ifm_dm_rep: 24 +ifm_dm_sz: 4800 +ifm_dm_pad_sz: 4800 + +MSG! read_ifm_ddr +buff dimension: +9600 +tiling dimension: +9600 +tile_traversal: +dim = 0, stride = 9600, wrap = 1 +tiling rep = 12 + +MSG! write_ifm_mem +buff dimension: +9600 +tiling dimension: +9600 +tile_traversal: +dim = 0, stride = 9600, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +9600 +tiling dimension: +9600 +tile_traversal: +dim = 0, stride = 9600, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +115200 1 +tiling dimension: +9600 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 9600, wrap = 12 +tiling rep = 1 +search dm seg, num = 672 +dm_seg = 672 +Construct tile layer is done +search dm seg, num = 13440 +dm_seg = 4480 +Construct tile layer is done +search dm seg, num = 13440 +dm_seg = 4480 +Construct tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +g0 print + + + graph tiling: +ifm_mem_rep: 20 +ifm_mem_sz: 672 +ifm_dm_rep: 20 +ifm_dm_sz: 672 +ifm_dm_pad_sz: 672 + +MSG! read_ifm_ddr +buff dimension: +672 +tiling dimension: +672 +tile_traversal: +dim = 0, stride = 672, wrap = 1 +tiling rep = 20 + +MSG! write_ifm_mem +buff dimension: +672 +tiling dimension: +672 +tile_traversal: +dim = 0, stride = 672, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +672 +tiling dimension: +672 +tile_traversal: +dim = 0, stride = 672, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +160 84 +tiling dimension: +8 84 +tile_traversal: +dim = 1, stride = 84, wrap = 1 +dim = 0, stride = 8, wrap = 20 +tiling rep = 1 +g1 print + + + graph tiling: +ifm_mem_rep: 1 +ifm_mem_sz: 13440 +ifm_dm_rep: 3 +ifm_dm_sz: 4480 +ifm_dm_pad_sz: 4480 + +MSG! read_ifm_ddr +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 + +MSG! write_ifm_mem +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +13440 1 +tiling dimension: +13440 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 +g2 print + + + graph tiling: +ifm_mem_rep: 12 +ifm_mem_sz: 13440 +ifm_dm_rep: 36 +ifm_dm_sz: 4480 +ifm_dm_pad_sz: 4480 + +MSG! read_ifm_ddr +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 12 + +MSG! write_ifm_mem +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +161280 1 +tiling dimension: +13440 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 13440, wrap = 12 +tiling rep = 1 +search dm seg, num = 672 +dm_seg = 672 +Construct tile layer is done +search dm seg, num = 13440 +dm_seg = 4480 +Construct tile layer is done +search dm seg, num = 13440 +dm_seg = 4480 +Construct tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +g0 print + + + graph tiling: +ifm_mem_rep: 20 +ifm_mem_sz: 672 +ifm_dm_rep: 20 +ifm_dm_sz: 672 +ifm_dm_pad_sz: 672 + +MSG! read_ifm_ddr +buff dimension: +672 +tiling dimension: +672 +tile_traversal: +dim = 0, stride = 672, wrap = 1 +tiling rep = 20 + +MSG! write_ifm_mem +buff dimension: +672 +tiling dimension: +672 +tile_traversal: +dim = 0, stride = 672, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +672 +tiling dimension: +672 +tile_traversal: +dim = 0, stride = 672, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +160 84 +tiling dimension: +8 84 +tile_traversal: +dim = 1, stride = 84, wrap = 1 +dim = 0, stride = 8, wrap = 20 +tiling rep = 1 +g1 print + + + graph tiling: +ifm_mem_rep: 1 +ifm_mem_sz: 13440 +ifm_dm_rep: 3 +ifm_dm_sz: 4480 +ifm_dm_pad_sz: 4480 + +MSG! read_ifm_ddr +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 + +MSG! write_ifm_mem +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +13440 1 +tiling dimension: +13440 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 +g2 print + + + graph tiling: +ifm_mem_rep: 12 +ifm_mem_sz: 13440 +ifm_dm_rep: 36 +ifm_dm_sz: 4480 +ifm_dm_pad_sz: 4480 + +MSG! read_ifm_ddr +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 12 + +MSG! write_ifm_mem +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +13440 +tiling dimension: +13440 +tile_traversal: +dim = 0, stride = 13440, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +161280 1 +tiling dimension: +13440 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 13440, wrap = 12 +tiling rep = 1 +search dm seg, num = 960 +dm_seg = 960 +Construct tile layer is done +search dm seg, num = 19200 +dm_seg = 4800 +Construct tile layer is done +search dm seg, num = 19200 +dm_seg = 4800 +Construct tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +g0 print + + + graph tiling: +ifm_mem_rep: 20 +ifm_mem_sz: 960 +ifm_dm_rep: 20 +ifm_dm_sz: 960 +ifm_dm_pad_sz: 960 + +MSG! read_ifm_ddr +buff dimension: +960 +tiling dimension: +960 +tile_traversal: +dim = 0, stride = 960, wrap = 1 +tiling rep = 20 + +MSG! write_ifm_mem +buff dimension: +960 +tiling dimension: +960 +tile_traversal: +dim = 0, stride = 960, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +960 +tiling dimension: +960 +tile_traversal: +dim = 0, stride = 960, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +160 120 +tiling dimension: +8 120 +tile_traversal: +dim = 1, stride = 120, wrap = 1 +dim = 0, stride = 8, wrap = 20 +tiling rep = 1 +g1 print + + + graph tiling: +ifm_mem_rep: 1 +ifm_mem_sz: 19200 +ifm_dm_rep: 4 +ifm_dm_sz: 4800 +ifm_dm_pad_sz: 4800 + +MSG! read_ifm_ddr +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 + +MSG! write_ifm_mem +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +19200 1 +tiling dimension: +19200 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 +g2 print + + + graph tiling: +ifm_mem_rep: 12 +ifm_mem_sz: 19200 +ifm_dm_rep: 48 +ifm_dm_sz: 4800 +ifm_dm_pad_sz: 4800 + +MSG! read_ifm_ddr +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 12 + +MSG! write_ifm_mem +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +230400 1 +tiling dimension: +19200 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 19200, wrap = 12 +tiling rep = 1 +search dm seg, num = 960 +dm_seg = 960 +Construct tile layer is done +search dm seg, num = 19200 +dm_seg = 4800 +Construct tile layer is done +search dm seg, num = 19200 +dm_seg = 4800 +Construct tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +g0 print + + + graph tiling: +ifm_mem_rep: 20 +ifm_mem_sz: 960 +ifm_dm_rep: 20 +ifm_dm_sz: 960 +ifm_dm_pad_sz: 960 + +MSG! read_ifm_ddr +buff dimension: +960 +tiling dimension: +960 +tile_traversal: +dim = 0, stride = 960, wrap = 1 +tiling rep = 20 + +MSG! write_ifm_mem +buff dimension: +960 +tiling dimension: +960 +tile_traversal: +dim = 0, stride = 960, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +960 +tiling dimension: +960 +tile_traversal: +dim = 0, stride = 960, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +160 120 +tiling dimension: +8 120 +tile_traversal: +dim = 1, stride = 120, wrap = 1 +dim = 0, stride = 8, wrap = 20 +tiling rep = 1 +g1 print + + + graph tiling: +ifm_mem_rep: 1 +ifm_mem_sz: 19200 +ifm_dm_rep: 4 +ifm_dm_sz: 4800 +ifm_dm_pad_sz: 4800 + +MSG! read_ifm_ddr +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 + +MSG! write_ifm_mem +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +19200 1 +tiling dimension: +19200 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 +g2 print + + + graph tiling: +ifm_mem_rep: 12 +ifm_mem_sz: 19200 +ifm_dm_rep: 48 +ifm_dm_sz: 4800 +ifm_dm_pad_sz: 4800 + +MSG! read_ifm_ddr +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 12 + +MSG! write_ifm_mem +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +19200 +tiling dimension: +19200 +tile_traversal: +dim = 0, stride = 19200, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +230400 1 +tiling dimension: +19200 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 19200, wrap = 12 +tiling rep = 1 +search dm seg, num = 128 +dm_seg = 128 +Construct tile layer is done +search dm seg, num = 2560 +dm_seg = 2560 +Construct tile layer is done +search dm seg, num = 2560 +dm_seg = 2560 +Construct tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +Connect L3 for tile layer is done +g0 print + + + graph tiling: +ifm_mem_rep: 20 +ifm_mem_sz: 128 +ifm_dm_rep: 20 +ifm_dm_sz: 128 +ifm_dm_pad_sz: 128 + +MSG! read_ifm_ddr +buff dimension: +128 +tiling dimension: +128 +tile_traversal: +dim = 0, stride = 128, wrap = 1 +tiling rep = 20 + +MSG! write_ifm_mem +buff dimension: +128 +tiling dimension: +128 +tile_traversal: +dim = 0, stride = 128, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +128 +tiling dimension: +128 +tile_traversal: +dim = 0, stride = 128, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +160 16 +tiling dimension: +8 16 +tile_traversal: +dim = 1, stride = 16, wrap = 1 +dim = 0, stride = 8, wrap = 20 +tiling rep = 1 +g1 print + + + graph tiling: +ifm_mem_rep: 1 +ifm_mem_sz: 2560 +ifm_dm_rep: 1 +ifm_dm_sz: 2560 +ifm_dm_pad_sz: 2560 + +MSG! read_ifm_ddr +buff dimension: +2560 +tiling dimension: +2560 +tile_traversal: +dim = 0, stride = 2560, wrap = 1 +tiling rep = 1 + +MSG! write_ifm_mem +buff dimension: +2560 +tiling dimension: +2560 +tile_traversal: +dim = 0, stride = 2560, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +2560 +tiling dimension: +2560 +tile_traversal: +dim = 0, stride = 2560, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +2560 1 +tiling dimension: +2560 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 2560, wrap = 1 +tiling rep = 1 +g2 print + + + graph tiling: +ifm_mem_rep: 12 +ifm_mem_sz: 2560 +ifm_dm_rep: 12 +ifm_dm_sz: 2560 +ifm_dm_pad_sz: 2560 + +MSG! read_ifm_ddr +buff dimension: +2560 +tiling dimension: +2560 +tile_traversal: +dim = 0, stride = 2560, wrap = 1 +tiling rep = 12 + +MSG! write_ifm_mem +buff dimension: +2560 +tiling dimension: +2560 +tile_traversal: +dim = 0, stride = 2560, wrap = 1 +tiling rep = 1 + +MSG! read_ofm_mem +buff dimension: +2560 +tiling dimension: +2560 +tile_traversal: +dim = 0, stride = 2560, wrap = 1 +tiling rep = 1 + +MSG! write_ofm_ddr +buff dimension: +30720 1 +tiling dimension: +2560 1 +tile_traversal: +dim = 1, stride = 1, wrap = 1 +dim = 0, stride = 2560, wrap = 12 +tiling rep = 1 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 20, 16, 12, +mt shape:8, 20, 16, 12, +mt stride:8, 20, 16, 12, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 6, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {7680, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {7680, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 20, 8, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 8, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 6, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 1 + .wts_mem_rep = 0 + .ofm_mem_rep = 1 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 20, 16, 12, +mt shape:8, 20, 16, 12, +mt stride:8, 20, 16, 12, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 8, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 8, 6, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {7680, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {7680, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 20, 8, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 8, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 6, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 1 + .wts_mem_rep = 0 + .ofm_mem_rep = 1 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 20, 16, 12, +mt shape:8, 20, 16, 12, +mt stride:8, 20, 16, 12, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 8, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 8, 6, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {7680, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {7680, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 20, 8, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 8, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 6, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 1 + .wts_mem_rep = 0 + .ofm_mem_rep = 1 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 20, 16, 12, +mt shape:8, 20, 16, 12, +mt stride:8, 20, 16, 12, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 6, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {7680, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {15360, }, + .tiling_dimension = {7680, }, + .offset = {7680, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 20, 8, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 8, 12, }, + .tiling_dimension = {8, 20, 8, 6, }, + .offset = {0, 0, 0, 6, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 1 + .wts_mem_rep = 0 + .ofm_mem_rep = 1 +using opt mode: 0 +TG-WARNING: write_mk_in0's access pattern is underspecified, aiecompiler will use a default setting which might change with time +TG-WARNING: read_mk_out0's access pattern is underspecified, aiecompiler will use a default setting which might change with time +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {6528, }, + .mk_in1_dim = {}, + .mk_out0_dim = {8192, }, + .mk_in0_size = 6528 + .mk_in1_size = 0 + .mk_out0_size = 8192 + .mk_in0_addr = {16384, }, + .mk_in1_addr = {}, + .mk_out0_addr = {0, }, + .lp_param = {12, 20, 1056964608, 1056964608, 3196059648, 3196059648, 4, 17563928, 19398913, 16843028, 17563936, 35651841, 16909073, 0, 0, 0, }, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + {30720, }, + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + {262144, }, + }, + .ifm_mem_size = {30720, }, + .wts_mem_size = {}, + .ofm_mem_size = {262144, }, + .read_ifm_mem = { + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 17, 4, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 3, .wrap = 2}, + {.dim = 3, .stride = 1, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 17, 4, 12, }, + .offset = {0, 0, 4, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 3, .wrap = 2}, + {.dim = 3, .stride = 1, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 17, 4, 12, }, + .offset = {0, 0, 8, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 3, .wrap = 2}, + {.dim = 3, .stride = 1, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 17, 4, 12, }, + .offset = {0, 0, 12, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 3, .wrap = 2}, + {.dim = 3, .stride = 1, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + { + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 1, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 3, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 4, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 5, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 6, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 7, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 8, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 9, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 10, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 11, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 12, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 13, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 14, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 15, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 1}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .mk_rep = 2 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 20, 16, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 20, 16, 12, }, + .tiling_dimension = {8, 20, 16, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {30720, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {30720, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 40, 16, 24, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 64, 16, 32, }, + .tiling_dimension = {8, 40, 16, 24, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 40, 16, 24, }, + .tiling_dimension = {8, 40, 16, 24, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 16, 24, }, + .tiling_dimension = {8, 40, 16, 24, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 1 + .wts_mem_rep = 0 + .ofm_mem_rep = 1 +connect_ifm_ddr +connect_ifm_ddr +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 40, 16, 24, +mt shape:8, 40, 16, 24, +mt stride:8, 40, 16, 24, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 40, 16, 24, }, + .tiling_dimension = {8, 40, 16, 11, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 16, 24, }, + .tiling_dimension = {8, 40, 16, 12, }, + .offset = {0, 0, 0, 11, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {117760, }, + .tiling_dimension = {56320, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {117760, }, + .tiling_dimension = {61440, }, + .offset = {56320, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {117760, }, + .tiling_dimension = {56320, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {117760, }, + .tiling_dimension = {61440, }, + .offset = {56320, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 40, 16, 23, }, + .tiling_dimension = {8, 40, 16, 11, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 16, 23, }, + .tiling_dimension = {8, 40, 16, 12, }, + .offset = {0, 0, 0, 11, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 1 + .wts_mem_rep = 0 + .ofm_mem_rep = 1 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 40, 10, 23, +mt shape:8, 40, 10, 23, +mt stride:8, 40, 10, 23, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 40, 5, 11, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 40, 5, 12, }, + .offset = {0, 0, 0, 11, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {17600, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {19200, }, + .offset = {17600, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {17600, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {19200, }, + .offset = {17600, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 40, 5, 23, }, + .tiling_dimension = {8, 40, 5, 11, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 5, 23, }, + .tiling_dimension = {8, 40, 5, 12, }, + .offset = {0, 0, 0, 11, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 1 + .wts_mem_rep = 0 + .ofm_mem_rep = 1 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 40, 10, 23, +mt shape:8, 40, 10, 23, +mt stride:8, 40, 10, 23, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 40, 5, 11, }, + .offset = {0, 0, 5, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 40, 5, 12, }, + .offset = {0, 0, 5, 11, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {17600, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {19200, }, + .offset = {17600, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {17600, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {19200, }, + .offset = {17600, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 40, 5, 23, }, + .tiling_dimension = {8, 40, 5, 11, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 5, 23, }, + .tiling_dimension = {8, 40, 5, 12, }, + .offset = {0, 0, 0, 11, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 1 + .wts_mem_rep = 0 + .ofm_mem_rep = 1 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 40, 10, 23, +mt shape:8, 40, 10, 23, +mt stride:8, 40, 10, 23, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 40, 5, 11, }, + .offset = {0, 0, 5, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 40, 5, 12, }, + .offset = {0, 0, 5, 11, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {17600, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {19200, }, + .offset = {17600, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {17600, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {19200, }, + .offset = {17600, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 40, 5, 23, }, + .tiling_dimension = {8, 40, 5, 11, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 5, 23, }, + .tiling_dimension = {8, 40, 5, 12, }, + .offset = {0, 0, 0, 11, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 1 + .wts_mem_rep = 0 + .ofm_mem_rep = 1 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 40, 10, 23, +mt shape:8, 40, 10, 23, +mt stride:8, 40, 10, 23, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 40, 5, 11, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 40, 5, 12, }, + .offset = {0, 0, 0, 11, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {17600, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {19200, }, + .offset = {17600, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {17600, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {36800, }, + .tiling_dimension = {19200, }, + .offset = {17600, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 40, 5, 23, }, + .tiling_dimension = {8, 40, 5, 11, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 5, 23, }, + .tiling_dimension = {8, 40, 5, 12, }, + .offset = {0, 0, 0, 11, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 1 + .wts_mem_rep = 0 + .ofm_mem_rep = 1 +using opt mode: 0 +TG-WARNING: write_mk_in0's access pattern is underspecified, aiecompiler will use a default setting which might change with time +TG-WARNING: read_mk_out0's access pattern is underspecified, aiecompiler will use a default setting which might change with time +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {9792, }, + .mk_in1_dim = {}, + .mk_out0_dim = {8192, }, + .mk_in0_size = 9792 + .mk_in1_size = 0 + .mk_out0_size = 8192 + .mk_in0_addr = {16384, }, + .mk_in1_addr = {}, + .mk_out0_addr = {0, }, + .lp_param = {23, 40, 1056964608, 1056964608, 3196059648, 3196059648, 4, 18284846, 36176129, 16913173, 101777952, 35651842, 16912146, 0, 0, 0, }, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + {61824, }, + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + {524288, }, + }, + .ifm_mem_size = {61824, }, + .wts_mem_size = {}, + .ofm_mem_size = {524288, }, + .read_ifm_mem = { + { + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 18, 4, 17, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 15, .wrap = 2}, + {.dim = 3, .stride = 6, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 18, 4, 17, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 15, .wrap = 2}, + {.dim = 3, .stride = 6, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + } + { + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 18, 4, 17, }, + .offset = {0, 0, 4, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 15, .wrap = 2}, + {.dim = 3, .stride = 6, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 18, 4, 17, }, + .offset = {0, 0, 4, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 15, .wrap = 2}, + {.dim = 3, .stride = 6, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + } + { + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 18, 4, 17, }, + .offset = {0, 0, 8, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 15, .wrap = 2}, + {.dim = 3, .stride = 6, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 18, 4, 17, }, + .offset = {0, 0, 8, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 15, .wrap = 2}, + {.dim = 3, .stride = 6, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + } + { + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 18, 4, 17, }, + .offset = {0, 0, 12, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 15, .wrap = 2}, + {.dim = 3, .stride = 6, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 18, 4, 17, }, + .offset = {0, 0, 12, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 15, .wrap = 2}, + {.dim = 3, .stride = 6, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + } + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + { + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 1, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 1, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 3, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 3, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + } + { + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 4, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 4, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 5, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 5, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 6, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 6, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 7, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 7, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + } + { + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 8, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 8, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 9, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 9, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 10, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 10, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 11, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 11, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + } + { + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 12, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 12, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 13, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 13, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 14, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 14, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 15, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 32, 1, 32, }, + .offset = {0, 0, 15, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 32, .wrap = 2}, + {.dim = 3, .stride = 32, .wrap = 2}, + } + .repetition = 1 + .buffer_iteration = 1 + } + } + }, + .mk_rep = 8 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 21, 10, 23, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 21, 10, 23, }, + .offset = {0, 19, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 21, 10, 23, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 40, 10, 23, }, + .tiling_dimension = {8, 21, 10, 23, }, + .offset = {0, 19, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 21, 10, 23, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 21, 10, 23, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + } + { + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 21, 10, 23, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 21, 16, 23, }, + .tiling_dimension = {8, 21, 10, 23, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 40, 10, 46, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 40, 10, 46, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + } + { + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 40, 10, 46, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 64, 16, 64, }, + .tiling_dimension = {8, 40, 10, 46, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 80, 10, 46, }, + .tiling_dimension = {8, 40, 10, 46, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 10, 46, }, + .tiling_dimension = {8, 40, 10, 46, }, + .offset = {0, 40, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 10, 46, }, + .tiling_dimension = {8, 40, 10, 46, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 10, 46, }, + .tiling_dimension = {8, 40, 10, 46, }, + .offset = {0, 40, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 2 + .wts_mem_rep = 0 + .ofm_mem_rep = 2 +connect_ifm_ddr +connect_ifm_ddr +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 80, 10, 46, +mt shape:8, 80, 4, 46, +mt stride:8, 80, 3, 46, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 80, 10, 46, }, + .tiling_dimension = {8, 80, 4, 22, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 2, .stride = 3, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 10, 46, }, + .tiling_dimension = {8, 80, 4, 23, }, + .offset = {0, 0, 0, 22, }, + .tiling_traversing ={ + {.dim = 2, .stride = 3, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {115200, }, + .tiling_dimension = {56320, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {115200, }, + .tiling_dimension = {58880, }, + .offset = {56320, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {115200, }, + .tiling_dimension = {56320, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {115200, }, + .tiling_dimension = {58880, }, + .offset = {56320, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 80, 10, 45, }, + .tiling_dimension = {8, 80, 4, 22, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 2, .stride = 3, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 10, 45, }, + .tiling_dimension = {8, 80, 4, 23, }, + .offset = {0, 0, 0, 22, }, + .tiling_traversing ={ + {.dim = 2, .stride = 3, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 3 + .wts_mem_rep = 0 + .ofm_mem_rep = 3 +debug slice hcwc8---- +input ddr shape:8, 80, 5, 45, +mt shape:8, 80, 5, 15, +mt stride:8, 80, 5, 15, +mt padded shape:8, 80, 5, 15, +dm shape:8, 80, 5, 3, +dm stride:8, 80, 5, 3, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 9600 + .mk_in1_size = 0 + .mk_out0_size = 5760 + .mk_in0_addr = {0, }, + .mk_in1_addr = {}, + .mk_out0_addr = {24576, }, + .lp_param = {40, 3, 80, 0, 20, 0, 9600, }, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + {65536, }, + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + {262144, }, + }, + .ifm_mem_size = {65536, }, + .wts_mem_size = {}, + .ofm_mem_size = {262144, }, + .read_ifm_mem = { + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .mk_rep = 15 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 80, 5, 7, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 80, 5, 8, }, + .offset = {0, 0, 0, 7, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {48000, }, + .tiling_dimension = {22400, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {48000, }, + .tiling_dimension = {25600, }, + .offset = {22400, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {28800, }, + .tiling_dimension = {13440, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {28800, }, + .tiling_dimension = {15360, }, + .offset = {13440, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 80, 3, 45, }, + .tiling_dimension = {8, 80, 3, 7, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 45, }, + .tiling_dimension = {8, 80, 3, 8, }, + .offset = {0, 0, 0, 7, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 3 + .wts_mem_rep = 0 + .ofm_mem_rep = 3 +debug slice hcwc8---- +input ddr shape:8, 80, 5, 45, +mt shape:8, 80, 5, 15, +mt stride:8, 80, 5, 15, +mt padded shape:8, 80, 5, 15, +dm shape:8, 80, 5, 3, +dm stride:8, 80, 5, 3, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 9600 + .mk_in1_size = 0 + .mk_out0_size = 5760 + .mk_in0_addr = {0, }, + .mk_in1_addr = {}, + .mk_out0_addr = {24576, }, + .lp_param = {40, 3, 80, 20, 40, 0, 9600, }, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + {65536, }, + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + {262144, }, + }, + .ifm_mem_size = {65536, }, + .wts_mem_size = {}, + .ofm_mem_size = {262144, }, + .read_ifm_mem = { + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .mk_rep = 15 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 80, 5, 7, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 80, 5, 8, }, + .offset = {0, 0, 0, 7, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {48000, }, + .tiling_dimension = {22400, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {48000, }, + .tiling_dimension = {25600, }, + .offset = {22400, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {28800, }, + .tiling_dimension = {13440, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {28800, }, + .tiling_dimension = {15360, }, + .offset = {13440, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 80, 3, 45, }, + .tiling_dimension = {8, 80, 3, 7, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 45, }, + .tiling_dimension = {8, 80, 3, 8, }, + .offset = {0, 0, 0, 7, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 3 + .wts_mem_rep = 0 + .ofm_mem_rep = 3 +[INFO] concatc8: enable kernel. +debug slice hcwc8---- +input ddr shape:8, 80, 5, 45, +mt shape:8, 80, 5, 15, +mt stride:8, 80, 5, 15, +mt padded shape:8, 80, 5, 15, +dm shape:8, 80, 5, 3, +dm stride:8, 80, 5, 3, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 9600 + .mk_in1_size = 0 + .mk_out0_size = 5760 + .mk_in0_addr = {0, }, + .mk_in1_addr = {}, + .mk_out0_addr = {24576, }, + .lp_param = {40, 3, 80, 20, 40, 0, 9600, }, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + {65536, }, + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + {262144, }, + }, + .ifm_mem_size = {65536, }, + .wts_mem_size = {}, + .ofm_mem_size = {262144, }, + .read_ifm_mem = { + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .mk_rep = 15 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 80, 5, 7, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 80, 5, 8, }, + .offset = {0, 0, 0, 7, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {48000, }, + .tiling_dimension = {22400, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {48000, }, + .tiling_dimension = {25600, }, + .offset = {22400, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {28800, }, + .tiling_dimension = {13440, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {28800, }, + .tiling_dimension = {15360, }, + .offset = {13440, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 80, 3, 45, }, + .tiling_dimension = {8, 80, 3, 7, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 45, }, + .tiling_dimension = {8, 80, 3, 8, }, + .offset = {0, 0, 0, 7, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 3 + .wts_mem_rep = 0 + .ofm_mem_rep = 3 +debug slice hcwc8---- +input ddr shape:8, 80, 5, 45, +mt shape:8, 80, 5, 15, +mt stride:8, 80, 5, 15, +mt padded shape:8, 80, 5, 15, +dm shape:8, 80, 5, 3, +dm stride:8, 80, 5, 3, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 9600 + .mk_in1_size = 0 + .mk_out0_size = 5760 + .mk_in0_addr = {0, }, + .mk_in1_addr = {}, + .mk_out0_addr = {24576, }, + .lp_param = {40, 3, 80, 0, 20, 0, 9600, }, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + {65536, }, + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + {262144, }, + }, + .ifm_mem_size = {65536, }, + .wts_mem_size = {}, + .ofm_mem_size = {262144, }, + .read_ifm_mem = { + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 15, }, + .tiling_dimension = {8, 80, 5, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 15, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 30, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 3, 60, }, + .tiling_dimension = {8, 80, 3, 3, }, + .offset = {0, 0, 0, 45, }, + .tiling_traversing ={ + {.dim = 3, .stride = 3, .wrap = 5}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .mk_rep = 15 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 80, 5, 7, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 80, 5, 8, }, + .offset = {0, 0, 0, 7, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {48000, }, + .tiling_dimension = {22400, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {48000, }, + .tiling_dimension = {25600, }, + .offset = {22400, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {28800, }, + .tiling_dimension = {13440, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {28800, }, + .tiling_dimension = {15360, }, + .offset = {13440, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 80, 3, 45, }, + .tiling_dimension = {8, 80, 3, 7, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 3, 45, }, + .tiling_dimension = {8, 80, 3, 8, }, + .offset = {0, 0, 0, 7, }, + .tiling_traversing ={ + {.dim = 3, .stride = 15, .wrap = 3}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 3 + .wts_mem_rep = 0 + .ofm_mem_rep = 3 +[INFO] concatc8: enable kernel. +[INFO] concatc8: enable kernel. +using opt mode: 2 +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + {8, 80, 5, 45, }, + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + {8, 160, 8, 90, }, + }, + .ifm_ddr_size = {144000, }, + .wts_ddr_size = {}, + .ofm_ddr_size = {921600, }, + .mk_in0_dim = {8, 10, 1, 34, }, + .mk_in1_dim = {}, + .mk_out0_dim = {8, 18, 1, 18, }, + .mk_in0_size = 2720 + .mk_in1_size = 0 + .mk_out0_size = 2592 + .mk_in0_addr = {20480, }, + .mk_in1_addr = {}, + .mk_out0_addr = {28672, }, + .lp_param = {32, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, }, + .num_ifm = 2 + .num_wts = 1 + .num_ofm = 2 + .ifm_mem_dim = { + {8, 8, 4, 32, }, + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + {8, 16, 4, 64, }, + }, + .ifm_mem_size = {8192, }, + .wts_mem_size = {}, + .ofm_mem_size = {32768, }, + .read_ifm_mem = { + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 1, 32, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 1, 32, }, + .offset = {0, 0, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 1, 32, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 1, 32, }, + .offset = {0, 0, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 0, 32, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 0, 48, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 1, 32, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 1, 48, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 2, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 2, 32, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 2, 48, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 3, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 3, 32, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 3, 48, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .mk_rep = 48 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 72, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 72, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 72, 0, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 72, 1, 13, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 72, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 72, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 72, 0, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 80, 5, 45, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 72, 1, 29, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 2 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 3 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 4 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 5 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 6 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 7 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 8 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 9 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 10 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 11 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 12 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 13 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 14 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 15 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 16 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 17 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 18 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 19 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 20 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 21 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 22 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 23 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 24 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 25 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 26 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 27 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 28 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 29 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 30 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 31 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 32 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 33 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 34 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 35 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 36 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 37 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 38 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 39 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 40 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 41 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 42 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 43 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 44 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 45 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 46 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 47 + } + } + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 2 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 3 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 4 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 5 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 6 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 7 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 8 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 9 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 10 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 11 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 12 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 13 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 14 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 15 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 16 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 17 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 18 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 19 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 20 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 21 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 22 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 23 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 24 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 25 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 26 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 27 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 28 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 29 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 30 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 31 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 32 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 33 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 34 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 35 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 36 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 37 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 38 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 39 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 40 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 41 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 42 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 43 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 44 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 45 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 46 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 47 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 2 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 3 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 4 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 5 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 6 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 7 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 8 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 9 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 10 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 11 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 12 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 13 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 14 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 15 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 16 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 17 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 18 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 19 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 20 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 21 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 22 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 23 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 24 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 25 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 26 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 27 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 28 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 29 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 30 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 31 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 32 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 33 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 34 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 35 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 36 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 37 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 38 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 39 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 40 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 41 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 42 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 43 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 44 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 45 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 46 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 47 + } + } + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 2 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 3 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 4 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 5 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 6 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 7 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 8 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 9 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 10 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 11 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 12 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 13 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 14 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 15 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 16 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 17 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 18 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 19 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 20 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 21 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 22 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 23 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 24 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 25 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 26 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 27 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 28 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 29 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 30 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 31 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 32 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 33 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 34 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 35 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 36 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 37 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 38 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 39 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 40 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 41 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 42 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 43 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 44 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 45 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 46 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 47 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 145, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 145, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 145, 0, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 145, 1, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 145, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 145, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 145, 2, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 8, 90, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 145, 3, 27, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 48 + .wts_mem_rep = 0 + .ofm_mem_rep = 48 +connect_ifm_ddr +connect_ifm_ddr +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 160, 4, 90, +mt shape:8, 160, 4, 24, +mt stride:8, 160, 4, 22, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 12, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 160, 2, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 2, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 12, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 4 + .wts_mem_rep = 0 + .ofm_mem_rep = 4 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 160, 4, 90, +mt shape:8, 160, 4, 24, +mt stride:8, 160, 4, 22, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 2, 12, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 160, 2, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 2, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 12, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 4 + .wts_mem_rep = 0 + .ofm_mem_rep = 4 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 160, 4, 90, +mt shape:8, 160, 4, 24, +mt stride:8, 160, 4, 22, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 2, 12, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 160, 2, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 2, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 12, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 4 + .wts_mem_rep = 0 + .ofm_mem_rep = 4 +debug slice hcwc8---- +SliceHCWC8 in bypassing kernel mode! +input ddr shape:8, 160, 4, 90, +mt shape:8, 160, 4, 24, +mt stride:8, 160, 4, 22, +mt padded shape:0, 0, 0, 0, +dm shape:0, 0, 0, 0, +dm stride:0, 0, 0, 0, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 0 + .mk_in1_size = 0 + .mk_out0_size = 0 + .mk_in0_addr = {}, + .mk_in1_addr = {}, + .mk_out0_addr = {}, + .lp_param = {}, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + }, + .ifm_mem_size = {}, + .wts_mem_size = {}, + .ofm_mem_size = {}, + .read_ifm_mem = { + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + }, + .mk_rep = 1 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 12, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 160, 2, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 2, 90, }, + .tiling_dimension = {8, 160, 2, 12, }, + .offset = {0, 0, 0, 12, }, + .tiling_traversing ={ + {.dim = 3, .stride = 22, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 4 + .wts_mem_rep = 0 + .ofm_mem_rep = 4 +using opt mode: 2 +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + {8, 160, 4, 90, }, + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + {8, 320, 5, 180, }, + }, + .ifm_ddr_size = {460800, }, + .wts_ddr_size = {}, + .ofm_ddr_size = {2304000, }, + .mk_in0_dim = {8, 10, 1, 34, }, + .mk_in1_dim = {}, + .mk_out0_dim = {8, 18, 1, 18, }, + .mk_in0_size = 2720 + .mk_in1_size = 0 + .mk_out0_size = 2592 + .mk_in0_addr = {20480, }, + .mk_in1_addr = {}, + .mk_out0_addr = {28672, }, + .lp_param = {32, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, }, + .num_ifm = 2 + .num_wts = 1 + .num_ofm = 2 + .ifm_mem_dim = { + {8, 8, 4, 32, }, + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + {8, 16, 4, 64, }, + }, + .ifm_mem_size = {8192, }, + .wts_mem_size = {}, + .ofm_mem_size = {32768, }, + .read_ifm_mem = { + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 1, 32, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 1, 32, }, + .offset = {0, 0, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 1, 32, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 1, 32, }, + .offset = {0, 0, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 0, 32, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 0, 48, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 1, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 1, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 1, 32, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 1, 48, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 2, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 2, 32, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 2, 48, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 3, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 3, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 3, 32, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 16, 1, 16, }, + .offset = {0, 0, 3, 48, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .mk_rep = 69 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 77, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 84, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 91, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 98, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 105, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 112, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 119, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 126, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 133, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 140, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 147, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 152, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 77, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 84, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 91, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 98, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 105, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 112, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 119, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 126, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 133, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 140, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 147, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 152, 0, 31, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 77, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 84, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 91, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 98, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 105, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 112, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 119, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 126, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 133, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 140, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 147, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 152, 0, 58, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 77, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 84, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 91, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 98, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 105, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 112, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 119, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 126, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 133, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 140, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 147, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 152, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 77, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 84, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 91, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 98, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 105, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 112, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 119, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 126, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 133, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 140, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 147, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 152, 0, 47, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 7, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 14, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 21, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 28, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 35, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 42, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 49, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 56, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 63, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 70, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 77, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 84, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 91, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 98, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 105, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 112, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 119, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 126, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 133, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 140, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 147, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 4, 90, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 152, 0, 74, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 2 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 3 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 4 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 5 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 6 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 7 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 8 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 9 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 10 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 11 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 12 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 13 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 14 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 15 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 16 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 17 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 18 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 19 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 20 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 21 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 22 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 23 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 24 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 25 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 26 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 27 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 28 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 29 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 30 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 31 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 32 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 33 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 34 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 35 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 36 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 37 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 38 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 39 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 40 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 41 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 42 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 43 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 44 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 45 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 46 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 47 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 48 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 49 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 50 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 51 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 52 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 53 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 54 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 55 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 56 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 57 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 58 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 59 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 60 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 61 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 62 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 63 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 64 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 65 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 66 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 67 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 68 + } + } + { + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 2 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 3 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 4 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 5 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 6 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 7 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 8 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 9 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 10 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 11 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 12 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 13 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 14 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 15 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 16 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 17 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 18 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 19 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 20 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 21 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 22 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 23 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 24 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 25 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 26 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 27 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 28 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 29 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 30 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 31 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 32 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 33 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 34 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 35 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 36 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 37 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 38 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 39 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 40 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 41 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 42 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 43 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 44 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 45 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 46 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 47 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 48 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 49 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 50 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 51 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 52 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 53 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 54 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 55 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 56 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 57 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 58 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 59 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 60 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 61 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 62 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 63 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 64 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 65 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 66 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 67 + } + { + .buffer_dimension = {8, 8, 4, 32, }, + .tiling_dimension = {8, 8, 4, 16, }, + .offset = {0, 0, 0, 16, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 68 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 2 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 3 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 4 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 5 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 6 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 7 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 8 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 9 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 10 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 11 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 12 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 13 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 14 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 15 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 16 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 17 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 18 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 19 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 20 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 21 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 22 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 62, }, + .offset = {0, 0, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 23 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 24 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 25 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 26 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 27 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 28 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 29 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 30 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 31 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 32 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 33 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 34 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 35 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 36 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 37 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 38 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 39 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 40 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 41 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 42 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 43 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 44 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 62, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 45 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 46 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 47 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 48 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 49 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 50 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 51 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 52 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 53 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 54 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 55 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 56 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 57 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 58 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 59 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 60 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 61 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 62 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 63 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 64 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 65 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 66 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 67 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 0, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 68 + } + } + { + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 1 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 2 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 3 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 4 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 5 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 6 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 7 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 8 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 9 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 10 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 11 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 12 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 13 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 14 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 15 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 16 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 17 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 18 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 19 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 20 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 21 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 22 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 62, }, + .offset = {0, 0, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 23 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 24 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 25 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 26 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 27 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 28 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 29 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 30 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 31 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 32 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 33 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 34 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 35 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 36 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 37 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 38 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 39 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 40 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 41 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 42 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 43 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 44 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 62, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 45 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 46 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 47 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 48 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 49 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 50 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 51 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 52 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 53 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 54 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 55 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 56 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 57 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 58 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 59 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 60 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 61 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 62 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 63 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 64 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 65 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 66 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 67 + } + { + .buffer_dimension = {8, 16, 4, 64, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 1, 2, 1, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 68 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 155, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 169, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 183, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 197, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 211, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 225, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 239, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 253, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 267, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 281, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 295, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 305, 0, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 62, }, + .offset = {0, 0, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 15, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 29, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 43, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 57, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 71, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 85, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 99, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 113, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 127, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 141, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 155, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 169, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 183, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 197, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 211, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 225, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 239, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 253, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 267, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 281, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 295, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 62, }, + .offset = {0, 305, 0, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 155, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 169, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 183, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 197, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 211, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 225, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 239, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 253, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 267, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 281, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 295, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 305, 0, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 155, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 169, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 183, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 197, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 211, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 225, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 239, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 253, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 267, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 281, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 295, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 305, 2, 0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 62, }, + .offset = {0, 0, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 15, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 29, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 43, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 57, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 71, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 85, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 99, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 113, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 127, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 141, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 155, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 169, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 183, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 197, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 211, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 225, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 239, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 253, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 267, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 281, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 62, }, + .offset = {0, 295, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 62, }, + .offset = {0, 305, 2, 63, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 0, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 15, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 29, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 43, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 57, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 71, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 85, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 99, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 113, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 127, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 141, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 155, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 169, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 183, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 197, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 211, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 225, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 239, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 253, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 267, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 281, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 14, 2, 63, }, + .offset = {0, 295, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 320, 5, 180, }, + .tiling_dimension = {8, 15, 2, 63, }, + .offset = {0, 305, 2, 117, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 69 + .wts_mem_rep = 0 + .ofm_mem_rep = 69 +connect_ifm_ddr +connect_ifm_ddr +debug slice hcwc8---- +input ddr shape:8, 320, 1, 180, +mt shape:8, 160, 1, 48, +mt stride:8, 160, 1, 44, +mt padded shape:8, 160, 1, 48, +dm shape:8, 160, 1, 12, +dm stride:8, 160, 1, 12, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 3840 + .mk_in1_size = 0 + .mk_out0_size = 960 + .mk_in0_addr = {0, 8192, }, + .mk_in1_addr = {}, + .mk_out0_addr = {16384, 24576, }, + .lp_param = {8, 3, 40, 3, 4, 1, 3840, }, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + {65536, }, + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + {65536, }, + }, + .ifm_mem_size = {65536, }, + .wts_mem_size = {}, + .ofm_mem_size = {65536, }, + .read_ifm_mem = { + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 12, }, + .offset = {0, 40, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 12, }, + .offset = {0, 80, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 12, }, + .offset = {0, 120, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 0, 0, 3, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 0, 0, 6, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 0, 0, 9, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 40, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 40, 0, 3, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 40, 0, 6, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 40, 0, 9, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 80, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 80, 0, 3, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 80, 0, 6, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 80, 0, 9, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 120, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 120, 0, 3, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 120, 0, 6, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 120, 0, 9, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .mk_rep = 32 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 320, 1, 180, }, + .tiling_dimension = {8, 160, 1, 24, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 2}, + {.dim = 3, .stride = 44, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 320, 1, 180, }, + .tiling_dimension = {8, 160, 1, 24, }, + .offset = {0, 0, 0, 24, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 2}, + {.dim = 3, .stride = 44, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 320, 1, 180, }, + .tiling_dimension = {8, 160, 1, 24, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 2}, + {.dim = 3, .stride = 44, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 320, 1, 180, }, + .tiling_dimension = {8, 160, 1, 24, }, + .offset = {0, 0, 0, 24, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 2}, + {.dim = 3, .stride = 44, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 8 + .wts_mem_rep = 0 + .ofm_mem_rep = 8 +debug slice hcwc8---- +input ddr shape:8, 320, 1, 180, +mt shape:8, 160, 1, 48, +mt stride:8, 160, 1, 44, +mt padded shape:8, 160, 1, 48, +dm shape:8, 160, 1, 12, +dm stride:8, 160, 1, 12, +tiling_paramter_t = { + .enable_kernel = 1 + .enable_wts = 0 + .ifm_ddr_dim = { + }, + .wts_ddr_dim = { + }, + .ofm_ddr_dim = { + }, + .ifm_ddr_size = {}, + .wts_ddr_size = {}, + .ofm_ddr_size = {}, + .mk_in0_dim = {}, + .mk_in1_dim = {}, + .mk_out0_dim = {}, + .mk_in0_size = 3840 + .mk_in1_size = 0 + .mk_out0_size = 960 + .mk_in0_addr = {0, 8192, }, + .mk_in1_addr = {}, + .mk_out0_addr = {16384, 24576, }, + .lp_param = {8, 3, 40, 0, 3, 1, 3840, }, + .num_ifm = 1 + .num_wts = 1 + .num_ofm = 1 + .ifm_mem_dim = { + {65536, }, + }, + .wts_mem_dim = { + }, + .ofm_mem_dim = { + {65536, }, + }, + .ifm_mem_size = {65536, }, + .wts_mem_size = {}, + .ofm_mem_size = {65536, }, + .read_ifm_mem = { + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 12, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 12, }, + .offset = {0, 40, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 12, }, + .offset = {0, 80, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 12, }, + .offset = {0, 120, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_mem = { + }, + .write_ofm_mem = { + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 0, 0, 3, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 0, 0, 6, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 0, 0, 9, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 40, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 40, 0, 3, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 40, 0, 6, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 40, 0, 9, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 80, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 80, 0, 3, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 80, 0, 6, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 80, 0, 9, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 120, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 120, 0, 3, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 120, 0, 6, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + { + .buffer_dimension = {8, 160, 1, 48, }, + .tiling_dimension = {8, 40, 1, 3, }, + .offset = {0, 120, 0, 9, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 1}, + {.dim = 3, .stride = 12, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .mk_rep = 32 + .read_ifm_ddr = { + { + { + .buffer_dimension = {8, 320, 1, 180, }, + .tiling_dimension = {8, 160, 1, 24, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 2}, + {.dim = 3, .stride = 44, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 320, 1, 180, }, + .tiling_dimension = {8, 160, 1, 24, }, + .offset = {0, 0, 0, 24, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 2}, + {.dim = 3, .stride = 44, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ifm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .read_wts_ddr = { + }, + .write_wts_mem = { + }, + .read_ofm_mem = { + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {0, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {61440, }, + .tiling_dimension = {30720, }, + .offset = {30720, }, + .tiling_traversing ={ + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .write_ofm_ddr = { + { + { + .buffer_dimension = {8, 320, 1, 180, }, + .tiling_dimension = {8, 160, 1, 24, }, + .offset = {0, 0, 0, 0, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 2}, + {.dim = 3, .stride = 44, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + { + { + .buffer_dimension = {8, 320, 1, 180, }, + .tiling_dimension = {8, 160, 1, 24, }, + .offset = {0, 0, 0, 24, }, + .tiling_traversing ={ + {.dim = 1, .stride = 160, .wrap = 2}, + {.dim = 3, .stride = 44, .wrap = 4}, + } + .repetition = 1 + .buffer_iteration = 0 + } + } + }, + .ifm_mem_rep = 8 + .wts_mem_rep = 0 + .ofm_mem_rep = 8 +INFO: [aiecompiler 77-749] Reading logical device aie2p_8x4_device.json +INFO: [aiecompiler 77-6447] Executing Cmd: aieir_be --time-passes=0 --Xelfgen="-j4" --Xelfgen="-j1" --heapsize=1792 --Xpreproc="-D__IO_BUFFER_FORCE_LIGHT_WEIGHT__" --Xpreproc="-DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1" --target=hw --include="/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend" --include="/usr/local/lib/python3.10/site-packages/include/aie_api" --include="/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common" --include="/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common" --include="/usr/local/lib/python3.10/dist-packages/vitis_mllib" --include="/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc" --include="/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf" --stacksize=1400 --enable-partition=0:4 --nodot-graph=true --adf-api-log-level=0 --event-trace-port=gmio --part=xc10AIE2P_ML-die-0x-e-S-es1 --json="Work/temp/top.json" +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +INFO: [aiecompiler 77-749] Reading logical device aie2p_8x4_device.json +WARNING: [aiecompiler 77-22869] Forcing --gen-elf-ctrl-pkts=true when --multi-layer-pm-reload-elf-ctrl-pkt is true. +WARNING: [aiecompiler 77-5902] Under --enable-multi-layer, disablefloorplanning is added to --Xmapper option. +WARNING: [aiecompiler 77-21917] When --enable-multi-layer=true, broadcast_enable_core is forced to false +WARNING: [aiecompiler 77-5758] When --enableMultiLayer=true, disable-dma-autostart is forced to true +WARNING: [aiecompiler 77-21918] When --enableMultiLayer=true, enable-core-processor-bus is forced to true +WARNING: [aiecompiler 77-6502] aiecompiler option broadcast_enable_core=true is not supported for part xc10AIE2P_ML-die-0x-e-S-es1; forcing it to false. +WARNING: [aiecompiler 77-23075] Forcing --multi-layer-ctrl-pkt-column-span=0 when --multi-layer-prebuilt-archive is specified. +WARNING: [aiecompiler 77-22616] Part xc10AIE2P_ML-die-0x-e-S-es1 does not support ECC Scrubbing; ECC Scrubbing is disabled. +multi_layer_preemption_layers: +674: Work/temp/top.json +INFO: [aiecompiler 77-757] Opening input file: Work/temp/top.json +INFO: [aiecompiler 77-6287] Emitting AIEIr in file: Work/temp/top_aieir_dump.txt +INFO: [aiecompiler 77-656] Processing Graph 'root' +INFO: [aiecompiler 77-21798] ### Entering UnPacker pass +INFO: [aiecompiler 77-21797] ### Done with UnPacker pass (72392752 secs) +INFO: [aiecompiler 77-5563] Reading multi-layer graph constraints from 'Work/temp/top_meir_constraints.json' +Column 3Row 0Channel 0 + +i0_pi0(compute_graph.Layer_8_l2_wts.in[0], compute_graph.Layer_13_l2_wts.in[0], compute_graph.Layer_14_l2_wts.in[0], compute_graph.Layer_15_l2_wts.in[0], compute_graph.Layer_16_l2_wts.in[0], compute_graph.Layer_18_l2_wts.in[0], compute_graph.Layer_19_l2_wts.in[0], compute_graph.Layer_20_l2_wts.in[0], compute_graph.Layer_22_l2_wts.in[0], compute_graph.Layer_23_l2_wts.in[0], compute_graph.Layer_24_l2_wts.in[0], compute_graph.Layer_27_l2_wts.in[0], compute_graph.Layer_28_l2_wts.in[0], compute_graph.Layer_34_l2_wts.in[0], compute_graph.Layer_35_l2_wts.in[0], compute_graph.Layer_36_l2_wts.in[0], compute_graph.Layer_39_l2_wts.in[0], compute_graph.Layer_40_l2_wts.in[0], compute_graph.Layer_46_l2_wts.in[0], compute_graph.Layer_47_l2_wts.in[0], compute_graph.Layer_48_l2_wts.in[0], compute_graph.Layer_51_l2_wts.in[0], compute_graph.Layer_52_l2_wts.in[0], compute_graph.Layer_58_l2_wts.in[0], compute_graph.Layer_59_l2_wts.in[0], compute_graph.Layer_65_l2_wts.in[0], compute_graph.Layer_70_l2_wts.in[0], compute_graph.Layer_71_l2_wts.in[0], compute_graph.Layer_76_l2_wts.in[0], compute_graph.Layer_81_l2_wts.in[0], compute_graph.Layer_82_l2_wts.in[0], compute_graph.Layer_87_l2_wts.in[0], compute_graph.Layer_92_l2_wts.in[0], compute_graph.Layer_249_l2_wts.in[0], compute_graph.Layer_93_l2_wts.in[0], compute_graph.Layer_98_l2_wts.in[0], compute_graph.Layer_103_l2_wts.in[0], compute_graph.Layer_104_l2_wts.in[0], compute_graph.Layer_109_l2_wts.in[0], compute_graph.Layer_116_l2_wts.in[0], compute_graph.Layer_117_l2_wts.in[0], compute_graph.Layer_123_l2_wts.in[0], compute_graph.Layer_124_l2_wts.in[0], compute_graph.Layer_129_l2_wts.in[0], compute_graph.Layer_136_l2_wts.in[0], compute_graph.Layer_137_l2_wts.in[0], compute_graph.Layer_143_l2_wts.in[0], compute_graph.Layer_144_l2_wts.in[0], compute_graph.Layer_149_l2_wts.in[0], compute_graph.Layer_156_l2_wts.in[0], compute_graph.Layer_157_l2_wts.in[0], compute_graph.Layer_163_l2_wts.in[0], compute_graph.Layer_164_l2_wts.in[0], compute_graph.Layer_169_l2_wts.in[0], compute_graph.Layer_176_l2_wts.in[0], compute_graph.Layer_177_l2_wts.in[0], compute_graph.Layer_183_l2_wts.in[0], compute_graph.Layer_184_l2_wts.in[0], compute_graph.Layer_189_l2_wts.in[0], compute_graph.Layer_196_l2_wts.in[0], compute_graph.Layer_197_l2_wts.in[0], compute_graph.Layer_203_l2_wts.in[0], compute_graph.Layer_204_l2_wts.in[0], compute_graph.Layer_211_l2_wts.in[0], compute_graph.Layer_214_l2_wts.in[0], compute_graph.Layer_218_l2_wts.in[0], compute_graph.Layer_226_l2_wts.in[0], compute_graph.Layer_237_l2_wts.in[0], compute_graph.Layer_241_l2_wts.in[0], compute_graph.Layer_257_l2_wts.in[0], compute_graph.Layer_261_l2_wts.in[0], compute_graph.Layer_269_l2_wts.in[0], compute_graph.Layer_276_l2_wts.in[0], compute_graph.Layer_280_l2_wts.in[0], compute_graph.Layer_288_l2_wts.in[0], compute_graph.Layer_295_l2_wts.in[0], compute_graph.Layer_296_l2_wts.in[0], compute_graph.Layer_297_l2_wts.in[0], compute_graph.templated_graph_260.wts.in[0], compute_graph.templated_graph_268.wts.in[0], compute_graph.templated_graph_273.wts.in[0]) + +Column 3Row 0Channel 1 + +i0_pi1(compute_graph.Layer_8_l2_wts.in[1], compute_graph.Layer_13_l2_wts.in[1], compute_graph.Layer_14_l2_wts.in[1], compute_graph.Layer_15_l2_wts.in[1], compute_graph.Layer_16_l2_wts.in[1], compute_graph.Layer_18_l2_wts.in[1], compute_graph.Layer_19_l2_wts.in[1], compute_graph.Layer_20_l2_wts.in[1], compute_graph.Layer_22_l2_wts.in[1], compute_graph.Layer_23_l2_wts.in[1], compute_graph.Layer_24_l2_wts.in[1], compute_graph.Layer_27_l2_wts.in[1], compute_graph.Layer_28_l2_wts.in[1], compute_graph.Layer_34_l2_wts.in[1], compute_graph.Layer_35_l2_wts.in[1], compute_graph.Layer_36_l2_wts.in[1], compute_graph.Layer_92_l2_wts.in[1], compute_graph.Layer_39_l2_wts.in[1], compute_graph.Layer_40_l2_wts.in[1], compute_graph.Layer_46_l2_wts.in[1], compute_graph.Layer_47_l2_wts.in[1], compute_graph.Layer_48_l2_wts.in[1], compute_graph.Layer_51_l2_wts.in[1], compute_graph.Layer_52_l2_wts.in[1], compute_graph.Layer_58_l2_wts.in[1], compute_graph.Layer_59_l2_wts.in[1], compute_graph.Layer_65_l2_wts.in[1], compute_graph.Layer_70_l2_wts.in[1], compute_graph.Layer_71_l2_wts.in[1], compute_graph.Layer_76_l2_wts.in[1], compute_graph.Layer_81_l2_wts.in[1], compute_graph.Layer_82_l2_wts.in[1], compute_graph.Layer_87_l2_wts.in[1], compute_graph.Layer_218_l2_wts.in[1], compute_graph.Layer_93_l2_wts.in[1], compute_graph.Layer_98_l2_wts.in[1], compute_graph.Layer_103_l2_wts.in[1], compute_graph.Layer_104_l2_wts.in[1], compute_graph.Layer_109_l2_wts.in[1], compute_graph.Layer_116_l2_wts.in[1], compute_graph.Layer_117_l2_wts.in[1], compute_graph.Layer_123_l2_wts.in[1], compute_graph.Layer_124_l2_wts.in[1], compute_graph.Layer_129_l2_wts.in[1], compute_graph.Layer_136_l2_wts.in[1], compute_graph.Layer_137_l2_wts.in[1], compute_graph.Layer_143_l2_wts.in[1], compute_graph.Layer_144_l2_wts.in[1], compute_graph.Layer_149_l2_wts.in[1], compute_graph.Layer_156_l2_wts.in[1], compute_graph.Layer_157_l2_wts.in[1], compute_graph.Layer_163_l2_wts.in[1], compute_graph.Layer_164_l2_wts.in[1], compute_graph.Layer_169_l2_wts.in[1], compute_graph.Layer_176_l2_wts.in[1], compute_graph.Layer_177_l2_wts.in[1], compute_graph.Layer_183_l2_wts.in[1], compute_graph.Layer_184_l2_wts.in[1], compute_graph.Layer_189_l2_wts.in[1], compute_graph.Layer_196_l2_wts.in[1], compute_graph.Layer_197_l2_wts.in[1], compute_graph.Layer_203_l2_wts.in[1], compute_graph.Layer_204_l2_wts.in[1], compute_graph.Layer_211_l2_wts.in[1], compute_graph.Layer_214_l2_wts.in[1], compute_graph.Layer_226_l2_wts.in[1], compute_graph.Layer_237_l2_wts.in[1], compute_graph.Layer_241_l2_wts.in[1], compute_graph.Layer_249_l2_wts.in[1], compute_graph.Layer_257_l2_wts.in[1], compute_graph.Layer_261_l2_wts.in[1], compute_graph.Layer_269_l2_wts.in[1], compute_graph.Layer_276_l2_wts.in[1], compute_graph.Layer_280_l2_wts.in[1], compute_graph.Layer_288_l2_wts.in[1], compute_graph.Layer_295_l2_wts.in[1], compute_graph.Layer_296_l2_wts.in[1], compute_graph.Layer_297_l2_wts.in[1]) + +Column 3Row 0Channel 0 + +i0_po0(compute_graph.Layer_8_l2_wts.out[0], compute_graph.Layer_13_l2_wts.out[0], compute_graph.Layer_14_l2_wts.out[0], compute_graph.Layer_15_l2_wts.out[0], compute_graph.Layer_16_l2_wts.out[0], compute_graph.Layer_18_l2_wts.out[0], compute_graph.Layer_19_l2_wts.out[0], compute_graph.Layer_20_l2_wts.out[0], compute_graph.Layer_22_l2_wts.out[0], compute_graph.Layer_23_l2_wts.out[0], compute_graph.Layer_24_l2_wts.out[0], compute_graph.Layer_27_l2_wts.out[0], compute_graph.Layer_28_l2_wts.out[0], compute_graph.Layer_34_l2_wts.out[0], compute_graph.Layer_35_l2_wts.out[0], compute_graph.Layer_36_l2_wts.out[0], compute_graph.Layer_92_l2_wts.out[0], compute_graph.Layer_39_l2_wts.out[0], compute_graph.Layer_40_l2_wts.out[0], compute_graph.Layer_46_l2_wts.out[0], compute_graph.Layer_47_l2_wts.out[0], compute_graph.Layer_48_l2_wts.out[0], compute_graph.Layer_51_l2_wts.out[0], compute_graph.Layer_52_l2_wts.out[0], compute_graph.Layer_58_l2_wts.out[0], compute_graph.Layer_59_l2_wts.out[0], compute_graph.Layer_65_l2_wts.out[0], compute_graph.Layer_70_l2_wts.out[0], compute_graph.Layer_71_l2_wts.out[0], compute_graph.Layer_76_l2_wts.out[0], compute_graph.Layer_81_l2_wts.out[0], compute_graph.Layer_82_l2_wts.out[0], compute_graph.Layer_87_l2_wts.out[0], compute_graph.Layer_218_l2_wts.out[0], compute_graph.Layer_93_l2_wts.out[0], compute_graph.Layer_98_l2_wts.out[0], compute_graph.Layer_103_l2_wts.out[0], compute_graph.Layer_104_l2_wts.out[0], compute_graph.Layer_109_l2_wts.out[0], compute_graph.Layer_116_l2_wts.out[0], compute_graph.Layer_117_l2_wts.out[0], compute_graph.Layer_123_l2_wts.out[0], compute_graph.Layer_124_l2_wts.out[0], compute_graph.Layer_129_l2_wts.out[0], compute_graph.Layer_136_l2_wts.out[0], compute_graph.Layer_137_l2_wts.out[0], compute_graph.Layer_143_l2_wts.out[0], compute_graph.Layer_144_l2_wts.out[0], compute_graph.Layer_149_l2_wts.out[0], compute_graph.Layer_156_l2_wts.out[0], compute_graph.Layer_157_l2_wts.out[0], compute_graph.Layer_163_l2_wts.out[0], compute_graph.Layer_164_l2_wts.out[0], compute_graph.Layer_169_l2_wts.out[0], compute_graph.Layer_176_l2_wts.out[0], compute_graph.Layer_177_l2_wts.out[0], compute_graph.Layer_183_l2_wts.out[0], compute_graph.Layer_184_l2_wts.out[0], compute_graph.Layer_189_l2_wts.out[0], compute_graph.Layer_196_l2_wts.out[0], compute_graph.Layer_197_l2_wts.out[0], compute_graph.Layer_203_l2_wts.out[0], compute_graph.Layer_204_l2_wts.out[0], compute_graph.Layer_211_l2_wts.out[0], compute_graph.Layer_214_l2_wts.out[0], compute_graph.Layer_226_l2_wts.out[0], compute_graph.Layer_237_l2_wts.out[0], compute_graph.Layer_241_l2_wts.out[0], compute_graph.Layer_249_l2_wts.out[0], compute_graph.Layer_257_l2_wts.out[0], compute_graph.Layer_261_l2_wts.out[0], compute_graph.Layer_269_l2_wts.out[0], compute_graph.Layer_276_l2_wts.out[0], compute_graph.Layer_280_l2_wts.out[0], compute_graph.Layer_288_l2_wts.out[0], compute_graph.Layer_295_l2_wts.out[0], compute_graph.Layer_296_l2_wts.out[0], compute_graph.Layer_297_l2_wts.out[0], compute_graph.templated_graph_260.wts.out[0], compute_graph.templated_graph_268.wts.out[0], compute_graph.templated_graph_273.wts.out[0]) + +Column 3Row 0Channel 1 + +i0_po1(compute_graph.Layer_8_l2_wts.out[1], compute_graph.Layer_13_l2_wts.out[1], compute_graph.Layer_14_l2_wts.out[1], compute_graph.Layer_15_l2_wts.out[1], compute_graph.Layer_16_l2_wts.out[1], compute_graph.Layer_18_l2_wts.out[1], compute_graph.Layer_19_l2_wts.out[1], compute_graph.Layer_20_l2_wts.out[1], compute_graph.Layer_22_l2_wts.out[1], compute_graph.Layer_23_l2_wts.out[1], compute_graph.Layer_24_l2_wts.out[1], compute_graph.Layer_27_l2_wts.out[1], compute_graph.Layer_28_l2_wts.out[1], compute_graph.Layer_34_l2_wts.out[1], compute_graph.Layer_35_l2_wts.out[1], compute_graph.Layer_36_l2_wts.out[1], compute_graph.Layer_92_l2_wts.out[1], compute_graph.Layer_39_l2_wts.out[1], compute_graph.Layer_40_l2_wts.out[1], compute_graph.Layer_46_l2_wts.out[1], compute_graph.Layer_47_l2_wts.out[1], compute_graph.Layer_48_l2_wts.out[1], compute_graph.Layer_51_l2_wts.out[1], compute_graph.Layer_52_l2_wts.out[1], compute_graph.Layer_58_l2_wts.out[1], compute_graph.Layer_59_l2_wts.out[1], compute_graph.Layer_65_l2_wts.out[1], compute_graph.Layer_70_l2_wts.out[1], compute_graph.Layer_71_l2_wts.out[1], compute_graph.Layer_76_l2_wts.out[1], compute_graph.Layer_81_l2_wts.out[1], compute_graph.Layer_82_l2_wts.out[1], compute_graph.Layer_87_l2_wts.out[1], compute_graph.Layer_218_l2_wts.out[1], compute_graph.Layer_93_l2_wts.out[1], compute_graph.Layer_98_l2_wts.out[1], compute_graph.Layer_103_l2_wts.out[1], compute_graph.Layer_104_l2_wts.out[1], compute_graph.Layer_109_l2_wts.out[1], compute_graph.Layer_116_l2_wts.out[1], compute_graph.Layer_117_l2_wts.out[1], compute_graph.Layer_123_l2_wts.out[1], compute_graph.Layer_124_l2_wts.out[1], compute_graph.Layer_129_l2_wts.out[1], compute_graph.Layer_136_l2_wts.out[1], compute_graph.Layer_137_l2_wts.out[1], compute_graph.Layer_143_l2_wts.out[1], compute_graph.Layer_144_l2_wts.out[1], compute_graph.Layer_149_l2_wts.out[1], compute_graph.Layer_156_l2_wts.out[1], compute_graph.Layer_157_l2_wts.out[1], compute_graph.Layer_163_l2_wts.out[1], compute_graph.Layer_164_l2_wts.out[1], compute_graph.Layer_169_l2_wts.out[1], compute_graph.Layer_176_l2_wts.out[1], compute_graph.Layer_177_l2_wts.out[1], compute_graph.Layer_183_l2_wts.out[1], compute_graph.Layer_184_l2_wts.out[1], compute_graph.Layer_189_l2_wts.out[1], compute_graph.Layer_196_l2_wts.out[1], compute_graph.Layer_197_l2_wts.out[1], compute_graph.Layer_203_l2_wts.out[1], compute_graph.Layer_204_l2_wts.out[1], compute_graph.Layer_211_l2_wts.out[1], compute_graph.Layer_214_l2_wts.out[1], compute_graph.Layer_226_l2_wts.out[1], compute_graph.Layer_237_l2_wts.out[1], compute_graph.Layer_241_l2_wts.out[1], compute_graph.Layer_249_l2_wts.out[1], compute_graph.Layer_257_l2_wts.out[1], compute_graph.Layer_261_l2_wts.out[1], compute_graph.Layer_269_l2_wts.out[1], compute_graph.Layer_276_l2_wts.out[1], compute_graph.Layer_280_l2_wts.out[1], compute_graph.Layer_288_l2_wts.out[1], compute_graph.Layer_295_l2_wts.out[1], compute_graph.Layer_296_l2_wts.out[1], compute_graph.Layer_297_l2_wts.out[1], compute_graph.templated_graph_260.wts.out[1], compute_graph.templated_graph_268.wts.out[1], compute_graph.templated_graph_273.wts.out[1]) + +Column 3Row 0Channel 2 + +i0_po2(compute_graph.Layer_8_l2_wts.out[2], compute_graph.Layer_13_l2_wts.out[2], compute_graph.Layer_14_l2_wts.out[2], compute_graph.Layer_15_l2_wts.out[2], compute_graph.Layer_16_l2_wts.out[2], compute_graph.Layer_18_l2_wts.out[2], compute_graph.Layer_19_l2_wts.out[2], compute_graph.Layer_20_l2_wts.out[2], compute_graph.Layer_22_l2_wts.out[2], compute_graph.Layer_23_l2_wts.out[2], compute_graph.Layer_24_l2_wts.out[2], compute_graph.Layer_27_l2_wts.out[2], compute_graph.Layer_28_l2_wts.out[2], compute_graph.Layer_34_l2_wts.out[2], compute_graph.Layer_35_l2_wts.out[2], compute_graph.Layer_36_l2_wts.out[2], compute_graph.Layer_92_l2_wts.out[2], compute_graph.Layer_39_l2_wts.out[2], compute_graph.Layer_40_l2_wts.out[2], compute_graph.Layer_46_l2_wts.out[2], compute_graph.Layer_47_l2_wts.out[2], compute_graph.Layer_48_l2_wts.out[2], compute_graph.Layer_51_l2_wts.out[2], compute_graph.Layer_52_l2_wts.out[2], compute_graph.Layer_58_l2_wts.out[2], compute_graph.Layer_59_l2_wts.out[2], compute_graph.Layer_65_l2_wts.out[2], compute_graph.Layer_70_l2_wts.out[2], compute_graph.Layer_71_l2_wts.out[2], compute_graph.Layer_76_l2_wts.out[2], compute_graph.Layer_81_l2_wts.out[2], compute_graph.Layer_82_l2_wts.out[2], compute_graph.Layer_87_l2_wts.out[2], compute_graph.Layer_218_l2_wts.out[2], compute_graph.Layer_93_l2_wts.out[2], compute_graph.Layer_98_l2_wts.out[2], compute_graph.Layer_103_l2_wts.out[2], compute_graph.Layer_104_l2_wts.out[2], compute_graph.Layer_109_l2_wts.out[2], compute_graph.Layer_116_l2_wts.out[2], compute_graph.Layer_117_l2_wts.out[2], compute_graph.Layer_123_l2_wts.out[2], compute_graph.Layer_124_l2_wts.out[2], compute_graph.Layer_129_l2_wts.out[2], compute_graph.Layer_136_l2_wts.out[2], compute_graph.Layer_137_l2_wts.out[2], compute_graph.Layer_143_l2_wts.out[2], compute_graph.Layer_144_l2_wts.out[2], compute_graph.Layer_149_l2_wts.out[2], compute_graph.Layer_156_l2_wts.out[2], compute_graph.Layer_157_l2_wts.out[2], compute_graph.Layer_163_l2_wts.out[2], compute_graph.Layer_164_l2_wts.out[2], compute_graph.Layer_169_l2_wts.out[2], compute_graph.Layer_176_l2_wts.out[2], compute_graph.Layer_177_l2_wts.out[2], compute_graph.Layer_183_l2_wts.out[2], compute_graph.Layer_184_l2_wts.out[2], compute_graph.Layer_189_l2_wts.out[2], compute_graph.Layer_196_l2_wts.out[2], compute_graph.Layer_197_l2_wts.out[2], compute_graph.Layer_203_l2_wts.out[2], compute_graph.Layer_204_l2_wts.out[2], compute_graph.Layer_211_l2_wts.out[2], compute_graph.Layer_214_l2_wts.out[2], compute_graph.Layer_226_l2_wts.out[2], compute_graph.Layer_237_l2_wts.out[2], compute_graph.Layer_241_l2_wts.out[2], compute_graph.Layer_249_l2_wts.out[2], compute_graph.Layer_257_l2_wts.out[2], compute_graph.Layer_261_l2_wts.out[2], compute_graph.Layer_269_l2_wts.out[2], compute_graph.Layer_276_l2_wts.out[2], compute_graph.Layer_280_l2_wts.out[2], compute_graph.Layer_288_l2_wts.out[2], compute_graph.Layer_295_l2_wts.out[2], compute_graph.Layer_296_l2_wts.out[2], compute_graph.Layer_297_l2_wts.out[2], compute_graph.templated_graph_260.wts.out[2], compute_graph.templated_graph_268.wts.out[2], compute_graph.templated_graph_273.wts.out[2]) + +Column 3Row 0Channel 3 + +i0_po3(compute_graph.Layer_8_l2_wts.out[3], compute_graph.Layer_13_l2_wts.out[3], compute_graph.Layer_14_l2_wts.out[3], compute_graph.Layer_15_l2_wts.out[3], compute_graph.Layer_16_l2_wts.out[3], compute_graph.Layer_18_l2_wts.out[3], compute_graph.Layer_19_l2_wts.out[3], compute_graph.Layer_20_l2_wts.out[3], compute_graph.Layer_22_l2_wts.out[3], compute_graph.Layer_23_l2_wts.out[3], compute_graph.Layer_24_l2_wts.out[3], compute_graph.Layer_27_l2_wts.out[3], compute_graph.Layer_28_l2_wts.out[3], compute_graph.Layer_34_l2_wts.out[3], compute_graph.Layer_35_l2_wts.out[3], compute_graph.Layer_36_l2_wts.out[3], compute_graph.Layer_39_l2_wts.out[3], compute_graph.Layer_40_l2_wts.out[3], compute_graph.Layer_46_l2_wts.out[3], compute_graph.Layer_47_l2_wts.out[3], compute_graph.Layer_48_l2_wts.out[3], compute_graph.Layer_51_l2_wts.out[3], compute_graph.Layer_52_l2_wts.out[3], compute_graph.Layer_58_l2_wts.out[3], compute_graph.Layer_59_l2_wts.out[3], compute_graph.Layer_65_l2_wts.out[3], compute_graph.Layer_70_l2_wts.out[3], compute_graph.Layer_71_l2_wts.out[3], compute_graph.Layer_76_l2_wts.out[3], compute_graph.Layer_81_l2_wts.out[3], compute_graph.Layer_82_l2_wts.out[3], compute_graph.Layer_87_l2_wts.out[3], compute_graph.Layer_92_l2_wts.out[3], compute_graph.Layer_93_l2_wts.out[3], compute_graph.Layer_98_l2_wts.out[3], compute_graph.Layer_103_l2_wts.out[3], compute_graph.Layer_104_l2_wts.out[3], compute_graph.Layer_109_l2_wts.out[3], compute_graph.Layer_116_l2_wts.out[3], compute_graph.Layer_117_l2_wts.out[3], compute_graph.Layer_123_l2_wts.out[3], compute_graph.Layer_124_l2_wts.out[3], compute_graph.Layer_129_l2_wts.out[3], compute_graph.Layer_136_l2_wts.out[3], compute_graph.Layer_137_l2_wts.out[3], compute_graph.Layer_143_l2_wts.out[3], compute_graph.Layer_144_l2_wts.out[3], compute_graph.Layer_149_l2_wts.out[3], compute_graph.Layer_156_l2_wts.out[3], compute_graph.Layer_157_l2_wts.out[3], compute_graph.Layer_163_l2_wts.out[3], compute_graph.Layer_164_l2_wts.out[3], compute_graph.Layer_169_l2_wts.out[3], compute_graph.Layer_176_l2_wts.out[3], compute_graph.Layer_177_l2_wts.out[3], compute_graph.Layer_183_l2_wts.out[3], compute_graph.Layer_184_l2_wts.out[3], compute_graph.Layer_189_l2_wts.out[3], compute_graph.Layer_196_l2_wts.out[3], compute_graph.Layer_197_l2_wts.out[3], compute_graph.Layer_203_l2_wts.out[3], compute_graph.Layer_204_l2_wts.out[3], compute_graph.Layer_211_l2_wts.out[3], compute_graph.Layer_214_l2_wts.out[3], compute_graph.Layer_218_l2_wts.out[3], compute_graph.Layer_226_l2_wts.out[3], compute_graph.Layer_237_l2_wts.out[3], compute_graph.Layer_241_l2_wts.out[3], compute_graph.Layer_249_l2_wts.out[3], compute_graph.Layer_257_l2_wts.out[3], compute_graph.Layer_261_l2_wts.out[3], compute_graph.Layer_269_l2_wts.out[3], compute_graph.Layer_276_l2_wts.out[3], compute_graph.Layer_280_l2_wts.out[3], compute_graph.Layer_288_l2_wts.out[3], compute_graph.Layer_295_l2_wts.out[3], compute_graph.Layer_296_l2_wts.out[3], compute_graph.Layer_297_l2_wts.out[3], compute_graph.templated_graph_260.wts.out[3], compute_graph.templated_graph_268.wts.out[3], compute_graph.templated_graph_273.wts.out[3]) + +Column 2Row 0Channel 0 + +i1_pi0(compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.in[0], compute_graph.L2_IFM_Buffer_for_input1_1_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.in[0], compute_graph.L2_IFM_Buffer_for_input1_2_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.in[0]) + +Column 2Row 0Channel 1 + +i1_pi1(compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.in[1], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.in[1], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.in[1], compute_graph.L2_IFM_Buffer_for_input1_1_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.in[1], compute_graph.L2_IFM_Buffer_for_input1_2_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.in[1]) + +Column 2Row 0Channel 0 + +i1_po0(compute_graph.l2_9.out[1], compute_graph.l2_11.out[1], compute_graph.l2_13.out[1], compute_graph.l2_14.out[0], compute_graph.l2_14.out[2], compute_graph.l2_16.out[0], compute_graph.l2_18.out[4], compute_graph.l2_59.out[4], compute_graph.templated_graph_49.trans_mt_ofm.out[0], compute_graph.templated_graph_56.g0.ifm_mem.out[0], compute_graph.templated_graph_56.g1.ifm_mem.out[0], compute_graph.templated_graph_56.g2.ifm_mem.out[0], compute_graph.templated_graph_64.ofm.out[0], compute_graph.templated_graph_64.ifm2.out[0], compute_graph.templated_graph_114.trans_mt_ofm.out[0], compute_graph.templated_graph_121.g0.ifm_mem.out[0], compute_graph.templated_graph_121.g1.ifm_mem.out[0], compute_graph.templated_graph_121.g2.ifm_mem.out[0], compute_graph.l2_164.out[4], compute_graph.l2_169.out[4], compute_graph.l2_184.out[4], compute_graph.l2_189.out[4], compute_graph.templated_graph_293.ofm_mem.out[0], compute_graph.l2_204.out[4], compute_graph.l2_234.out[4], compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.out[0], compute_graph.l2_257.out[0], compute_graph.l2_276.out[0], compute_graph.l2_280.out[0], compute_graph.l2_288.out[0], compute_graph.l2_289.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.out[0], compute_graph.l2_295.out[0], compute_graph.templated_graph_2.ifm.out[0], compute_graph.templated_graph_3.ofm.out[0], compute_graph.templated_graph_3.ifm2.out[0], compute_graph.templated_graph_4.trans_mt_ofm.out[0], compute_graph.templated_graph_5.ofm.out[0], compute_graph.templated_graph_5.ifm2.out[0], compute_graph.templated_graph_17.ofm.out[0], compute_graph.templated_graph_17.ifm2.out[0], compute_graph.templated_graph_21.ofm.out[0], compute_graph.templated_graph_21.ifm2.out[0], compute_graph.templated_graph_25.trans_mt_ofm.out[0], compute_graph.templated_graph_32.g0.ifm_mem.out[0], compute_graph.templated_graph_32.g1.ifm_mem.out[0], compute_graph.templated_graph_32.g2.ifm_mem.out[0], compute_graph.templated_graph_37.trans_mt_ofm.out[0], compute_graph.templated_graph_44.g0.ifm_mem.out[0], compute_graph.templated_graph_44.g1.ifm_mem.out[0], compute_graph.templated_graph_44.g2.ifm_mem.out[0], compute_graph.templated_graph_134.trans_mt_ofm.out[0], compute_graph.templated_graph_141.g0.ifm_mem.out[0], compute_graph.templated_graph_141.g1.ifm_mem.out[0], compute_graph.templated_graph_141.g2.ifm_mem.out[0], compute_graph.templated_graph_154.trans_mt_ofm.out[0], compute_graph.templated_graph_161.g0.ifm_mem.out[0], compute_graph.templated_graph_161.g1.ifm_mem.out[0], compute_graph.templated_graph_161.g2.ifm_mem.out[0], compute_graph.templated_graph_174.trans_mt_ofm.out[0], compute_graph.templated_graph_181.g0.ifm_mem.out[0], compute_graph.templated_graph_181.g1.ifm_mem.out[0], compute_graph.templated_graph_181.g2.ifm_mem.out[0], compute_graph.templated_graph_194.trans_mt_ofm.out[0], compute_graph.templated_graph_201.g0.ifm_mem.out[0], compute_graph.templated_graph_201.g1.ifm_mem.out[0], compute_graph.templated_graph_201.g2.ifm_mem.out[0], compute_graph.templated_graph_209.trans_mt_ofm.out[0], compute_graph.templated_graph_213.g0.ifm_mem.out[0], compute_graph.templated_graph_213.g1.ifm_mem.out[0], compute_graph.templated_graph_213.g2.ifm_mem.out[0], compute_graph.templated_graph_215.ifm.out[0], compute_graph.templated_graph_216.ifm.out[0], compute_graph.templated_graph_220.ifm.out[0], compute_graph.templated_graph_223.ifm.out[0], compute_graph.templated_graph_231.ofm_mem.out[0], compute_graph.templated_graph_232.ifm.out[0], compute_graph.templated_graph_238.ifm.out[0], compute_graph.templated_graph_239.ifm.out[0], compute_graph.templated_graph_243.ifm.out[0], compute_graph.templated_graph_246.ifm.out[0], compute_graph.templated_graph_254.ofm_mem.out[0], compute_graph.templated_graph_255.ifm.out[0], compute_graph.templated_graph_258.ofm.out[0], compute_graph.templated_graph_259.ofm.out[0], compute_graph.templated_graph_260.ofm.out[0], compute_graph.templated_graph_263.ofm.out[0], compute_graph.templated_graph_266.ofm.out[0], compute_graph.templated_graph_268.ofm.out[0], compute_graph.templated_graph_273.ofm.out[0], compute_graph.templated_graph_274.ofm_mem.out[0], compute_graph.templated_graph_277.ifm.out[0], compute_graph.templated_graph_278.ifm.out[0], compute_graph.templated_graph_282.ifm.out[0], compute_graph.templated_graph_285.ifm.out[0], compute_graph.templated_graph_298.ofm.out[0], compute_graph.templated_graph_300.ofm.out[0]) + +Column 2Row 0Channel 1 + +i1_po1(compute_graph.l2_14.out[1], compute_graph.l2_14.out[3], compute_graph.l2_16.out[1], compute_graph.l2_18.out[5], compute_graph.l2_59.out[5], compute_graph.l2_164.out[5], compute_graph.l2_169.out[5], compute_graph.l2_184.out[5], compute_graph.l2_189.out[5], compute_graph.templated_graph_293.ofm_mem.out[1], compute_graph.l2_204.out[5], compute_graph.l2_234.out[5], compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.out[1], compute_graph.l2_257.out[1], compute_graph.l2_276.out[1], compute_graph.l2_280.out[1], compute_graph.l2_288.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.out[1], compute_graph.l2_295.out[1], compute_graph.templated_graph_2.ifm.out[1], compute_graph.templated_graph_215.ifm.out[1], compute_graph.templated_graph_216.ifm.out[1], compute_graph.templated_graph_220.ifm.out[1], compute_graph.templated_graph_223.ifm.out[1], compute_graph.templated_graph_231.ofm_mem.out[1], compute_graph.templated_graph_232.ifm.out[1], compute_graph.templated_graph_238.ifm.out[1], compute_graph.templated_graph_239.ifm.out[1], compute_graph.templated_graph_243.ifm.out[1], compute_graph.templated_graph_246.ifm.out[1], compute_graph.templated_graph_254.ofm_mem.out[1], compute_graph.templated_graph_255.ifm.out[1], compute_graph.templated_graph_258.ofm.out[1], compute_graph.templated_graph_259.ofm.out[1], compute_graph.templated_graph_263.ofm.out[1], compute_graph.templated_graph_266.ofm.out[1], compute_graph.templated_graph_274.ofm_mem.out[1], compute_graph.templated_graph_277.ifm.out[1], compute_graph.templated_graph_278.ifm.out[1], compute_graph.templated_graph_282.ifm.out[1], compute_graph.templated_graph_285.ifm.out[1], compute_graph.templated_graph_298.ofm.out[1], compute_graph.templated_graph_300.ofm.out[1]) + +Column 0Row 0Channel 0 + +i2_pi0(compute_graph.L2_IFM_Buffer_for_input0_0_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.in[0], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.in[0], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.in[0], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.in[0], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.in[0], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.in[0], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.in[0], compute_graph.templated_graph_49.trans_mt_ifm.in[0], compute_graph.templated_graph_56.g0.ifm_mem.in[0], compute_graph.templated_graph_56.g1.ifm_mem.in[0], compute_graph.templated_graph_56.g2.ifm_mem.in[0], compute_graph.templated_graph_64.ifm.in[0], compute_graph.templated_graph_64.ifm2.in[0], compute_graph.templated_graph_114.trans_mt_ifm.in[0], compute_graph.templated_graph_121.g0.ifm_mem.in[0], compute_graph.templated_graph_121.g1.ifm_mem.in[0], compute_graph.templated_graph_121.g2.ifm_mem.in[0], compute_graph.templated_graph_134.trans_mt_ifm.in[0], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.in[0], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.in[0], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.in[0], compute_graph.L2_IFM_Buffer_for_input4_0_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.in[0], compute_graph.L2_IFM_Buffer_for_input4_1_port1.in[0], compute_graph.L2_IFM_Buffer_for_input4_2_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.in[0], compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.in[0], compute_graph.L2_IFM_Buffer_for_input3_0_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.in[0], compute_graph.L2_IFM_Buffer_for_input3_1_port1.in[0], compute_graph.L2_IFM_Buffer_for_input3_2_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.in[0], compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.in[0], compute_graph.L2_IFM_Buffer_for_input2_1_port1.in[0], compute_graph.L2_IFM_Buffer_for_input2_2_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.in[0], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.in[0], compute_graph.L2_IFM_Buffer_for_input1_0_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[0], compute_graph.templated_graph_2.ifm.in[0], compute_graph.templated_graph_3.ifm.in[0], compute_graph.templated_graph_3.ifm2.in[0], compute_graph.templated_graph_4.trans_mt_ifm.in[0], compute_graph.templated_graph_5.ifm.in[0], compute_graph.templated_graph_5.ifm2.in[0], compute_graph.templated_graph_17.ifm.in[0], compute_graph.templated_graph_17.ifm2.in[0], compute_graph.templated_graph_21.ifm.in[0], compute_graph.templated_graph_21.ifm2.in[0], compute_graph.templated_graph_25.trans_mt_ifm.in[0], compute_graph.templated_graph_32.g0.ifm_mem.in[0], compute_graph.templated_graph_32.g1.ifm_mem.in[0], compute_graph.templated_graph_32.g2.ifm_mem.in[0], compute_graph.templated_graph_37.trans_mt_ifm.in[0], compute_graph.templated_graph_44.g0.ifm_mem.in[0], compute_graph.templated_graph_44.g1.ifm_mem.in[0], compute_graph.templated_graph_44.g2.ifm_mem.in[0], compute_graph.templated_graph_141.g0.ifm_mem.in[0], compute_graph.templated_graph_141.g1.ifm_mem.in[0], compute_graph.templated_graph_141.g2.ifm_mem.in[0], compute_graph.templated_graph_154.trans_mt_ifm.in[0], compute_graph.templated_graph_161.g0.ifm_mem.in[0], compute_graph.templated_graph_161.g1.ifm_mem.in[0], compute_graph.templated_graph_161.g2.ifm_mem.in[0], compute_graph.templated_graph_174.trans_mt_ifm.in[0], compute_graph.templated_graph_181.g0.ifm_mem.in[0], compute_graph.templated_graph_181.g1.ifm_mem.in[0], compute_graph.templated_graph_181.g2.ifm_mem.in[0], compute_graph.templated_graph_194.trans_mt_ifm.in[0], compute_graph.templated_graph_201.g0.ifm_mem.in[0], compute_graph.templated_graph_201.g1.ifm_mem.in[0], compute_graph.templated_graph_201.g2.ifm_mem.in[0], compute_graph.templated_graph_209.trans_mt_ifm.in[0], compute_graph.templated_graph_213.g0.ifm_mem.in[0], compute_graph.templated_graph_213.g1.ifm_mem.in[0], compute_graph.templated_graph_213.g2.ifm_mem.in[0], compute_graph.templated_graph_215.ifm.in[0], compute_graph.templated_graph_216.ifm.in[0], compute_graph.templated_graph_220.ifm.in[0], compute_graph.templated_graph_223.ifm.in[0], compute_graph.templated_graph_231.ifm_mem.in[0], compute_graph.templated_graph_232.ifm.in[0], compute_graph.templated_graph_238.ifm.in[0], compute_graph.templated_graph_239.ifm.in[0], compute_graph.templated_graph_243.ifm.in[0], compute_graph.templated_graph_246.ifm.in[0], compute_graph.templated_graph_254.ifm_mem.in[0], compute_graph.templated_graph_255.ifm.in[0], compute_graph.templated_graph_258.ifm.in[0], compute_graph.templated_graph_259.ifm.in[0], compute_graph.templated_graph_260.ifm.in[0], compute_graph.templated_graph_263.ifm.in[0], compute_graph.templated_graph_266.ifm.in[0], compute_graph.templated_graph_268.ifm.in[0], compute_graph.templated_graph_273.ifm.in[0], compute_graph.templated_graph_274.ifm_mem.in[0], compute_graph.templated_graph_277.ifm.in[0], compute_graph.templated_graph_278.ifm.in[0], compute_graph.templated_graph_282.ifm.in[0], compute_graph.templated_graph_285.ifm.in[0], compute_graph.templated_graph_293.ifm_mem.in[0], compute_graph.templated_graph_298.ifm.in[0], compute_graph.templated_graph_300.ifm.in[0]) + +Column 0Row 0Channel 1 + +i2_pi1(compute_graph.L2_IFM_Buffer_for_input0_0_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.in[1], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.in[1], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.in[1], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.in[1], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.in[1], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.in[1], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.in[1], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.in[1], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.in[1], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.in[1], compute_graph.L2_IFM_Buffer_for_input4_0_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.in[1], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.in[1], compute_graph.L2_IFM_Buffer_for_input4_1_port1.in[1], compute_graph.L2_IFM_Buffer_for_input4_2_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.in[1], compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.in[1], compute_graph.L2_IFM_Buffer_for_input3_0_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.in[1], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.in[1], compute_graph.L2_IFM_Buffer_for_input3_1_port1.in[1], compute_graph.L2_IFM_Buffer_for_input3_2_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.in[1], compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.in[1], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.in[1], compute_graph.L2_IFM_Buffer_for_input2_1_port1.in[1], compute_graph.L2_IFM_Buffer_for_input2_2_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.in[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.in[1], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.in[1], compute_graph.L2_IFM_Buffer_for_input1_0_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[1], compute_graph.templated_graph_2.ifm.in[1], compute_graph.templated_graph_215.ifm.in[1], compute_graph.templated_graph_216.ifm.in[1], compute_graph.templated_graph_220.ifm.in[1], compute_graph.templated_graph_223.ifm.in[1], compute_graph.templated_graph_231.ifm_mem.in[1], compute_graph.templated_graph_232.ifm.in[1], compute_graph.templated_graph_238.ifm.in[1], compute_graph.templated_graph_239.ifm.in[1], compute_graph.templated_graph_243.ifm.in[1], compute_graph.templated_graph_246.ifm.in[1], compute_graph.templated_graph_254.ifm_mem.in[1], compute_graph.templated_graph_255.ifm.in[1], compute_graph.templated_graph_258.ifm.in[1], compute_graph.templated_graph_259.ifm.in[1], compute_graph.templated_graph_263.ifm.in[1], compute_graph.templated_graph_266.ifm.in[1], compute_graph.templated_graph_274.ifm_mem.in[1], compute_graph.templated_graph_277.ifm.in[1], compute_graph.templated_graph_278.ifm.in[1], compute_graph.templated_graph_282.ifm.in[1], compute_graph.templated_graph_285.ifm.in[1], compute_graph.templated_graph_293.ifm_mem.in[1], compute_graph.templated_graph_298.ifm.in[1], compute_graph.templated_graph_300.ifm.in[1]) + +Column 0Row 0Channel 0 + +i2_po0(compute_graph.l2_1.out[0], compute_graph.l2_6.out[0], compute_graph.l2_7.out[0], compute_graph.l2_8.out[0], compute_graph.l2_9.out[0], compute_graph.l2_10.out[0], compute_graph.l2_11.out[0], compute_graph.l2_12.out[0], compute_graph.l2_13.out[0], compute_graph.l2_15.out[0], compute_graph.l2_19.out[0], compute_graph.l2_20.out[0], compute_graph.l2_22.out[4], compute_graph.l2_24.out[0], compute_graph.l2_24.out[2], compute_graph.l2_31.out[0], compute_graph.l2_34.out[4], compute_graph.l2_36.out[0], compute_graph.l2_36.out[2], compute_graph.l2_43.out[0], compute_graph.l2_46.out[4], compute_graph.l2_48.out[0], compute_graph.l2_48.out[2], compute_graph.l2_192.out[0], compute_graph.l2_55.out[0], compute_graph.l2_58.out[4], compute_graph.l2_62.out[0], compute_graph.l2_63.out[0], compute_graph.l2_113.out[0], compute_graph.l2_113.out[2], compute_graph.l2_120.out[0], compute_graph.l2_123.out[4], compute_graph.l2_133.out[0], compute_graph.l2_133.out[2], compute_graph.l2_140.out[0], compute_graph.l2_153.out[0], compute_graph.l2_153.out[2], compute_graph.l2_160.out[0], compute_graph.l2_163.out[4], compute_graph.l2_167.out[0], compute_graph.l2_168.out[0], compute_graph.l2_172.out[0], compute_graph.l2_173.out[0], compute_graph.l2_180.out[0], compute_graph.l2_183.out[4], compute_graph.l2_187.out[0], compute_graph.l2_188.out[0], compute_graph.l2_193.out[0], compute_graph.l2_200.out[0], compute_graph.l2_207.out[0], compute_graph.l2_208.out[0], compute_graph.l2_212.out[0], compute_graph.l2_214.out[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.out[0], compute_graph.L2_IFM_Buffer_for_input4_0_port1.out[0], compute_graph.l2_219.out[0], compute_graph.l2_222.out[0], compute_graph.l2_224.out[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.out[0], compute_graph.l2_229.out[0], compute_graph.l2_229.out[2], compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.out[0], compute_graph.l2_233.out[0], compute_graph.l2_233.out[2], compute_graph.l2_235.out[0], compute_graph.l2_237.out[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.out[0], compute_graph.L2_IFM_Buffer_for_input3_0_port1.out[0], compute_graph.l2_242.out[0], compute_graph.l2_245.out[0], compute_graph.l2_247.out[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.out[0], compute_graph.l2_252.out[0], compute_graph.l2_252.out[2], compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.out[0], compute_graph.l2_262.out[0], compute_graph.l2_265.out[0], compute_graph.l2_267.out[0], compute_graph.l2_272.out[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.out[0], compute_graph.L2_IFM_Buffer_for_input1_0_port1.out[0], compute_graph.l2_281.out[0], compute_graph.l2_283.out[0], compute_graph.l2_284.out[0], compute_graph.l2_286.out[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.out[0], compute_graph.l2_289.out[0], compute_graph.l2_290.out[0], compute_graph.l2_291.out[0], compute_graph.l2_291.out[2], compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.out[0], compute_graph.l2_296.out[0], compute_graph.l2_297.out[0], compute_graph.L2_OFM_Buffer_layer_299.out[0], compute_graph.l2_301.out[0], compute_graph.L2_OFM_Buffer_layer_302.out[0]) + +Column 0Row 0Channel 1 + +i2_po1(compute_graph.l2_1.out[1], compute_graph.l2_6.out[1], compute_graph.l2_7.out[1], compute_graph.l2_8.out[1], compute_graph.l2_10.out[1], compute_graph.l2_12.out[1], compute_graph.l2_15.out[1], compute_graph.l2_19.out[1], compute_graph.l2_20.out[1], compute_graph.l2_22.out[5], compute_graph.l2_24.out[1], compute_graph.l2_24.out[3], compute_graph.l2_31.out[1], compute_graph.l2_34.out[5], compute_graph.l2_36.out[1], compute_graph.l2_36.out[3], compute_graph.l2_43.out[1], compute_graph.l2_46.out[5], compute_graph.l2_48.out[1], compute_graph.l2_48.out[3], compute_graph.l2_192.out[1], compute_graph.l2_55.out[1], compute_graph.l2_58.out[5], compute_graph.l2_62.out[1], compute_graph.l2_63.out[1], compute_graph.l2_113.out[1], compute_graph.l2_113.out[3], compute_graph.l2_120.out[1], compute_graph.l2_123.out[5], compute_graph.l2_133.out[1], compute_graph.l2_133.out[3], compute_graph.l2_140.out[1], compute_graph.l2_153.out[1], compute_graph.l2_153.out[3], compute_graph.l2_160.out[1], compute_graph.l2_163.out[5], compute_graph.l2_167.out[1], compute_graph.l2_168.out[1], compute_graph.l2_172.out[1], compute_graph.l2_173.out[1], compute_graph.l2_180.out[1], compute_graph.l2_183.out[5], compute_graph.l2_187.out[1], compute_graph.l2_188.out[1], compute_graph.l2_193.out[1], compute_graph.l2_200.out[1], compute_graph.l2_207.out[1], compute_graph.l2_208.out[1], compute_graph.l2_212.out[1], compute_graph.l2_214.out[1], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.out[1], compute_graph.L2_IFM_Buffer_for_input4_0_port1.out[1], compute_graph.l2_219.out[1], compute_graph.l2_222.out[1], compute_graph.l2_224.out[1], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.out[1], compute_graph.l2_229.out[1], compute_graph.l2_229.out[3], compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.out[1], compute_graph.l2_233.out[1], compute_graph.l2_233.out[3], compute_graph.l2_235.out[1], compute_graph.l2_237.out[1], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.out[1], compute_graph.L2_IFM_Buffer_for_input3_0_port1.out[1], compute_graph.l2_242.out[1], compute_graph.l2_245.out[1], compute_graph.l2_247.out[1], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.out[1], compute_graph.l2_252.out[1], compute_graph.l2_252.out[3], compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.out[1], compute_graph.l2_262.out[1], compute_graph.l2_265.out[1], compute_graph.l2_267.out[1], compute_graph.l2_272.out[1], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.out[1], compute_graph.L2_IFM_Buffer_for_input1_0_port1.out[1], compute_graph.l2_281.out[1], compute_graph.l2_283.out[1], compute_graph.l2_284.out[1], compute_graph.l2_286.out[1], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.out[1], compute_graph.l2_290.out[1], compute_graph.l2_291.out[1], compute_graph.l2_291.out[3], compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.out[1], compute_graph.l2_296.out[1], compute_graph.l2_297.out[1], compute_graph.L2_OFM_Buffer_layer_299.out[1], compute_graph.l2_301.out[1], compute_graph.L2_OFM_Buffer_layer_302.out[1]) + +Column 1Row 0Channel 0 + +i3_po0(compute_graph.l2_54.out[0], compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[0], compute_graph.l2_18.out[0], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[0], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[0], compute_graph.l2_22.out[0], compute_graph.l2_23.out[0], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[0], compute_graph.l2_26.out[0], compute_graph.l2_27.out[0], compute_graph.l2_28.out[0], compute_graph.l2_29.out[0], compute_graph.l2_30.out[0], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[0], compute_graph.l2_33.out[0], compute_graph.l2_34.out[0], compute_graph.l2_35.out[0], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[0], compute_graph.l2_38.out[0], compute_graph.l2_39.out[0], compute_graph.l2_40.out[0], compute_graph.l2_41.out[0], compute_graph.l2_42.out[0], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[0], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[0], compute_graph.l2_45.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[0], compute_graph.l2_46.out[0], compute_graph.l2_47.out[0], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[0], compute_graph.l2_50.out[0], compute_graph.l2_51.out[0], compute_graph.l2_52.out[0], compute_graph.l2_53.out[0], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[0], compute_graph.l2_57.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[0], compute_graph.l2_58.out[0], compute_graph.l2_59.out[0], compute_graph.l2_60.out[0], compute_graph.l2_61.out[0], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[0], compute_graph.l2_65.out[0], compute_graph.l2_65.out[4], compute_graph.l2_66.out[0], compute_graph.l2_67.out[0], compute_graph.l2_68.out[0], compute_graph.l2_69.out[0], compute_graph.l2_70.out[0], compute_graph.l2_70.out[4], compute_graph.l2_71.out[0], compute_graph.l2_71.out[4], compute_graph.l2_72.out[0], compute_graph.l2_73.out[0], compute_graph.l2_74.out[0], compute_graph.l2_75.out[0], compute_graph.l2_76.out[0], compute_graph.l2_76.out[4], compute_graph.l2_77.out[0], compute_graph.l2_78.out[0], compute_graph.l2_79.out[0], compute_graph.l2_80.out[0], compute_graph.l2_81.out[0], compute_graph.l2_81.out[4], compute_graph.l2_82.out[0], compute_graph.l2_82.out[4], compute_graph.l2_83.out[0], compute_graph.l2_84.out[0], compute_graph.l2_85.out[0], compute_graph.l2_86.out[0], compute_graph.l2_87.out[0], compute_graph.l2_87.out[4], compute_graph.l2_88.out[0], compute_graph.l2_89.out[0], compute_graph.l2_90.out[0], compute_graph.l2_91.out[0], compute_graph.l2_92.out[0], compute_graph.l2_92.out[4], compute_graph.l2_93.out[0], compute_graph.l2_93.out[4], compute_graph.l2_94.out[0], compute_graph.l2_95.out[0], compute_graph.l2_96.out[0], compute_graph.l2_97.out[0], compute_graph.l2_98.out[0], compute_graph.l2_98.out[4], compute_graph.l2_99.out[0], compute_graph.l2_100.out[0], compute_graph.l2_101.out[0], compute_graph.l2_102.out[0], compute_graph.l2_103.out[0], compute_graph.l2_104.out[0], compute_graph.l2_104.out[4], compute_graph.l2_105.out[0], compute_graph.l2_106.out[0], compute_graph.l2_107.out[0], compute_graph.l2_108.out[0], compute_graph.l2_109.out[0], compute_graph.l2_109.out[4], compute_graph.l2_110.out[0], compute_graph.l2_111.out[0], compute_graph.l2_112.out[0], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[0], compute_graph.l2_115.out[0], compute_graph.l2_116.out[0], compute_graph.l2_117.out[0], compute_graph.l2_118.out[0], compute_graph.l2_119.out[0], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[0], compute_graph.l2_122.out[0], compute_graph.l2_123.out[0], compute_graph.l2_124.out[0], compute_graph.l2_124.out[4], compute_graph.l2_125.out[0], compute_graph.l2_126.out[0], compute_graph.l2_127.out[0], compute_graph.l2_128.out[0], compute_graph.l2_129.out[0], compute_graph.l2_129.out[4], compute_graph.l2_130.out[0], compute_graph.l2_131.out[0], compute_graph.l2_132.out[0], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[0], compute_graph.l2_135.out[0], compute_graph.l2_136.out[0], compute_graph.l2_137.out[0], compute_graph.l2_138.out[0], compute_graph.l2_139.out[0], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[0], compute_graph.l2_142.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[0], compute_graph.l2_143.out[0], compute_graph.l2_144.out[0], compute_graph.l2_144.out[4], compute_graph.l2_145.out[0], compute_graph.l2_146.out[0], compute_graph.l2_147.out[0], compute_graph.l2_148.out[0], compute_graph.l2_149.out[0], compute_graph.l2_149.out[4], compute_graph.l2_150.out[0], compute_graph.l2_151.out[0], compute_graph.l2_152.out[0], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[0], compute_graph.l2_155.out[0], compute_graph.l2_156.out[0], compute_graph.l2_157.out[0], compute_graph.l2_158.out[0], compute_graph.l2_159.out[0], compute_graph.l2_162.out[0], compute_graph.l2_163.out[0], compute_graph.l2_164.out[0], compute_graph.l2_165.out[0], compute_graph.l2_166.out[0], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[0], compute_graph.l2_169.out[0], compute_graph.l2_170.out[0], compute_graph.l2_171.out[0], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[0], compute_graph.l2_175.out[0], compute_graph.l2_176.out[0], compute_graph.l2_177.out[0], compute_graph.l2_178.out[0], compute_graph.l2_179.out[0], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[0], compute_graph.l2_182.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[0], compute_graph.l2_183.out[0], compute_graph.l2_184.out[0], compute_graph.l2_185.out[0], compute_graph.l2_186.out[0], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[0], compute_graph.l2_189.out[0], compute_graph.l2_190.out[0], compute_graph.l2_191.out[0], compute_graph.templated_graph_293.ifm_mem.out[0], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[0], compute_graph.l2_195.out[0], compute_graph.l2_196.out[0], compute_graph.l2_197.out[0], compute_graph.l2_198.out[0], compute_graph.l2_199.out[0], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[0], compute_graph.l2_202.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[0], compute_graph.l2_203.out[0], compute_graph.l2_204.out[0], compute_graph.l2_205.out[0], compute_graph.l2_206.out[0], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[0], compute_graph.l2_210.out[0], compute_graph.l2_211.out[0], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[0], compute_graph.l2_218.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[0], compute_graph.l2_221.out[0], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[0], compute_graph.l2_226.out[0], compute_graph.l2_227.out[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[0], compute_graph.l2_228.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[0], compute_graph.l2_234.out[0], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[0], compute_graph.l2_241.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[0], compute_graph.l2_244.out[0], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[0], compute_graph.l2_249.out[0], compute_graph.l2_250.out[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[0], compute_graph.l2_251.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[0], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[0], compute_graph.l2_261.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[0], compute_graph.l2_264.out[0], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[0], compute_graph.l2_269.out[0], compute_graph.l2_270.out[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[0], compute_graph.l2_271.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[0], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[0], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[0], compute_graph.templated_graph_4.trans_mt_ifm.out[0], compute_graph.templated_graph_231.ifm_mem.out[0], compute_graph.templated_graph_254.ifm_mem.out[0], compute_graph.templated_graph_258.ifm.out[0], compute_graph.templated_graph_259.ifm.out[0], compute_graph.templated_graph_260.ifm.out[0], compute_graph.templated_graph_263.ifm.out[0], compute_graph.templated_graph_266.ifm.out[0], compute_graph.templated_graph_268.ifm.out[0], compute_graph.templated_graph_273.ifm.out[0], compute_graph.templated_graph_274.ifm_mem.out[0], compute_graph.templated_graph_298.ifm.out[0], compute_graph.templated_graph_300.ifm.out[0]) + +Column 1Row 0Channel 1 + +i3_po1(compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[1], compute_graph.l2_54.out[1], compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[1], compute_graph.l2_18.out[1], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[1], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[1], compute_graph.l2_22.out[1], compute_graph.l2_23.out[1], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[1], compute_graph.l2_26.out[1], compute_graph.l2_27.out[1], compute_graph.l2_28.out[1], compute_graph.l2_29.out[1], compute_graph.l2_30.out[1], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[1], compute_graph.l2_33.out[1], compute_graph.l2_34.out[1], compute_graph.l2_35.out[1], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[1], compute_graph.l2_38.out[1], compute_graph.l2_39.out[1], compute_graph.l2_40.out[1], compute_graph.l2_41.out[1], compute_graph.l2_42.out[1], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[1], compute_graph.l2_45.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[1], compute_graph.l2_46.out[1], compute_graph.l2_47.out[1], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[1], compute_graph.l2_50.out[1], compute_graph.l2_51.out[1], compute_graph.l2_52.out[1], compute_graph.l2_53.out[1], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[1], compute_graph.l2_57.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[1], compute_graph.l2_58.out[1], compute_graph.l2_59.out[1], compute_graph.l2_60.out[1], compute_graph.l2_61.out[1], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[1], compute_graph.l2_65.out[1], compute_graph.l2_65.out[5], compute_graph.l2_66.out[1], compute_graph.l2_67.out[1], compute_graph.l2_68.out[1], compute_graph.l2_69.out[1], compute_graph.l2_70.out[1], compute_graph.l2_70.out[5], compute_graph.l2_71.out[1], compute_graph.l2_71.out[5], compute_graph.l2_72.out[1], compute_graph.l2_73.out[1], compute_graph.l2_74.out[1], compute_graph.l2_75.out[1], compute_graph.l2_76.out[1], compute_graph.l2_76.out[5], compute_graph.l2_77.out[1], compute_graph.l2_78.out[1], compute_graph.l2_79.out[1], compute_graph.l2_80.out[1], compute_graph.l2_81.out[1], compute_graph.l2_81.out[5], compute_graph.l2_82.out[1], compute_graph.l2_82.out[5], compute_graph.l2_83.out[1], compute_graph.l2_84.out[1], compute_graph.l2_85.out[1], compute_graph.l2_86.out[1], compute_graph.l2_87.out[1], compute_graph.l2_87.out[5], compute_graph.l2_88.out[1], compute_graph.l2_89.out[1], compute_graph.l2_90.out[1], compute_graph.l2_91.out[1], compute_graph.l2_92.out[1], compute_graph.l2_92.out[5], compute_graph.l2_93.out[1], compute_graph.l2_93.out[5], compute_graph.l2_94.out[1], compute_graph.l2_95.out[1], compute_graph.l2_96.out[1], compute_graph.l2_97.out[1], compute_graph.l2_98.out[1], compute_graph.l2_98.out[5], compute_graph.l2_99.out[1], compute_graph.l2_100.out[1], compute_graph.l2_101.out[1], compute_graph.l2_102.out[1], compute_graph.l2_103.out[1], compute_graph.l2_104.out[1], compute_graph.l2_104.out[5], compute_graph.l2_105.out[1], compute_graph.l2_106.out[1], compute_graph.l2_107.out[1], compute_graph.l2_108.out[1], compute_graph.l2_109.out[1], compute_graph.l2_109.out[5], compute_graph.l2_110.out[1], compute_graph.l2_111.out[1], compute_graph.l2_112.out[1], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[1], compute_graph.l2_115.out[1], compute_graph.l2_116.out[1], compute_graph.l2_117.out[1], compute_graph.l2_118.out[1], compute_graph.l2_119.out[1], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[1], compute_graph.l2_122.out[1], compute_graph.l2_123.out[1], compute_graph.l2_124.out[1], compute_graph.l2_124.out[5], compute_graph.l2_125.out[1], compute_graph.l2_126.out[1], compute_graph.l2_127.out[1], compute_graph.l2_128.out[1], compute_graph.l2_129.out[1], compute_graph.l2_129.out[5], compute_graph.l2_130.out[1], compute_graph.l2_131.out[1], compute_graph.l2_132.out[1], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[1], compute_graph.l2_135.out[1], compute_graph.l2_136.out[1], compute_graph.l2_137.out[1], compute_graph.l2_138.out[1], compute_graph.l2_139.out[1], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[1], compute_graph.l2_142.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[1], compute_graph.l2_143.out[1], compute_graph.l2_144.out[1], compute_graph.l2_144.out[5], compute_graph.l2_145.out[1], compute_graph.l2_146.out[1], compute_graph.l2_147.out[1], compute_graph.l2_148.out[1], compute_graph.l2_149.out[1], compute_graph.l2_149.out[5], compute_graph.l2_150.out[1], compute_graph.l2_151.out[1], compute_graph.l2_152.out[1], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[1], compute_graph.l2_155.out[1], compute_graph.l2_156.out[1], compute_graph.l2_157.out[1], compute_graph.l2_158.out[1], compute_graph.l2_159.out[1], compute_graph.l2_162.out[1], compute_graph.l2_163.out[1], compute_graph.l2_164.out[1], compute_graph.l2_165.out[1], compute_graph.l2_166.out[1], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[1], compute_graph.l2_169.out[1], compute_graph.l2_170.out[1], compute_graph.l2_171.out[1], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[1], compute_graph.l2_175.out[1], compute_graph.l2_176.out[1], compute_graph.l2_177.out[1], compute_graph.l2_178.out[1], compute_graph.l2_179.out[1], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[1], compute_graph.l2_182.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[1], compute_graph.l2_183.out[1], compute_graph.l2_184.out[1], compute_graph.l2_185.out[1], compute_graph.l2_186.out[1], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[1], compute_graph.l2_189.out[1], compute_graph.l2_190.out[1], compute_graph.l2_191.out[1], compute_graph.templated_graph_293.ifm_mem.out[1], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[1], compute_graph.l2_195.out[1], compute_graph.l2_196.out[1], compute_graph.l2_197.out[1], compute_graph.l2_198.out[1], compute_graph.l2_199.out[1], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[1], compute_graph.l2_202.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[1], compute_graph.l2_203.out[1], compute_graph.l2_204.out[1], compute_graph.l2_205.out[1], compute_graph.l2_206.out[1], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[1], compute_graph.l2_210.out[1], compute_graph.l2_211.out[1], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[1], compute_graph.l2_218.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[1], compute_graph.l2_221.out[1], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[1], compute_graph.l2_226.out[1], compute_graph.l2_227.out[1], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[1], compute_graph.l2_228.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[1], compute_graph.l2_234.out[1], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[1], compute_graph.l2_241.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[1], compute_graph.l2_244.out[1], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[1], compute_graph.l2_249.out[1], compute_graph.l2_250.out[1], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[1], compute_graph.l2_251.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[1], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[1], compute_graph.l2_261.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[1], compute_graph.l2_264.out[1], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[1], compute_graph.l2_269.out[1], compute_graph.l2_270.out[1], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[1], compute_graph.l2_271.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[1], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[1], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[1], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[1], compute_graph.templated_graph_4.trans_mt_ifm.out[1], compute_graph.templated_graph_231.ifm_mem.out[1], compute_graph.templated_graph_254.ifm_mem.out[1], compute_graph.templated_graph_258.ifm.out[1], compute_graph.templated_graph_259.ifm.out[1], compute_graph.templated_graph_260.ifm.out[1], compute_graph.templated_graph_263.ifm.out[1], compute_graph.templated_graph_266.ifm.out[1], compute_graph.templated_graph_268.ifm.out[1], compute_graph.templated_graph_273.ifm.out[1], compute_graph.templated_graph_274.ifm_mem.out[1], compute_graph.templated_graph_298.ifm.out[1], compute_graph.templated_graph_300.ifm.out[1]) + +Column 1Row 0Channel 2 + +i3_po2(compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[2], compute_graph.l2_54.out[2], compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[2], compute_graph.l2_18.out[2], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[2], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[2], compute_graph.l2_22.out[2], compute_graph.l2_23.out[2], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[2], compute_graph.l2_26.out[2], compute_graph.l2_27.out[2], compute_graph.l2_28.out[2], compute_graph.l2_29.out[2], compute_graph.l2_30.out[2], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[2], compute_graph.l2_33.out[2], compute_graph.l2_34.out[2], compute_graph.l2_35.out[2], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[2], compute_graph.l2_38.out[2], compute_graph.l2_39.out[2], compute_graph.l2_40.out[2], compute_graph.l2_41.out[2], compute_graph.l2_42.out[2], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[2], compute_graph.l2_45.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[2], compute_graph.l2_46.out[2], compute_graph.l2_47.out[2], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[2], compute_graph.l2_50.out[2], compute_graph.l2_51.out[2], compute_graph.l2_52.out[2], compute_graph.l2_53.out[2], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[2], compute_graph.l2_57.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[2], compute_graph.l2_58.out[2], compute_graph.l2_59.out[2], compute_graph.l2_60.out[2], compute_graph.l2_61.out[2], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[2], compute_graph.l2_65.out[2], compute_graph.l2_65.out[6], compute_graph.l2_66.out[2], compute_graph.l2_67.out[2], compute_graph.l2_68.out[2], compute_graph.l2_69.out[2], compute_graph.l2_70.out[2], compute_graph.l2_70.out[6], compute_graph.l2_71.out[2], compute_graph.l2_71.out[6], compute_graph.l2_72.out[2], compute_graph.l2_73.out[2], compute_graph.l2_74.out[2], compute_graph.l2_75.out[2], compute_graph.l2_76.out[2], compute_graph.l2_76.out[6], compute_graph.l2_77.out[2], compute_graph.l2_78.out[2], compute_graph.l2_79.out[2], compute_graph.l2_80.out[2], compute_graph.l2_81.out[2], compute_graph.l2_81.out[6], compute_graph.l2_82.out[2], compute_graph.l2_82.out[6], compute_graph.l2_83.out[2], compute_graph.l2_84.out[2], compute_graph.l2_85.out[2], compute_graph.l2_86.out[2], compute_graph.l2_87.out[2], compute_graph.l2_87.out[6], compute_graph.l2_88.out[2], compute_graph.l2_89.out[2], compute_graph.l2_90.out[2], compute_graph.l2_91.out[2], compute_graph.l2_92.out[2], compute_graph.l2_92.out[6], compute_graph.l2_93.out[2], compute_graph.l2_93.out[6], compute_graph.l2_94.out[2], compute_graph.l2_95.out[2], compute_graph.l2_96.out[2], compute_graph.l2_97.out[2], compute_graph.l2_98.out[2], compute_graph.l2_98.out[6], compute_graph.l2_99.out[2], compute_graph.l2_100.out[2], compute_graph.l2_101.out[2], compute_graph.l2_102.out[2], compute_graph.l2_103.out[2], compute_graph.l2_104.out[2], compute_graph.l2_104.out[6], compute_graph.l2_105.out[2], compute_graph.l2_106.out[2], compute_graph.l2_107.out[2], compute_graph.l2_108.out[2], compute_graph.l2_109.out[2], compute_graph.l2_109.out[6], compute_graph.l2_110.out[2], compute_graph.l2_111.out[2], compute_graph.l2_112.out[2], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[2], compute_graph.l2_115.out[2], compute_graph.l2_116.out[2], compute_graph.l2_117.out[2], compute_graph.l2_118.out[2], compute_graph.l2_119.out[2], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[2], compute_graph.l2_122.out[2], compute_graph.l2_123.out[2], compute_graph.l2_124.out[2], compute_graph.l2_124.out[6], compute_graph.l2_125.out[2], compute_graph.l2_126.out[2], compute_graph.l2_127.out[2], compute_graph.l2_128.out[2], compute_graph.l2_129.out[2], compute_graph.l2_129.out[6], compute_graph.l2_130.out[2], compute_graph.l2_131.out[2], compute_graph.l2_132.out[2], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[2], compute_graph.l2_135.out[2], compute_graph.l2_136.out[2], compute_graph.l2_137.out[2], compute_graph.l2_138.out[2], compute_graph.l2_139.out[2], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[2], compute_graph.l2_142.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[2], compute_graph.l2_143.out[2], compute_graph.l2_144.out[2], compute_graph.l2_144.out[6], compute_graph.l2_145.out[2], compute_graph.l2_146.out[2], compute_graph.l2_147.out[2], compute_graph.l2_148.out[2], compute_graph.l2_149.out[2], compute_graph.l2_149.out[6], compute_graph.l2_150.out[2], compute_graph.l2_151.out[2], compute_graph.l2_152.out[2], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[2], compute_graph.l2_155.out[2], compute_graph.l2_156.out[2], compute_graph.l2_157.out[2], compute_graph.l2_158.out[2], compute_graph.l2_159.out[2], compute_graph.l2_162.out[2], compute_graph.l2_163.out[2], compute_graph.l2_164.out[2], compute_graph.l2_165.out[2], compute_graph.l2_166.out[2], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[2], compute_graph.l2_169.out[2], compute_graph.l2_170.out[2], compute_graph.l2_171.out[2], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[2], compute_graph.l2_175.out[2], compute_graph.l2_176.out[2], compute_graph.l2_177.out[2], compute_graph.l2_178.out[2], compute_graph.l2_179.out[2], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[2], compute_graph.l2_182.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[2], compute_graph.l2_183.out[2], compute_graph.l2_184.out[2], compute_graph.l2_185.out[2], compute_graph.l2_186.out[2], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[2], compute_graph.l2_189.out[2], compute_graph.l2_190.out[2], compute_graph.l2_191.out[2], compute_graph.templated_graph_293.ifm_mem.out[2], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[2], compute_graph.l2_195.out[2], compute_graph.l2_196.out[2], compute_graph.l2_197.out[2], compute_graph.l2_198.out[2], compute_graph.l2_199.out[2], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[2], compute_graph.l2_202.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[2], compute_graph.l2_203.out[2], compute_graph.l2_204.out[2], compute_graph.l2_205.out[2], compute_graph.l2_206.out[2], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[2], compute_graph.l2_210.out[2], compute_graph.l2_211.out[2], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[2], compute_graph.l2_218.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[2], compute_graph.l2_221.out[2], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[2], compute_graph.l2_226.out[2], compute_graph.l2_227.out[2], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[2], compute_graph.l2_228.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[2], compute_graph.l2_234.out[2], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[2], compute_graph.l2_241.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[2], compute_graph.l2_244.out[2], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[2], compute_graph.l2_249.out[2], compute_graph.l2_250.out[2], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[2], compute_graph.l2_251.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[2], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[2], compute_graph.l2_261.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[2], compute_graph.l2_264.out[2], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[2], compute_graph.l2_269.out[2], compute_graph.l2_270.out[2], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[2], compute_graph.l2_271.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[2], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[2], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[2], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[2], compute_graph.templated_graph_4.trans_mt_ifm.out[2], compute_graph.templated_graph_231.ifm_mem.out[2], compute_graph.templated_graph_254.ifm_mem.out[2], compute_graph.templated_graph_258.ifm.out[2], compute_graph.templated_graph_259.ifm.out[2], compute_graph.templated_graph_260.ifm.out[2], compute_graph.templated_graph_263.ifm.out[2], compute_graph.templated_graph_266.ifm.out[2], compute_graph.templated_graph_268.ifm.out[2], compute_graph.templated_graph_273.ifm.out[2], compute_graph.templated_graph_274.ifm_mem.out[2], compute_graph.templated_graph_298.ifm.out[2], compute_graph.templated_graph_300.ifm.out[2]) + +Column 1Row 0Channel 3 + +i3_po3(compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[3], compute_graph.l2_54.out[3], compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[3], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[3], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[3], compute_graph.l2_18.out[3], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[3], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[3], compute_graph.l2_22.out[3], compute_graph.l2_23.out[3], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[3], compute_graph.l2_26.out[3], compute_graph.l2_27.out[3], compute_graph.l2_28.out[3], compute_graph.l2_29.out[3], compute_graph.l2_30.out[3], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[3], compute_graph.l2_33.out[3], compute_graph.l2_34.out[3], compute_graph.l2_35.out[3], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[3], compute_graph.l2_38.out[3], compute_graph.l2_39.out[3], compute_graph.l2_40.out[3], compute_graph.l2_41.out[3], compute_graph.l2_42.out[3], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[3], compute_graph.l2_45.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[3], compute_graph.l2_46.out[3], compute_graph.l2_47.out[3], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[3], compute_graph.l2_50.out[3], compute_graph.l2_51.out[3], compute_graph.l2_52.out[3], compute_graph.l2_53.out[3], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[3], compute_graph.l2_57.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[3], compute_graph.l2_58.out[3], compute_graph.l2_59.out[3], compute_graph.l2_60.out[3], compute_graph.l2_61.out[3], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[3], compute_graph.l2_65.out[3], compute_graph.l2_65.out[7], compute_graph.l2_66.out[3], compute_graph.l2_67.out[3], compute_graph.l2_68.out[3], compute_graph.l2_69.out[3], compute_graph.l2_70.out[3], compute_graph.l2_70.out[7], compute_graph.l2_71.out[3], compute_graph.l2_71.out[7], compute_graph.l2_72.out[3], compute_graph.l2_73.out[3], compute_graph.l2_74.out[3], compute_graph.l2_75.out[3], compute_graph.l2_76.out[3], compute_graph.l2_76.out[7], compute_graph.l2_77.out[3], compute_graph.l2_78.out[3], compute_graph.l2_79.out[3], compute_graph.l2_80.out[3], compute_graph.l2_81.out[3], compute_graph.l2_81.out[7], compute_graph.l2_82.out[3], compute_graph.l2_82.out[7], compute_graph.l2_83.out[3], compute_graph.l2_84.out[3], compute_graph.l2_85.out[3], compute_graph.l2_86.out[3], compute_graph.l2_87.out[3], compute_graph.l2_87.out[7], compute_graph.l2_88.out[3], compute_graph.l2_89.out[3], compute_graph.l2_90.out[3], compute_graph.l2_91.out[3], compute_graph.l2_92.out[3], compute_graph.l2_92.out[7], compute_graph.l2_93.out[3], compute_graph.l2_93.out[7], compute_graph.l2_94.out[3], compute_graph.l2_95.out[3], compute_graph.l2_96.out[3], compute_graph.l2_97.out[3], compute_graph.l2_98.out[3], compute_graph.l2_98.out[7], compute_graph.l2_99.out[3], compute_graph.l2_100.out[3], compute_graph.l2_101.out[3], compute_graph.l2_102.out[3], compute_graph.l2_103.out[3], compute_graph.l2_104.out[3], compute_graph.l2_104.out[7], compute_graph.l2_105.out[3], compute_graph.l2_106.out[3], compute_graph.l2_107.out[3], compute_graph.l2_108.out[3], compute_graph.l2_109.out[3], compute_graph.l2_109.out[7], compute_graph.l2_110.out[3], compute_graph.l2_111.out[3], compute_graph.l2_112.out[3], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[3], compute_graph.l2_115.out[3], compute_graph.l2_116.out[3], compute_graph.l2_117.out[3], compute_graph.l2_118.out[3], compute_graph.l2_119.out[3], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[3], compute_graph.l2_122.out[3], compute_graph.l2_123.out[3], compute_graph.l2_124.out[3], compute_graph.l2_124.out[7], compute_graph.l2_125.out[3], compute_graph.l2_126.out[3], compute_graph.l2_127.out[3], compute_graph.l2_128.out[3], compute_graph.l2_129.out[3], compute_graph.l2_129.out[7], compute_graph.l2_130.out[3], compute_graph.l2_131.out[3], compute_graph.l2_132.out[3], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[3], compute_graph.l2_135.out[3], compute_graph.l2_136.out[3], compute_graph.l2_137.out[3], compute_graph.l2_138.out[3], compute_graph.l2_139.out[3], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[3], compute_graph.l2_142.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[3], compute_graph.l2_143.out[3], compute_graph.l2_144.out[3], compute_graph.l2_144.out[7], compute_graph.l2_145.out[3], compute_graph.l2_146.out[3], compute_graph.l2_147.out[3], compute_graph.l2_148.out[3], compute_graph.l2_149.out[3], compute_graph.l2_149.out[7], compute_graph.l2_150.out[3], compute_graph.l2_151.out[3], compute_graph.l2_152.out[3], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[3], compute_graph.l2_155.out[3], compute_graph.l2_156.out[3], compute_graph.l2_157.out[3], compute_graph.l2_158.out[3], compute_graph.l2_159.out[3], compute_graph.l2_162.out[3], compute_graph.l2_163.out[3], compute_graph.l2_164.out[3], compute_graph.l2_165.out[3], compute_graph.l2_166.out[3], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[3], compute_graph.l2_169.out[3], compute_graph.l2_170.out[3], compute_graph.l2_171.out[3], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[3], compute_graph.l2_175.out[3], compute_graph.l2_176.out[3], compute_graph.l2_177.out[3], compute_graph.l2_178.out[3], compute_graph.l2_179.out[3], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[3], compute_graph.l2_182.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[3], compute_graph.l2_183.out[3], compute_graph.l2_184.out[3], compute_graph.l2_185.out[3], compute_graph.l2_186.out[3], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[3], compute_graph.l2_189.out[3], compute_graph.l2_190.out[3], compute_graph.l2_191.out[3], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[3], compute_graph.l2_195.out[3], compute_graph.l2_196.out[3], compute_graph.l2_197.out[3], compute_graph.l2_198.out[3], compute_graph.l2_199.out[3], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[3], compute_graph.l2_202.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[3], compute_graph.l2_203.out[3], compute_graph.l2_204.out[3], compute_graph.l2_205.out[3], compute_graph.l2_206.out[3], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[3], compute_graph.l2_210.out[3], compute_graph.l2_211.out[3], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[3], compute_graph.l2_218.out[3], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[3], compute_graph.l2_221.out[3], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[3], compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[3], compute_graph.l2_226.out[3], compute_graph.l2_227.out[3], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[3], compute_graph.l2_228.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[3], compute_graph.l2_234.out[3], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[3], compute_graph.l2_241.out[3], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[3], compute_graph.l2_244.out[3], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[3], compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[3], compute_graph.l2_249.out[3], compute_graph.l2_250.out[3], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[3], compute_graph.l2_251.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[3], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[3], compute_graph.l2_261.out[3], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[3], compute_graph.l2_264.out[3], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[3], compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[3], compute_graph.l2_269.out[3], compute_graph.l2_270.out[3], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[3], compute_graph.l2_271.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[3], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[3], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[3], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[3], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[3], compute_graph.templated_graph_4.trans_mt_ifm.out[3], compute_graph.templated_graph_231.ifm_mem.out[3], compute_graph.templated_graph_254.ifm_mem.out[3], compute_graph.templated_graph_258.ifm.out[3], compute_graph.templated_graph_259.ifm.out[3], compute_graph.templated_graph_260.ifm.out[3], compute_graph.templated_graph_263.ifm.out[3], compute_graph.templated_graph_266.ifm.out[3], compute_graph.templated_graph_268.ifm.out[3], compute_graph.templated_graph_273.ifm.out[3], compute_graph.templated_graph_274.ifm_mem.out[3], compute_graph.templated_graph_293.ifm_mem.out[3], compute_graph.templated_graph_298.ifm.out[3], compute_graph.templated_graph_300.ifm.out[3]) + +Column 1Row 0Channel 0 + +i3_pi0(compute_graph.l2_45.in[0], compute_graph.l2_55.in[0], compute_graph.l2_1.in[0], compute_graph.l2_6.in[0], compute_graph.l2_7.in[0], compute_graph.l2_8.in[0], compute_graph.l2_9.in[0], compute_graph.l2_10.in[0], compute_graph.l2_11.in[0], compute_graph.l2_12.in[0], compute_graph.l2_13.in[0], compute_graph.l2_14.in[0], compute_graph.l2_15.in[0], compute_graph.l2_16.in[0], compute_graph.l2_18.in[0], compute_graph.l2_19.in[0], compute_graph.l2_20.in[0], compute_graph.l2_22.in[0], compute_graph.l2_23.in[0], compute_graph.l2_24.in[0], compute_graph.l2_26.in[0], compute_graph.l2_27.in[0], compute_graph.l2_28.in[0], compute_graph.l2_29.in[0], compute_graph.l2_30.in[0], compute_graph.l2_31.in[0], compute_graph.l2_33.in[0], compute_graph.l2_34.in[0], compute_graph.l2_35.in[0], compute_graph.l2_36.in[0], compute_graph.l2_38.in[0], compute_graph.l2_39.in[0], compute_graph.l2_40.in[0], compute_graph.l2_41.in[0], compute_graph.l2_42.in[0], compute_graph.l2_43.in[0], compute_graph.l2_162.in[0], compute_graph.l2_46.in[0], compute_graph.l2_47.in[0], compute_graph.l2_48.in[0], compute_graph.l2_50.in[0], compute_graph.l2_51.in[0], compute_graph.l2_52.in[0], compute_graph.l2_53.in[0], compute_graph.l2_54.in[0], compute_graph.l2_193.in[0], compute_graph.l2_57.in[0], compute_graph.l2_58.in[0], compute_graph.l2_59.in[0], compute_graph.l2_60.in[0], compute_graph.l2_61.in[0], compute_graph.l2_62.in[0], compute_graph.l2_63.in[0], compute_graph.l2_65.in[0], compute_graph.l2_66.in[0], compute_graph.l2_67.in[0], compute_graph.l2_68.in[0], compute_graph.l2_69.in[0], compute_graph.l2_70.in[0], compute_graph.l2_71.in[0], compute_graph.l2_72.in[0], compute_graph.l2_73.in[0], compute_graph.l2_74.in[0], compute_graph.l2_75.in[0], compute_graph.l2_76.in[0], compute_graph.l2_77.in[0], compute_graph.l2_78.in[0], compute_graph.l2_79.in[0], compute_graph.l2_80.in[0], compute_graph.l2_81.in[0], compute_graph.l2_82.in[0], compute_graph.l2_83.in[0], compute_graph.l2_84.in[0], compute_graph.l2_85.in[0], compute_graph.l2_86.in[0], compute_graph.l2_87.in[0], compute_graph.l2_88.in[0], compute_graph.l2_89.in[0], compute_graph.l2_90.in[0], compute_graph.l2_91.in[0], compute_graph.l2_92.in[0], compute_graph.l2_93.in[0], compute_graph.l2_94.in[0], compute_graph.l2_95.in[0], compute_graph.l2_96.in[0], compute_graph.l2_97.in[0], compute_graph.l2_98.in[0], compute_graph.l2_99.in[0], compute_graph.l2_100.in[0], compute_graph.l2_101.in[0], compute_graph.l2_102.in[0], compute_graph.l2_103.in[0], compute_graph.l2_104.in[0], compute_graph.l2_105.in[0], compute_graph.l2_106.in[0], compute_graph.l2_107.in[0], compute_graph.l2_108.in[0], compute_graph.l2_109.in[0], compute_graph.l2_110.in[0], compute_graph.l2_111.in[0], compute_graph.l2_112.in[0], compute_graph.l2_113.in[0], compute_graph.l2_115.in[0], compute_graph.l2_116.in[0], compute_graph.l2_117.in[0], compute_graph.l2_118.in[0], compute_graph.l2_119.in[0], compute_graph.l2_120.in[0], compute_graph.l2_122.in[0], compute_graph.l2_123.in[0], compute_graph.l2_124.in[0], compute_graph.l2_125.in[0], compute_graph.l2_126.in[0], compute_graph.l2_127.in[0], compute_graph.l2_128.in[0], compute_graph.l2_129.in[0], compute_graph.l2_130.in[0], compute_graph.l2_131.in[0], compute_graph.l2_132.in[0], compute_graph.l2_133.in[0], compute_graph.l2_135.in[0], compute_graph.l2_136.in[0], compute_graph.l2_137.in[0], compute_graph.l2_138.in[0], compute_graph.l2_139.in[0], compute_graph.l2_140.in[0], compute_graph.l2_142.in[0], compute_graph.l2_143.in[0], compute_graph.l2_144.in[0], compute_graph.l2_145.in[0], compute_graph.l2_146.in[0], compute_graph.l2_147.in[0], compute_graph.l2_148.in[0], compute_graph.l2_149.in[0], compute_graph.l2_150.in[0], compute_graph.l2_151.in[0], compute_graph.l2_152.in[0], compute_graph.l2_153.in[0], compute_graph.l2_155.in[0], compute_graph.l2_156.in[0], compute_graph.l2_157.in[0], compute_graph.l2_158.in[0], compute_graph.l2_159.in[0], compute_graph.l2_160.in[0], compute_graph.l2_163.in[0], compute_graph.l2_164.in[0], compute_graph.l2_165.in[0], compute_graph.l2_166.in[0], compute_graph.l2_167.in[0], compute_graph.l2_168.in[0], compute_graph.l2_169.in[0], compute_graph.l2_170.in[0], compute_graph.l2_171.in[0], compute_graph.l2_172.in[0], compute_graph.l2_173.in[0], compute_graph.l2_175.in[0], compute_graph.l2_176.in[0], compute_graph.l2_177.in[0], compute_graph.l2_178.in[0], compute_graph.l2_179.in[0], compute_graph.l2_180.in[0], compute_graph.l2_182.in[0], compute_graph.l2_183.in[0], compute_graph.l2_184.in[0], compute_graph.l2_185.in[0], compute_graph.l2_186.in[0], compute_graph.l2_187.in[0], compute_graph.l2_188.in[0], compute_graph.l2_189.in[0], compute_graph.l2_190.in[0], compute_graph.l2_191.in[0], compute_graph.l2_192.in[0], compute_graph.l2_195.in[0], compute_graph.l2_196.in[0], compute_graph.l2_197.in[0], compute_graph.l2_198.in[0], compute_graph.l2_199.in[0], compute_graph.l2_200.in[0], compute_graph.l2_202.in[0], compute_graph.l2_203.in[0], compute_graph.l2_204.in[0], compute_graph.l2_205.in[0], compute_graph.l2_206.in[0], compute_graph.l2_207.in[0], compute_graph.l2_208.in[0], compute_graph.l2_210.in[0], compute_graph.l2_211.in[0], compute_graph.l2_212.in[0], compute_graph.l2_214.in[0], compute_graph.l2_218.in[0], compute_graph.l2_219.in[0], compute_graph.l2_221.in[0], compute_graph.l2_222.in[0], compute_graph.l2_224.in[0], compute_graph.l2_226.in[0], compute_graph.l2_227.in[0], compute_graph.l2_228.in[0], compute_graph.l2_229.in[0], compute_graph.l2_233.in[0], compute_graph.l2_234.in[0], compute_graph.l2_235.in[0], compute_graph.l2_237.in[0], compute_graph.l2_241.in[0], compute_graph.l2_242.in[0], compute_graph.l2_244.in[0], compute_graph.l2_245.in[0], compute_graph.l2_247.in[0], compute_graph.l2_249.in[0], compute_graph.l2_250.in[0], compute_graph.l2_251.in[0], compute_graph.l2_252.in[0], compute_graph.l2_257.in[0], compute_graph.l2_261.in[0], compute_graph.l2_262.in[0], compute_graph.l2_264.in[0], compute_graph.l2_265.in[0], compute_graph.l2_267.in[0], compute_graph.l2_269.in[0], compute_graph.l2_270.in[0], compute_graph.l2_271.in[0], compute_graph.l2_272.in[0], compute_graph.l2_276.in[0], compute_graph.l2_280.in[0], compute_graph.l2_281.in[0], compute_graph.l2_283.in[0], compute_graph.l2_284.in[0], compute_graph.l2_286.in[0], compute_graph.l2_288.in[0], compute_graph.l2_289.in[0], compute_graph.l2_290.in[0], compute_graph.l2_291.in[0], compute_graph.l2_295.in[0], compute_graph.l2_296.in[0], compute_graph.l2_297.in[0], compute_graph.L2_OFM_Buffer_layer_299.in[0], compute_graph.l2_301.in[0], compute_graph.L2_OFM_Buffer_layer_302.in[0], compute_graph.templated_graph_4.trans_mt_ofm.in[0], compute_graph.templated_graph_231.ofm_mem.in[0], compute_graph.templated_graph_254.ofm_mem.in[0], compute_graph.templated_graph_258.ofm.in[0], compute_graph.templated_graph_259.ofm.in[0], compute_graph.templated_graph_260.ofm.in[0], compute_graph.templated_graph_263.ofm.in[0], compute_graph.templated_graph_266.ofm.in[0], compute_graph.templated_graph_268.ofm.in[0], compute_graph.templated_graph_273.ofm.in[0], compute_graph.templated_graph_274.ofm_mem.in[0], compute_graph.templated_graph_293.ofm_mem.in[0], compute_graph.templated_graph_298.ofm.in[0], compute_graph.templated_graph_300.ofm.in[0]) + +Column 1Row 0Channel 1 + +i3_pi1(compute_graph.l2_45.in[1], compute_graph.l2_55.in[1], compute_graph.l2_1.in[1], compute_graph.l2_6.in[1], compute_graph.l2_7.in[1], compute_graph.l2_8.in[1], compute_graph.l2_9.in[1], compute_graph.l2_10.in[1], compute_graph.l2_11.in[1], compute_graph.l2_12.in[1], compute_graph.l2_13.in[1], compute_graph.l2_14.in[1], compute_graph.l2_15.in[1], compute_graph.l2_16.in[1], compute_graph.l2_18.in[1], compute_graph.l2_19.in[1], compute_graph.l2_20.in[1], compute_graph.l2_22.in[1], compute_graph.l2_23.in[1], compute_graph.l2_24.in[1], compute_graph.l2_26.in[1], compute_graph.l2_27.in[1], compute_graph.l2_28.in[1], compute_graph.l2_29.in[1], compute_graph.l2_30.in[1], compute_graph.l2_31.in[1], compute_graph.l2_33.in[1], compute_graph.l2_34.in[1], compute_graph.l2_35.in[1], compute_graph.l2_36.in[1], compute_graph.l2_38.in[1], compute_graph.l2_39.in[1], compute_graph.l2_40.in[1], compute_graph.l2_41.in[1], compute_graph.l2_42.in[1], compute_graph.l2_43.in[1], compute_graph.l2_162.in[1], compute_graph.l2_46.in[1], compute_graph.l2_47.in[1], compute_graph.l2_48.in[1], compute_graph.l2_50.in[1], compute_graph.l2_51.in[1], compute_graph.l2_52.in[1], compute_graph.l2_53.in[1], compute_graph.l2_54.in[1], compute_graph.l2_57.in[1], compute_graph.l2_58.in[1], compute_graph.l2_59.in[1], compute_graph.l2_60.in[1], compute_graph.l2_61.in[1], compute_graph.l2_62.in[1], compute_graph.l2_63.in[1], compute_graph.l2_65.in[1], compute_graph.l2_66.in[1], compute_graph.l2_67.in[1], compute_graph.l2_68.in[1], compute_graph.l2_69.in[1], compute_graph.l2_70.in[1], compute_graph.l2_71.in[1], compute_graph.l2_72.in[1], compute_graph.l2_73.in[1], compute_graph.l2_74.in[1], compute_graph.l2_75.in[1], compute_graph.l2_76.in[1], compute_graph.l2_77.in[1], compute_graph.l2_78.in[1], compute_graph.l2_79.in[1], compute_graph.l2_80.in[1], compute_graph.l2_81.in[1], compute_graph.l2_82.in[1], compute_graph.l2_83.in[1], compute_graph.l2_84.in[1], compute_graph.l2_85.in[1], compute_graph.l2_86.in[1], compute_graph.l2_87.in[1], compute_graph.l2_88.in[1], compute_graph.l2_89.in[1], compute_graph.l2_90.in[1], compute_graph.l2_91.in[1], compute_graph.l2_92.in[1], compute_graph.l2_93.in[1], compute_graph.l2_94.in[1], compute_graph.l2_95.in[1], compute_graph.l2_96.in[1], compute_graph.l2_97.in[1], compute_graph.l2_98.in[1], compute_graph.l2_99.in[1], compute_graph.l2_100.in[1], compute_graph.l2_101.in[1], compute_graph.l2_102.in[1], compute_graph.l2_103.in[1], compute_graph.l2_104.in[1], compute_graph.l2_105.in[1], compute_graph.l2_106.in[1], compute_graph.l2_107.in[1], compute_graph.l2_108.in[1], compute_graph.l2_109.in[1], compute_graph.l2_110.in[1], compute_graph.l2_111.in[1], compute_graph.l2_112.in[1], compute_graph.l2_113.in[1], compute_graph.l2_115.in[1], compute_graph.l2_116.in[1], compute_graph.l2_117.in[1], compute_graph.l2_118.in[1], compute_graph.l2_119.in[1], compute_graph.l2_120.in[1], compute_graph.l2_122.in[1], compute_graph.l2_123.in[1], compute_graph.l2_124.in[1], compute_graph.l2_125.in[1], compute_graph.l2_126.in[1], compute_graph.l2_127.in[1], compute_graph.l2_128.in[1], compute_graph.l2_129.in[1], compute_graph.l2_130.in[1], compute_graph.l2_131.in[1], compute_graph.l2_132.in[1], compute_graph.l2_133.in[1], compute_graph.l2_135.in[1], compute_graph.l2_136.in[1], compute_graph.l2_137.in[1], compute_graph.l2_138.in[1], compute_graph.l2_139.in[1], compute_graph.l2_140.in[1], compute_graph.l2_142.in[1], compute_graph.l2_143.in[1], compute_graph.l2_144.in[1], compute_graph.l2_145.in[1], compute_graph.l2_146.in[1], compute_graph.l2_147.in[1], compute_graph.l2_148.in[1], compute_graph.l2_149.in[1], compute_graph.l2_150.in[1], compute_graph.l2_151.in[1], compute_graph.l2_152.in[1], compute_graph.l2_153.in[1], compute_graph.l2_155.in[1], compute_graph.l2_156.in[1], compute_graph.l2_157.in[1], compute_graph.l2_158.in[1], compute_graph.l2_159.in[1], compute_graph.l2_160.in[1], compute_graph.l2_163.in[1], compute_graph.l2_164.in[1], compute_graph.l2_165.in[1], compute_graph.l2_166.in[1], compute_graph.l2_167.in[1], compute_graph.l2_168.in[1], compute_graph.l2_169.in[1], compute_graph.l2_170.in[1], compute_graph.l2_171.in[1], compute_graph.l2_172.in[1], compute_graph.l2_173.in[1], compute_graph.l2_175.in[1], compute_graph.l2_176.in[1], compute_graph.l2_177.in[1], compute_graph.l2_178.in[1], compute_graph.l2_179.in[1], compute_graph.l2_180.in[1], compute_graph.l2_182.in[1], compute_graph.l2_183.in[1], compute_graph.l2_184.in[1], compute_graph.l2_185.in[1], compute_graph.l2_186.in[1], compute_graph.l2_187.in[1], compute_graph.l2_188.in[1], compute_graph.l2_189.in[1], compute_graph.l2_190.in[1], compute_graph.l2_191.in[1], compute_graph.l2_192.in[1], compute_graph.l2_193.in[1], compute_graph.l2_195.in[1], compute_graph.l2_196.in[1], compute_graph.l2_197.in[1], compute_graph.l2_198.in[1], compute_graph.l2_199.in[1], compute_graph.l2_200.in[1], compute_graph.l2_202.in[1], compute_graph.l2_203.in[1], compute_graph.l2_204.in[1], compute_graph.l2_205.in[1], compute_graph.l2_206.in[1], compute_graph.l2_207.in[1], compute_graph.l2_208.in[1], compute_graph.l2_210.in[1], compute_graph.l2_211.in[1], compute_graph.l2_212.in[1], compute_graph.l2_214.in[1], compute_graph.l2_218.in[1], compute_graph.l2_219.in[1], compute_graph.l2_221.in[1], compute_graph.l2_222.in[1], compute_graph.l2_224.in[1], compute_graph.l2_226.in[1], compute_graph.l2_227.in[1], compute_graph.l2_228.in[1], compute_graph.l2_229.in[1], compute_graph.l2_233.in[1], compute_graph.l2_234.in[1], compute_graph.l2_235.in[1], compute_graph.l2_237.in[1], compute_graph.l2_241.in[1], compute_graph.l2_242.in[1], compute_graph.l2_244.in[1], compute_graph.l2_245.in[1], compute_graph.l2_247.in[1], compute_graph.l2_249.in[1], compute_graph.l2_250.in[1], compute_graph.l2_251.in[1], compute_graph.l2_252.in[1], compute_graph.l2_257.in[1], compute_graph.l2_261.in[1], compute_graph.l2_262.in[1], compute_graph.l2_264.in[1], compute_graph.l2_265.in[1], compute_graph.l2_267.in[1], compute_graph.l2_269.in[1], compute_graph.l2_270.in[1], compute_graph.l2_271.in[1], compute_graph.l2_272.in[1], compute_graph.l2_276.in[1], compute_graph.l2_280.in[1], compute_graph.l2_281.in[1], compute_graph.l2_283.in[1], compute_graph.l2_284.in[1], compute_graph.l2_286.in[1], compute_graph.l2_288.in[1], compute_graph.l2_289.in[1], compute_graph.l2_290.in[1], compute_graph.l2_291.in[1], compute_graph.l2_295.in[1], compute_graph.l2_296.in[1], compute_graph.l2_297.in[1], compute_graph.L2_OFM_Buffer_layer_299.in[1], compute_graph.l2_301.in[1], compute_graph.L2_OFM_Buffer_layer_302.in[1], compute_graph.templated_graph_4.trans_mt_ofm.in[1], compute_graph.templated_graph_231.ofm_mem.in[1], compute_graph.templated_graph_254.ofm_mem.in[1], compute_graph.templated_graph_258.ofm.in[1], compute_graph.templated_graph_259.ofm.in[1], compute_graph.templated_graph_260.ofm.in[1], compute_graph.templated_graph_263.ofm.in[1], compute_graph.templated_graph_266.ofm.in[1], compute_graph.templated_graph_268.ofm.in[1], compute_graph.templated_graph_273.ofm.in[1], compute_graph.templated_graph_274.ofm_mem.in[1], compute_graph.templated_graph_293.ofm_mem.in[1], compute_graph.templated_graph_298.ofm.in[1], compute_graph.templated_graph_300.ofm.in[1]) + +Column 1Row 0Channel 2 + +i3_pi2(compute_graph.l2_45.in[2], compute_graph.l2_1.in[2], compute_graph.l2_6.in[2], compute_graph.l2_7.in[2], compute_graph.l2_8.in[2], compute_graph.l2_9.in[2], compute_graph.l2_10.in[2], compute_graph.l2_11.in[2], compute_graph.l2_12.in[2], compute_graph.l2_13.in[2], compute_graph.l2_14.in[2], compute_graph.l2_15.in[2], compute_graph.l2_16.in[2], compute_graph.l2_18.in[2], compute_graph.l2_19.in[2], compute_graph.l2_20.in[2], compute_graph.l2_22.in[2], compute_graph.l2_23.in[2], compute_graph.l2_24.in[2], compute_graph.l2_26.in[2], compute_graph.l2_27.in[2], compute_graph.l2_28.in[2], compute_graph.l2_29.in[2], compute_graph.l2_30.in[2], compute_graph.l2_31.in[2], compute_graph.l2_33.in[2], compute_graph.l2_34.in[2], compute_graph.l2_35.in[2], compute_graph.l2_36.in[2], compute_graph.l2_38.in[2], compute_graph.l2_39.in[2], compute_graph.l2_40.in[2], compute_graph.l2_41.in[2], compute_graph.l2_42.in[2], compute_graph.l2_43.in[2], compute_graph.l2_46.in[2], compute_graph.l2_47.in[2], compute_graph.l2_48.in[2], compute_graph.l2_50.in[2], compute_graph.l2_51.in[2], compute_graph.l2_52.in[2], compute_graph.l2_53.in[2], compute_graph.l2_54.in[2], compute_graph.l2_55.in[2], compute_graph.l2_57.in[2], compute_graph.l2_58.in[2], compute_graph.l2_59.in[2], compute_graph.l2_60.in[2], compute_graph.l2_61.in[2], compute_graph.l2_62.in[2], compute_graph.l2_63.in[2], compute_graph.l2_65.in[2], compute_graph.l2_66.in[2], compute_graph.l2_67.in[2], compute_graph.l2_68.in[2], compute_graph.l2_69.in[2], compute_graph.l2_70.in[2], compute_graph.l2_71.in[2], compute_graph.l2_72.in[2], compute_graph.l2_73.in[2], compute_graph.l2_74.in[2], compute_graph.l2_75.in[2], compute_graph.l2_76.in[2], compute_graph.l2_77.in[2], compute_graph.l2_78.in[2], compute_graph.l2_79.in[2], compute_graph.l2_80.in[2], compute_graph.l2_81.in[2], compute_graph.l2_82.in[2], compute_graph.l2_83.in[2], compute_graph.l2_84.in[2], compute_graph.l2_85.in[2], compute_graph.l2_86.in[2], compute_graph.l2_87.in[2], compute_graph.l2_88.in[2], compute_graph.l2_89.in[2], compute_graph.l2_90.in[2], compute_graph.l2_91.in[2], compute_graph.l2_92.in[2], compute_graph.l2_93.in[2], compute_graph.l2_94.in[2], compute_graph.l2_95.in[2], compute_graph.l2_96.in[2], compute_graph.l2_97.in[2], compute_graph.l2_98.in[2], compute_graph.l2_99.in[2], compute_graph.l2_100.in[2], compute_graph.l2_101.in[2], compute_graph.l2_102.in[2], compute_graph.l2_103.in[2], compute_graph.l2_104.in[2], compute_graph.l2_105.in[2], compute_graph.l2_106.in[2], compute_graph.l2_107.in[2], compute_graph.l2_108.in[2], compute_graph.l2_109.in[2], compute_graph.l2_110.in[2], compute_graph.l2_111.in[2], compute_graph.l2_112.in[2], compute_graph.l2_113.in[2], compute_graph.l2_115.in[2], compute_graph.l2_116.in[2], compute_graph.l2_117.in[2], compute_graph.l2_118.in[2], compute_graph.l2_119.in[2], compute_graph.l2_120.in[2], compute_graph.l2_122.in[2], compute_graph.l2_123.in[2], compute_graph.l2_124.in[2], compute_graph.l2_125.in[2], compute_graph.l2_126.in[2], compute_graph.l2_127.in[2], compute_graph.l2_128.in[2], compute_graph.l2_129.in[2], compute_graph.l2_130.in[2], compute_graph.l2_131.in[2], compute_graph.l2_132.in[2], compute_graph.l2_133.in[2], compute_graph.l2_135.in[2], compute_graph.l2_136.in[2], compute_graph.l2_137.in[2], compute_graph.l2_138.in[2], compute_graph.l2_139.in[2], compute_graph.l2_140.in[2], compute_graph.l2_142.in[2], compute_graph.l2_143.in[2], compute_graph.l2_144.in[2], compute_graph.l2_145.in[2], compute_graph.l2_146.in[2], compute_graph.l2_147.in[2], compute_graph.l2_148.in[2], compute_graph.l2_149.in[2], compute_graph.l2_150.in[2], compute_graph.l2_151.in[2], compute_graph.l2_152.in[2], compute_graph.l2_153.in[2], compute_graph.l2_155.in[2], compute_graph.l2_156.in[2], compute_graph.l2_157.in[2], compute_graph.l2_158.in[2], compute_graph.l2_159.in[2], compute_graph.l2_160.in[2], compute_graph.l2_162.in[2], compute_graph.l2_163.in[2], compute_graph.l2_164.in[2], compute_graph.l2_165.in[2], compute_graph.l2_166.in[2], compute_graph.l2_167.in[2], compute_graph.l2_168.in[2], compute_graph.l2_169.in[2], compute_graph.l2_170.in[2], compute_graph.l2_171.in[2], compute_graph.l2_172.in[2], compute_graph.l2_173.in[2], compute_graph.l2_175.in[2], compute_graph.l2_176.in[2], compute_graph.l2_177.in[2], compute_graph.l2_178.in[2], compute_graph.l2_179.in[2], compute_graph.l2_180.in[2], compute_graph.l2_182.in[2], compute_graph.l2_183.in[2], compute_graph.l2_184.in[2], compute_graph.l2_185.in[2], compute_graph.l2_186.in[2], compute_graph.l2_187.in[2], compute_graph.l2_188.in[2], compute_graph.l2_189.in[2], compute_graph.l2_190.in[2], compute_graph.l2_191.in[2], compute_graph.l2_192.in[2], compute_graph.l2_193.in[2], compute_graph.l2_195.in[2], compute_graph.l2_196.in[2], compute_graph.l2_197.in[2], compute_graph.l2_198.in[2], compute_graph.l2_199.in[2], compute_graph.l2_200.in[2], compute_graph.l2_202.in[2], compute_graph.l2_203.in[2], compute_graph.l2_204.in[2], compute_graph.l2_205.in[2], compute_graph.l2_206.in[2], compute_graph.l2_207.in[2], compute_graph.l2_208.in[2], compute_graph.l2_210.in[2], compute_graph.l2_211.in[2], compute_graph.l2_212.in[2], compute_graph.l2_214.in[2], compute_graph.l2_218.in[2], compute_graph.l2_219.in[2], compute_graph.l2_221.in[2], compute_graph.l2_222.in[2], compute_graph.l2_224.in[2], compute_graph.l2_226.in[2], compute_graph.l2_227.in[2], compute_graph.l2_228.in[2], compute_graph.l2_229.in[2], compute_graph.l2_233.in[2], compute_graph.l2_234.in[2], compute_graph.l2_235.in[2], compute_graph.l2_237.in[2], compute_graph.l2_241.in[2], compute_graph.l2_242.in[2], compute_graph.l2_244.in[2], compute_graph.l2_245.in[2], compute_graph.l2_247.in[2], compute_graph.l2_249.in[2], compute_graph.l2_250.in[2], compute_graph.l2_251.in[2], compute_graph.l2_252.in[2], compute_graph.l2_257.in[2], compute_graph.l2_261.in[2], compute_graph.l2_262.in[2], compute_graph.l2_264.in[2], compute_graph.l2_265.in[2], compute_graph.l2_267.in[2], compute_graph.l2_269.in[2], compute_graph.l2_270.in[2], compute_graph.l2_271.in[2], compute_graph.l2_272.in[2], compute_graph.l2_276.in[2], compute_graph.l2_280.in[2], compute_graph.l2_281.in[2], compute_graph.l2_283.in[2], compute_graph.l2_284.in[2], compute_graph.l2_286.in[2], compute_graph.l2_288.in[2], compute_graph.l2_289.in[2], compute_graph.l2_290.in[2], compute_graph.l2_291.in[2], compute_graph.l2_295.in[2], compute_graph.l2_296.in[2], compute_graph.l2_297.in[2], compute_graph.L2_OFM_Buffer_layer_299.in[2], compute_graph.l2_301.in[2], compute_graph.L2_OFM_Buffer_layer_302.in[2], compute_graph.templated_graph_4.trans_mt_ofm.in[2], compute_graph.templated_graph_231.ofm_mem.in[2], compute_graph.templated_graph_254.ofm_mem.in[2], compute_graph.templated_graph_258.ofm.in[2], compute_graph.templated_graph_259.ofm.in[2], compute_graph.templated_graph_260.ofm.in[2], compute_graph.templated_graph_263.ofm.in[2], compute_graph.templated_graph_266.ofm.in[2], compute_graph.templated_graph_268.ofm.in[2], compute_graph.templated_graph_273.ofm.in[2], compute_graph.templated_graph_274.ofm_mem.in[2], compute_graph.templated_graph_293.ofm_mem.in[2], compute_graph.templated_graph_298.ofm.in[2], compute_graph.templated_graph_300.ofm.in[2]) + +Column 1Row 0Channel 3 + +i3_pi3(compute_graph.l2_1.in[3], compute_graph.l2_6.in[3], compute_graph.l2_7.in[3], compute_graph.l2_8.in[3], compute_graph.l2_9.in[3], compute_graph.l2_10.in[3], compute_graph.l2_11.in[3], compute_graph.l2_12.in[3], compute_graph.l2_13.in[3], compute_graph.l2_14.in[3], compute_graph.l2_15.in[3], compute_graph.l2_16.in[3], compute_graph.l2_18.in[3], compute_graph.l2_19.in[3], compute_graph.l2_20.in[3], compute_graph.l2_22.in[3], compute_graph.l2_23.in[3], compute_graph.l2_24.in[3], compute_graph.l2_26.in[3], compute_graph.l2_27.in[3], compute_graph.l2_28.in[3], compute_graph.l2_29.in[3], compute_graph.l2_30.in[3], compute_graph.l2_31.in[3], compute_graph.l2_33.in[3], compute_graph.l2_34.in[3], compute_graph.l2_35.in[3], compute_graph.l2_36.in[3], compute_graph.l2_38.in[3], compute_graph.l2_39.in[3], compute_graph.l2_40.in[3], compute_graph.l2_41.in[3], compute_graph.l2_42.in[3], compute_graph.l2_43.in[3], compute_graph.l2_45.in[3], compute_graph.l2_46.in[3], compute_graph.l2_47.in[3], compute_graph.l2_48.in[3], compute_graph.l2_50.in[3], compute_graph.l2_51.in[3], compute_graph.l2_52.in[3], compute_graph.l2_53.in[3], compute_graph.l2_54.in[3], compute_graph.l2_55.in[3], compute_graph.l2_57.in[3], compute_graph.l2_58.in[3], compute_graph.l2_59.in[3], compute_graph.l2_60.in[3], compute_graph.l2_61.in[3], compute_graph.l2_62.in[3], compute_graph.l2_63.in[3], compute_graph.l2_65.in[3], compute_graph.l2_66.in[3], compute_graph.l2_67.in[3], compute_graph.l2_68.in[3], compute_graph.l2_69.in[3], compute_graph.l2_70.in[3], compute_graph.l2_71.in[3], compute_graph.l2_72.in[3], compute_graph.l2_73.in[3], compute_graph.l2_74.in[3], compute_graph.l2_75.in[3], compute_graph.l2_76.in[3], compute_graph.l2_77.in[3], compute_graph.l2_78.in[3], compute_graph.l2_79.in[3], compute_graph.l2_80.in[3], compute_graph.l2_81.in[3], compute_graph.l2_82.in[3], compute_graph.l2_83.in[3], compute_graph.l2_84.in[3], compute_graph.l2_85.in[3], compute_graph.l2_86.in[3], compute_graph.l2_87.in[3], compute_graph.l2_88.in[3], compute_graph.l2_89.in[3], compute_graph.l2_90.in[3], compute_graph.l2_91.in[3], compute_graph.l2_92.in[3], compute_graph.l2_93.in[3], compute_graph.l2_94.in[3], compute_graph.l2_95.in[3], compute_graph.l2_96.in[3], compute_graph.l2_97.in[3], compute_graph.l2_98.in[3], compute_graph.l2_99.in[3], compute_graph.l2_100.in[3], compute_graph.l2_101.in[3], compute_graph.l2_102.in[3], compute_graph.l2_103.in[3], compute_graph.l2_104.in[3], compute_graph.l2_105.in[3], compute_graph.l2_106.in[3], compute_graph.l2_107.in[3], compute_graph.l2_108.in[3], compute_graph.l2_109.in[3], compute_graph.l2_110.in[3], compute_graph.l2_111.in[3], compute_graph.l2_112.in[3], compute_graph.l2_113.in[3], compute_graph.l2_115.in[3], compute_graph.l2_116.in[3], compute_graph.l2_117.in[3], compute_graph.l2_118.in[3], compute_graph.l2_119.in[3], compute_graph.l2_120.in[3], compute_graph.l2_122.in[3], compute_graph.l2_123.in[3], compute_graph.l2_124.in[3], compute_graph.l2_125.in[3], compute_graph.l2_126.in[3], compute_graph.l2_127.in[3], compute_graph.l2_128.in[3], compute_graph.l2_129.in[3], compute_graph.l2_130.in[3], compute_graph.l2_131.in[3], compute_graph.l2_132.in[3], compute_graph.l2_133.in[3], compute_graph.l2_135.in[3], compute_graph.l2_136.in[3], compute_graph.l2_137.in[3], compute_graph.l2_138.in[3], compute_graph.l2_139.in[3], compute_graph.l2_140.in[3], compute_graph.l2_142.in[3], compute_graph.l2_143.in[3], compute_graph.l2_144.in[3], compute_graph.l2_145.in[3], compute_graph.l2_146.in[3], compute_graph.l2_147.in[3], compute_graph.l2_148.in[3], compute_graph.l2_149.in[3], compute_graph.l2_150.in[3], compute_graph.l2_151.in[3], compute_graph.l2_152.in[3], compute_graph.l2_153.in[3], compute_graph.l2_155.in[3], compute_graph.l2_156.in[3], compute_graph.l2_157.in[3], compute_graph.l2_158.in[3], compute_graph.l2_159.in[3], compute_graph.l2_160.in[3], compute_graph.l2_162.in[3], compute_graph.l2_163.in[3], compute_graph.l2_164.in[3], compute_graph.l2_165.in[3], compute_graph.l2_166.in[3], compute_graph.l2_167.in[3], compute_graph.l2_168.in[3], compute_graph.l2_169.in[3], compute_graph.l2_170.in[3], compute_graph.l2_171.in[3], compute_graph.l2_172.in[3], compute_graph.l2_173.in[3], compute_graph.l2_175.in[3], compute_graph.l2_176.in[3], compute_graph.l2_177.in[3], compute_graph.l2_178.in[3], compute_graph.l2_179.in[3], compute_graph.l2_180.in[3], compute_graph.l2_182.in[3], compute_graph.l2_183.in[3], compute_graph.l2_184.in[3], compute_graph.l2_185.in[3], compute_graph.l2_186.in[3], compute_graph.l2_187.in[3], compute_graph.l2_188.in[3], compute_graph.l2_189.in[3], compute_graph.l2_190.in[3], compute_graph.l2_191.in[3], compute_graph.l2_192.in[3], compute_graph.templated_graph_293.ofm_mem.in[3], compute_graph.l2_193.in[3], compute_graph.l2_195.in[3], compute_graph.l2_196.in[3], compute_graph.l2_197.in[3], compute_graph.l2_198.in[3], compute_graph.l2_199.in[3], compute_graph.l2_200.in[3], compute_graph.l2_202.in[3], compute_graph.l2_203.in[3], compute_graph.l2_204.in[3], compute_graph.l2_205.in[3], compute_graph.l2_206.in[3], compute_graph.l2_207.in[3], compute_graph.l2_208.in[3], compute_graph.l2_210.in[3], compute_graph.l2_211.in[3], compute_graph.l2_212.in[3], compute_graph.l2_214.in[3], compute_graph.l2_218.in[3], compute_graph.l2_219.in[3], compute_graph.l2_221.in[3], compute_graph.l2_222.in[3], compute_graph.l2_224.in[3], compute_graph.l2_226.in[3], compute_graph.l2_227.in[3], compute_graph.l2_228.in[3], compute_graph.l2_229.in[3], compute_graph.l2_233.in[3], compute_graph.l2_234.in[3], compute_graph.l2_235.in[3], compute_graph.l2_237.in[3], compute_graph.l2_241.in[3], compute_graph.l2_242.in[3], compute_graph.l2_244.in[3], compute_graph.l2_245.in[3], compute_graph.l2_247.in[3], compute_graph.l2_249.in[3], compute_graph.l2_250.in[3], compute_graph.l2_251.in[3], compute_graph.l2_252.in[3], compute_graph.l2_257.in[3], compute_graph.l2_261.in[3], compute_graph.l2_262.in[3], compute_graph.l2_264.in[3], compute_graph.l2_265.in[3], compute_graph.l2_267.in[3], compute_graph.l2_269.in[3], compute_graph.l2_270.in[3], compute_graph.l2_271.in[3], compute_graph.l2_272.in[3], compute_graph.l2_276.in[3], compute_graph.l2_280.in[3], compute_graph.l2_281.in[3], compute_graph.l2_283.in[3], compute_graph.l2_284.in[3], compute_graph.l2_286.in[3], compute_graph.l2_288.in[3], compute_graph.l2_289.in[3], compute_graph.l2_290.in[3], compute_graph.l2_291.in[3], compute_graph.l2_295.in[3], compute_graph.l2_296.in[3], compute_graph.l2_297.in[3], compute_graph.L2_OFM_Buffer_layer_299.in[3], compute_graph.l2_301.in[3], compute_graph.L2_OFM_Buffer_layer_302.in[3], compute_graph.templated_graph_4.trans_mt_ofm.in[3], compute_graph.templated_graph_231.ofm_mem.in[3], compute_graph.templated_graph_254.ofm_mem.in[3], compute_graph.templated_graph_258.ofm.in[3], compute_graph.templated_graph_259.ofm.in[3], compute_graph.templated_graph_260.ofm.in[3], compute_graph.templated_graph_263.ofm.in[3], compute_graph.templated_graph_266.ofm.in[3], compute_graph.templated_graph_268.ofm.in[3], compute_graph.templated_graph_273.ofm.in[3], compute_graph.templated_graph_274.ofm_mem.in[3], compute_graph.templated_graph_298.ofm.in[3], compute_graph.templated_graph_300.ofm.in[3]) + +Column 1Row 0Channel 4 + +i3_po4(compute_graph.templated_graph_49.trans_mt_ifm.out[0], compute_graph.templated_graph_64.ifm.out[0], compute_graph.templated_graph_114.trans_mt_ifm.out[0], compute_graph.templated_graph_134.trans_mt_ifm.out[0], compute_graph.templated_graph_3.ifm.out[0], compute_graph.templated_graph_5.ifm.out[0], compute_graph.templated_graph_17.ifm.out[0], compute_graph.templated_graph_21.ifm.out[0], compute_graph.templated_graph_25.trans_mt_ifm.out[0], compute_graph.templated_graph_37.trans_mt_ifm.out[0], compute_graph.templated_graph_154.trans_mt_ifm.out[0], compute_graph.templated_graph_174.trans_mt_ifm.out[0], compute_graph.templated_graph_194.trans_mt_ifm.out[0], compute_graph.templated_graph_209.trans_mt_ifm.out[0]) + +Column 1Row 0Channel 5 + +i3_pi4(compute_graph.templated_graph_49.trans_mt_ofm.in[0], compute_graph.templated_graph_64.ofm.in[0], compute_graph.templated_graph_114.trans_mt_ofm.in[0], compute_graph.templated_graph_134.trans_mt_ofm.in[0], compute_graph.templated_graph_3.ofm.in[0], compute_graph.templated_graph_5.ofm.in[0], compute_graph.templated_graph_17.ofm.in[0], compute_graph.templated_graph_21.ofm.in[0], compute_graph.templated_graph_25.trans_mt_ofm.in[0], compute_graph.templated_graph_37.trans_mt_ofm.in[0], compute_graph.templated_graph_154.trans_mt_ofm.in[0], compute_graph.templated_graph_174.trans_mt_ofm.in[0], compute_graph.templated_graph_194.trans_mt_ofm.in[0], compute_graph.templated_graph_209.trans_mt_ofm.in[0]) + +Column 0Row 0Channel 0 + +i4_pi0(compute_graph.templated_graph_4.trans_comp_nd[0][0].in[0], compute_graph.flexml_layers[36].compute_node[0][0].in[0], compute_graph.templated_graph_231.mk[0][0].in[0], compute_graph.templated_graph_254.mk[0][0].in[0], compute_graph.templated_graph_258.compute_node[0][0].in[0], compute_graph.templated_graph_259.compute_node[0][0].in[0], compute_graph.templated_graph_260.mk[0][0].in[0], compute_graph.templated_graph_263.compute_node[0][0].in[0], compute_graph.templated_graph_266.compute_node[0][0].in[0], compute_graph.templated_graph_268.mk[0][0].in[0], compute_graph.templated_graph_273.mk[0][0].in[0], compute_graph.templated_graph_274.mk[0][0].in[0], compute_graph.flexml_layers[63].compute_node[0][0].in[0], compute_graph.flexml_layers[64].compute_node[0][0].in[0], compute_graph.templated_graph_293.mk[0][0].in[0], compute_graph.templated_graph_298.compute_node[0][0].in[0], compute_graph.templated_graph_300.compute_node[0][0].in[0], compute_graph.flexml_layers[0].compute_node[0][0].in[0], compute_graph.flexml_layers[1].compute_node[0][0].in[0], compute_graph.flexml_layers[1].compute_node[0][0].in[2], compute_graph.flexml_layers[2].compute_node[0][0].in[0], compute_graph.flexml_layers[2].compute_node[0][0].in[2], compute_graph.flexml_layers[3].compute_node[0][0].in[0], compute_graph.flexml_layers[4].compute_node[0][0].in[0], compute_graph.flexml_layers[5].compute_node[0][0].in[0], compute_graph.flexml_layers[6].compute_node[0][0].in[0], compute_graph.flexml_layers[7].compute_node[0][0].in[0], compute_graph.flexml_layers[7].compute_node[0][0].in[2], compute_graph.flexml_layers[8].compute_node[0][0].in[0], compute_graph.flexml_layers[9].compute_node[0][0].in[0], compute_graph.flexml_layers[9].compute_node[0][0].in[2], compute_graph.flexml_layers[10].compute_node[0][0].in[0], compute_graph.flexml_layers[11].compute_node[0][0].in[0], compute_graph.flexml_layers[12].compute_node[0][0].in[0], compute_graph.flexml_layers[13].compute_node[0][0].in[0], compute_graph.flexml_layers[14].compute_node[0][0].in[0], compute_graph.flexml_layers[15].compute_node[0][0].in[0], compute_graph.flexml_layers[15].compute_node[0][0].in[2], compute_graph.flexml_layers[16].compute_node[0][0].in[0], compute_graph.flexml_layers[17].compute_node[0][0].in[0], compute_graph.flexml_layers[18].compute_node[0][0].in[0], compute_graph.flexml_layers[19].compute_node[0][0].in[0], compute_graph.flexml_layers[20].compute_node[0][0].in[0], compute_graph.flexml_layers[21].compute_node[0][0].in[0], compute_graph.flexml_layers[22].compute_node[0][0].in[0], compute_graph.flexml_layers[23].compute_node[0][0].in[0], compute_graph.flexml_layers[24].compute_node[0][0].in[0], compute_graph.flexml_layers[24].compute_node[0][0].in[2], compute_graph.flexml_layers[25].compute_node[0][0].in[0], compute_graph.flexml_layers[26].compute_node[0][0].in[0], compute_graph.flexml_layers[27].compute_node[0][0].in[0], compute_graph.flexml_layers[28].compute_node[0][0].in[0], compute_graph.flexml_layers[29].compute_node[0][0].in[0], compute_graph.flexml_layers[30].compute_node[0][0].in[0], compute_graph.flexml_layers[31].compute_node[0][0].in[0], compute_graph.flexml_layers[32].compute_node[0][0].in[0], compute_graph.flexml_layers[33].compute_node[0][0].in[0], compute_graph.flexml_layers[34].compute_node[0][0].in[0], compute_graph.flexml_layers[34].compute_node[0][0].in[2], compute_graph.flexml_layers[35].compute_node[0][0].in[0], compute_graph.flexml_layers[35].compute_node[0][0].in[2], compute_graph.flexml_layers[144].compute_node[0][0].in[0], compute_graph.flexml_layers[145].compute_node[0][0].in[0], compute_graph.flexml_layers[37].compute_node[0][0].in[0], compute_graph.flexml_layers[38].compute_node[0][0].in[0], compute_graph.flexml_layers[39].compute_node[0][0].in[0], compute_graph.flexml_layers[40].compute_node[0][0].in[0], compute_graph.flexml_layers[41].compute_node[0][0].in[0], compute_graph.flexml_layers[42].compute_node[0][0].in[0], compute_graph.flexml_layers[43].compute_node[0][0].in[0], compute_graph.flexml_layers[44].compute_node[0][0].in[0], compute_graph.flexml_layers[44].compute_node[0][0].in[2], compute_graph.flexml_layers[45].compute_node[0][0].in[0], compute_graph.flexml_layers[45].compute_node[0][0].in[2], compute_graph.flexml_layers[46].compute_node[0][0].in[0], compute_graph.flexml_layers[47].compute_node[0][0].in[0], compute_graph.flexml_layers[48].compute_node[0][0].in[0], compute_graph.flexml_layers[49].compute_node[0][0].in[0], compute_graph.flexml_layers[50].compute_node[0][0].in[0], compute_graph.flexml_layers[50].compute_node[0][0].in[2], compute_graph.flexml_layers[51].compute_node[0][0].in[0], compute_graph.flexml_layers[52].compute_node[0][0].in[0], compute_graph.flexml_layers[53].compute_node[0][0].in[0], compute_graph.flexml_layers[54].compute_node[0][0].in[0], compute_graph.flexml_layers[55].compute_node[0][0].in[0], compute_graph.flexml_layers[55].compute_node[0][0].in[2], compute_graph.flexml_layers[56].compute_node[0][0].in[0], compute_graph.flexml_layers[57].compute_node[0][0].in[0], compute_graph.flexml_layers[58].compute_node[0][0].in[0], compute_graph.flexml_layers[59].compute_node[0][0].in[0], compute_graph.flexml_layers[60].compute_node[0][0].in[0], compute_graph.flexml_layers[61].compute_node[0][0].in[0], compute_graph.flexml_layers[61].compute_node[0][0].in[2], compute_graph.flexml_layers[62].compute_node[0][0].in[0], compute_graph.flexml_layers[201].compute_node[0][0].in[0], compute_graph.flexml_layers[202].compute_node[0][0].in[0], compute_graph.flexml_layers[65].compute_node[0][0].in[0], compute_graph.flexml_layers[66].compute_node[0][0].in[0], compute_graph.flexml_layers[66].compute_node[0][0].in[2], compute_graph.flexml_layers[67].compute_node[0][0].in[0], compute_graph.flexml_layers[67].compute_node[0][0].in[2], compute_graph.flexml_layers[68].compute_node[0][0].in[0], compute_graph.flexml_layers[69].compute_node[0][0].in[0], compute_graph.flexml_layers[70].compute_node[0][0].in[0], compute_graph.flexml_layers[71].compute_node[0][0].in[0], compute_graph.flexml_layers[72].compute_node[0][0].in[0], compute_graph.flexml_layers[72].compute_node[0][0].in[2], compute_graph.flexml_layers[73].compute_node[0][0].in[0], compute_graph.flexml_layers[74].compute_node[0][0].in[0], compute_graph.flexml_layers[75].compute_node[0][0].in[0], compute_graph.flexml_layers[76].compute_node[0][0].in[0], compute_graph.flexml_layers[77].compute_node[0][0].in[0], compute_graph.flexml_layers[77].compute_node[0][0].in[2], compute_graph.flexml_layers[78].compute_node[0][0].in[0], compute_graph.flexml_layers[78].compute_node[0][0].in[2], compute_graph.flexml_layers[79].compute_node[0][0].in[0], compute_graph.flexml_layers[80].compute_node[0][0].in[0], compute_graph.flexml_layers[81].compute_node[0][0].in[0], compute_graph.flexml_layers[82].compute_node[0][0].in[0], compute_graph.flexml_layers[83].compute_node[0][0].in[0], compute_graph.flexml_layers[83].compute_node[0][0].in[2], compute_graph.flexml_layers[84].compute_node[0][0].in[0], compute_graph.flexml_layers[85].compute_node[0][0].in[0], compute_graph.flexml_layers[86].compute_node[0][0].in[0], compute_graph.flexml_layers[87].compute_node[0][0].in[0], compute_graph.flexml_layers[88].compute_node[0][0].in[0], compute_graph.flexml_layers[88].compute_node[0][0].in[2], compute_graph.flexml_layers[89].compute_node[0][0].in[0], compute_graph.flexml_layers[89].compute_node[0][0].in[2], compute_graph.flexml_layers[90].compute_node[0][0].in[0], compute_graph.flexml_layers[91].compute_node[0][0].in[0], compute_graph.flexml_layers[92].compute_node[0][0].in[0], compute_graph.flexml_layers[93].compute_node[0][0].in[0], compute_graph.flexml_layers[94].compute_node[0][0].in[0], compute_graph.flexml_layers[94].compute_node[0][0].in[2], compute_graph.flexml_layers[95].compute_node[0][0].in[0], compute_graph.flexml_layers[96].compute_node[0][0].in[0], compute_graph.flexml_layers[97].compute_node[0][0].in[0], compute_graph.flexml_layers[98].compute_node[0][0].in[0], compute_graph.flexml_layers[99].compute_node[0][0].in[0], compute_graph.flexml_layers[99].compute_node[0][0].in[2], compute_graph.flexml_layers[100].compute_node[0][0].in[0], compute_graph.flexml_layers[101].compute_node[0][0].in[0], compute_graph.flexml_layers[102].compute_node[0][0].in[0], compute_graph.flexml_layers[103].compute_node[0][0].in[0], compute_graph.flexml_layers[104].compute_node[0][0].in[0], compute_graph.flexml_layers[105].compute_node[0][0].in[0], compute_graph.flexml_layers[106].compute_node[0][0].in[0], compute_graph.flexml_layers[106].compute_node[0][0].in[2], compute_graph.flexml_layers[107].compute_node[0][0].in[0], compute_graph.flexml_layers[108].compute_node[0][0].in[0], compute_graph.flexml_layers[109].compute_node[0][0].in[0], compute_graph.flexml_layers[110].compute_node[0][0].in[0], compute_graph.flexml_layers[111].compute_node[0][0].in[0], compute_graph.flexml_layers[112].compute_node[0][0].in[0], compute_graph.flexml_layers[112].compute_node[0][0].in[2], compute_graph.flexml_layers[113].compute_node[0][0].in[0], compute_graph.flexml_layers[114].compute_node[0][0].in[0], compute_graph.flexml_layers[115].compute_node[0][0].in[0], compute_graph.flexml_layers[116].compute_node[0][0].in[0], compute_graph.flexml_layers[117].compute_node[0][0].in[0], compute_graph.flexml_layers[117].compute_node[0][0].in[2], compute_graph.flexml_layers[118].compute_node[0][0].in[0], compute_graph.flexml_layers[119].compute_node[0][0].in[0], compute_graph.flexml_layers[120].compute_node[0][0].in[0], compute_graph.flexml_layers[121].compute_node[0][0].in[0], compute_graph.flexml_layers[122].compute_node[0][0].in[0], compute_graph.flexml_layers[123].compute_node[0][0].in[0], compute_graph.flexml_layers[124].compute_node[0][0].in[0], compute_graph.flexml_layers[124].compute_node[0][0].in[2], compute_graph.flexml_layers[125].compute_node[0][0].in[0], compute_graph.flexml_layers[125].compute_node[0][0].in[2], compute_graph.flexml_layers[126].compute_node[0][0].in[0], compute_graph.flexml_layers[127].compute_node[0][0].in[0], compute_graph.flexml_layers[128].compute_node[0][0].in[0], compute_graph.flexml_layers[129].compute_node[0][0].in[0], compute_graph.flexml_layers[130].compute_node[0][0].in[0], compute_graph.flexml_layers[130].compute_node[0][0].in[2], compute_graph.flexml_layers[131].compute_node[0][0].in[0], compute_graph.flexml_layers[132].compute_node[0][0].in[0], compute_graph.flexml_layers[133].compute_node[0][0].in[0], compute_graph.flexml_layers[134].compute_node[0][0].in[0], compute_graph.flexml_layers[135].compute_node[0][0].in[0], compute_graph.flexml_layers[135].compute_node[0][0].in[2], compute_graph.flexml_layers[136].compute_node[0][0].in[0], compute_graph.flexml_layers[137].compute_node[0][0].in[0], compute_graph.flexml_layers[138].compute_node[0][0].in[0], compute_graph.flexml_layers[139].compute_node[0][0].in[0], compute_graph.flexml_layers[140].compute_node[0][0].in[0], compute_graph.flexml_layers[141].compute_node[0][0].in[0], compute_graph.flexml_layers[142].compute_node[0][0].in[0], compute_graph.flexml_layers[142].compute_node[0][0].in[2], compute_graph.flexml_layers[143].compute_node[0][0].in[0], compute_graph.flexml_layers[146].compute_node[0][0].in[0], compute_graph.flexml_layers[147].compute_node[0][0].in[0], compute_graph.flexml_layers[148].compute_node[0][0].in[0], compute_graph.flexml_layers[148].compute_node[0][0].in[2], compute_graph.flexml_layers[149].compute_node[0][0].in[0], compute_graph.flexml_layers[150].compute_node[0][0].in[0], compute_graph.flexml_layers[151].compute_node[0][0].in[0], compute_graph.flexml_layers[152].compute_node[0][0].in[0], compute_graph.flexml_layers[153].compute_node[0][0].in[0], compute_graph.flexml_layers[153].compute_node[0][0].in[2], compute_graph.flexml_layers[154].compute_node[0][0].in[0], compute_graph.flexml_layers[155].compute_node[0][0].in[0], compute_graph.flexml_layers[156].compute_node[0][0].in[0], compute_graph.flexml_layers[157].compute_node[0][0].in[0], compute_graph.flexml_layers[158].compute_node[0][0].in[0], compute_graph.flexml_layers[159].compute_node[0][0].in[0], compute_graph.flexml_layers[160].compute_node[0][0].in[0], compute_graph.flexml_layers[160].compute_node[0][0].in[2], compute_graph.flexml_layers[161].compute_node[0][0].in[0], compute_graph.flexml_layers[161].compute_node[0][0].in[2], compute_graph.flexml_layers[162].compute_node[0][0].in[0], compute_graph.flexml_layers[163].compute_node[0][0].in[0], compute_graph.flexml_layers[164].compute_node[0][0].in[0], compute_graph.flexml_layers[165].compute_node[0][0].in[0], compute_graph.flexml_layers[166].compute_node[0][0].in[0], compute_graph.flexml_layers[166].compute_node[0][0].in[2], compute_graph.flexml_layers[167].compute_node[0][0].in[0], compute_graph.flexml_layers[168].compute_node[0][0].in[0], compute_graph.flexml_layers[169].compute_node[0][0].in[0], compute_graph.flexml_layers[170].compute_node[0][0].in[0], compute_graph.flexml_layers[171].compute_node[0][0].in[0], compute_graph.flexml_layers[171].compute_node[0][0].in[2], compute_graph.flexml_layers[172].compute_node[0][0].in[0], compute_graph.flexml_layers[173].compute_node[0][0].in[0], compute_graph.flexml_layers[174].compute_node[0][0].in[0], compute_graph.flexml_layers[175].compute_node[0][0].in[0], compute_graph.flexml_layers[176].compute_node[0][0].in[0], compute_graph.flexml_layers[177].compute_node[0][0].in[0], compute_graph.flexml_layers[178].compute_node[0][0].in[0], compute_graph.flexml_layers[178].compute_node[0][0].in[2], compute_graph.flexml_layers[179].compute_node[0][0].in[0], compute_graph.flexml_layers[179].compute_node[0][0].in[2], compute_graph.flexml_layers[180].compute_node[0][0].in[0], compute_graph.flexml_layers[181].compute_node[0][0].in[0], compute_graph.flexml_layers[182].compute_node[0][0].in[0], compute_graph.flexml_layers[183].compute_node[0][0].in[0], compute_graph.flexml_layers[184].compute_node[0][0].in[0], compute_graph.flexml_layers[184].compute_node[0][0].in[2], compute_graph.flexml_layers[185].compute_node[0][0].in[0], compute_graph.flexml_layers[186].compute_node[0][0].in[0], compute_graph.flexml_layers[187].compute_node[0][0].in[0], compute_graph.flexml_layers[188].compute_node[0][0].in[0], compute_graph.flexml_layers[188].compute_node[0][0].in[2], compute_graph.flexml_layers[189].compute_node[0][0].in[0], compute_graph.flexml_layers[190].compute_node[0][0].in[0], compute_graph.flexml_layers[191].compute_node[0][0].in[0], compute_graph.flexml_layers[191].compute_node[0][0].in[2], compute_graph.flexml_layers[192].compute_node[0][0].in[0], compute_graph.flexml_layers[192].compute_node[0][0].in[2], compute_graph.flexml_layers[193].compute_node[0][0].in[0], compute_graph.flexml_layers[193].compute_node[0][0].in[2], compute_graph.flexml_layers[194].compute_node[0][0].in[0], compute_graph.flexml_layers[195].compute_node[0][0].in[0], compute_graph.flexml_layers[196].compute_node[0][0].in[0], compute_graph.flexml_layers[196].compute_node[0][0].in[2], compute_graph.flexml_layers[197].compute_node[0][0].in[0], compute_graph.flexml_layers[197].compute_node[0][0].in[2], compute_graph.flexml_layers[198].compute_node[0][0].in[0], compute_graph.flexml_layers[199].compute_node[0][0].in[0], compute_graph.flexml_layers[200].compute_node[0][0].in[0], compute_graph.flexml_layers[203].compute_node[0][0].in[0], compute_graph.flexml_layers[204].compute_node[0][0].in[0], compute_graph.flexml_layers[204].compute_node[0][0].in[2], compute_graph.flexml_layers[205].compute_node[0][0].in[0], compute_graph.flexml_layers[205].compute_node[0][0].in[2], compute_graph.flexml_layers[206].compute_node[0][0].in[0], compute_graph.flexml_layers[206].compute_node[0][0].in[2], compute_graph.flexml_layers[207].compute_node[0][0].in[0], compute_graph.flexml_layers[208].compute_node[0][0].in[0], compute_graph.flexml_layers[209].compute_node[0][0].in[0], compute_graph.flexml_layers[209].compute_node[0][0].in[2], compute_graph.flexml_layers[210].compute_node[0][0].in[0], compute_graph.flexml_layers[210].compute_node[0][0].in[2], compute_graph.flexml_layers[211].compute_node[0][0].in[0], compute_graph.flexml_layers[212].compute_node[0][0].in[0], compute_graph.flexml_layers[213].compute_node[0][0].in[0], compute_graph.flexml_layers[214].compute_node[0][0].in[0], compute_graph.flexml_layers[214].compute_node[0][0].in[2], compute_graph.flexml_layers[215].compute_node[0][0].in[0], compute_graph.flexml_layers[215].compute_node[0][0].in[2], compute_graph.flexml_layers[216].compute_node[0][0].in[0], compute_graph.flexml_layers[216].compute_node[0][0].in[2], compute_graph.flexml_layers[217].compute_node[0][0].in[0], compute_graph.flexml_layers[218].compute_node[0][0].in[0], compute_graph.flexml_layers[219].compute_node[0][0].in[0], compute_graph.flexml_layers[219].compute_node[0][0].in[2], compute_graph.flexml_layers[220].compute_node[0][0].in[0], compute_graph.flexml_layers[220].compute_node[0][0].in[2], compute_graph.flexml_layers[221].compute_node[0][0].in[0], compute_graph.flexml_layers[222].compute_node[0][0].in[0], compute_graph.flexml_layers[223].compute_node[0][0].in[0], compute_graph.flexml_layers[224].compute_node[0][0].in[0], compute_graph.flexml_layers[224].compute_node[0][0].in[2], compute_graph.flexml_layers[225].compute_node[0][0].in[0], compute_graph.flexml_layers[225].compute_node[0][0].in[2], compute_graph.flexml_layers[226].compute_node[0][0].in[0], compute_graph.flexml_layers[226].compute_node[0][0].in[2], compute_graph.flexml_layers[227].compute_node[0][0].in[0], compute_graph.flexml_layers[228].compute_node[0][0].in[0], compute_graph.flexml_layers[229].compute_node[0][0].in[0], compute_graph.flexml_layers[229].compute_node[0][0].in[2], compute_graph.flexml_layers[230].compute_node[0][0].in[0], compute_graph.flexml_layers[230].compute_node[0][0].in[2], compute_graph.flexml_layers[231].compute_node[0][0].in[0], compute_graph.flexml_layers[232].compute_node[0][0].in[0], compute_graph.flexml_layers[233].compute_node[0][0].in[0], compute_graph.flexml_layers[234].compute_node[0][0].in[0], compute_graph.flexml_layers[235].compute_node[0][0].in[0], compute_graph.flexml_layers[235].compute_node[0][0].in[2], compute_graph.flexml_layers[236].compute_node[0][0].in[0]) + +Column 0Row 0Channel 0 + +i4_po0(compute_graph.templated_graph_4.trans_comp_nd[0][0].out[0], compute_graph.flexml_layers[36].compute_node[0][0].out[0], compute_graph.templated_graph_231.mk[0][0].out[0], compute_graph.templated_graph_254.mk[0][0].out[0], compute_graph.templated_graph_258.compute_node[0][0].out[0], compute_graph.templated_graph_259.compute_node[0][0].out[0], compute_graph.templated_graph_260.mk[0][0].out[0], compute_graph.templated_graph_263.compute_node[0][0].out[0], compute_graph.templated_graph_266.compute_node[0][0].out[0], compute_graph.templated_graph_268.mk[0][0].out[0], compute_graph.templated_graph_273.mk[0][0].out[0], compute_graph.templated_graph_274.mk[0][0].out[0], compute_graph.flexml_layers[63].compute_node[0][0].out[0], compute_graph.flexml_layers[64].compute_node[0][0].out[0], compute_graph.templated_graph_293.mk[0][0].out[0], compute_graph.templated_graph_298.compute_node[0][0].out[0], compute_graph.templated_graph_300.compute_node[0][0].out[0], compute_graph.flexml_layers[0].compute_node[0][0].out[0], compute_graph.flexml_layers[1].compute_node[0][0].out[0], compute_graph.flexml_layers[2].compute_node[0][0].out[0], compute_graph.flexml_layers[3].compute_node[0][0].out[0], compute_graph.flexml_layers[4].compute_node[0][0].out[0], compute_graph.flexml_layers[5].compute_node[0][0].out[0], compute_graph.flexml_layers[6].compute_node[0][0].out[0], compute_graph.flexml_layers[7].compute_node[0][0].out[0], compute_graph.flexml_layers[8].compute_node[0][0].out[0], compute_graph.flexml_layers[9].compute_node[0][0].out[0], compute_graph.flexml_layers[10].compute_node[0][0].out[0], compute_graph.flexml_layers[11].compute_node[0][0].out[0], compute_graph.flexml_layers[12].compute_node[0][0].out[0], compute_graph.flexml_layers[13].compute_node[0][0].out[0], compute_graph.flexml_layers[14].compute_node[0][0].out[0], compute_graph.flexml_layers[15].compute_node[0][0].out[0], compute_graph.flexml_layers[16].compute_node[0][0].out[0], compute_graph.flexml_layers[17].compute_node[0][0].out[0], compute_graph.flexml_layers[18].compute_node[0][0].out[0], compute_graph.flexml_layers[19].compute_node[0][0].out[0], compute_graph.flexml_layers[20].compute_node[0][0].out[0], compute_graph.flexml_layers[21].compute_node[0][0].out[0], compute_graph.flexml_layers[22].compute_node[0][0].out[0], compute_graph.flexml_layers[23].compute_node[0][0].out[0], compute_graph.flexml_layers[24].compute_node[0][0].out[0], compute_graph.flexml_layers[25].compute_node[0][0].out[0], compute_graph.flexml_layers[26].compute_node[0][0].out[0], compute_graph.flexml_layers[27].compute_node[0][0].out[0], compute_graph.flexml_layers[28].compute_node[0][0].out[0], compute_graph.flexml_layers[29].compute_node[0][0].out[0], compute_graph.flexml_layers[30].compute_node[0][0].out[0], compute_graph.flexml_layers[31].compute_node[0][0].out[0], compute_graph.flexml_layers[32].compute_node[0][0].out[0], compute_graph.flexml_layers[33].compute_node[0][0].out[0], compute_graph.flexml_layers[34].compute_node[0][0].out[0], compute_graph.flexml_layers[35].compute_node[0][0].out[0], compute_graph.flexml_layers[144].compute_node[0][0].out[0], compute_graph.flexml_layers[145].compute_node[0][0].out[0], compute_graph.flexml_layers[37].compute_node[0][0].out[0], compute_graph.flexml_layers[38].compute_node[0][0].out[0], compute_graph.flexml_layers[39].compute_node[0][0].out[0], compute_graph.flexml_layers[40].compute_node[0][0].out[0], compute_graph.flexml_layers[41].compute_node[0][0].out[0], compute_graph.flexml_layers[42].compute_node[0][0].out[0], compute_graph.flexml_layers[43].compute_node[0][0].out[0], compute_graph.flexml_layers[44].compute_node[0][0].out[0], compute_graph.flexml_layers[45].compute_node[0][0].out[0], compute_graph.flexml_layers[46].compute_node[0][0].out[0], compute_graph.flexml_layers[47].compute_node[0][0].out[0], compute_graph.flexml_layers[48].compute_node[0][0].out[0], compute_graph.flexml_layers[49].compute_node[0][0].out[0], compute_graph.flexml_layers[50].compute_node[0][0].out[0], compute_graph.flexml_layers[51].compute_node[0][0].out[0], compute_graph.flexml_layers[52].compute_node[0][0].out[0], compute_graph.flexml_layers[53].compute_node[0][0].out[0], compute_graph.flexml_layers[54].compute_node[0][0].out[0], compute_graph.flexml_layers[55].compute_node[0][0].out[0], compute_graph.flexml_layers[56].compute_node[0][0].out[0], compute_graph.flexml_layers[57].compute_node[0][0].out[0], compute_graph.flexml_layers[58].compute_node[0][0].out[0], compute_graph.flexml_layers[59].compute_node[0][0].out[0], compute_graph.flexml_layers[60].compute_node[0][0].out[0], compute_graph.flexml_layers[61].compute_node[0][0].out[0], compute_graph.flexml_layers[62].compute_node[0][0].out[0], compute_graph.flexml_layers[201].compute_node[0][0].out[0], compute_graph.flexml_layers[202].compute_node[0][0].out[0], compute_graph.flexml_layers[65].compute_node[0][0].out[0], compute_graph.flexml_layers[66].compute_node[0][0].out[0], compute_graph.flexml_layers[67].compute_node[0][0].out[0], compute_graph.flexml_layers[68].compute_node[0][0].out[0], compute_graph.flexml_layers[69].compute_node[0][0].out[0], compute_graph.flexml_layers[70].compute_node[0][0].out[0], compute_graph.flexml_layers[71].compute_node[0][0].out[0], compute_graph.flexml_layers[72].compute_node[0][0].out[0], compute_graph.flexml_layers[73].compute_node[0][0].out[0], compute_graph.flexml_layers[74].compute_node[0][0].out[0], compute_graph.flexml_layers[75].compute_node[0][0].out[0], compute_graph.flexml_layers[76].compute_node[0][0].out[0], compute_graph.flexml_layers[77].compute_node[0][0].out[0], compute_graph.flexml_layers[78].compute_node[0][0].out[0], compute_graph.flexml_layers[79].compute_node[0][0].out[0], compute_graph.flexml_layers[80].compute_node[0][0].out[0], compute_graph.flexml_layers[81].compute_node[0][0].out[0], compute_graph.flexml_layers[82].compute_node[0][0].out[0], compute_graph.flexml_layers[83].compute_node[0][0].out[0], compute_graph.flexml_layers[84].compute_node[0][0].out[0], compute_graph.flexml_layers[85].compute_node[0][0].out[0], compute_graph.flexml_layers[86].compute_node[0][0].out[0], compute_graph.flexml_layers[87].compute_node[0][0].out[0], compute_graph.flexml_layers[88].compute_node[0][0].out[0], compute_graph.flexml_layers[89].compute_node[0][0].out[0], compute_graph.flexml_layers[90].compute_node[0][0].out[0], compute_graph.flexml_layers[91].compute_node[0][0].out[0], compute_graph.flexml_layers[92].compute_node[0][0].out[0], compute_graph.flexml_layers[93].compute_node[0][0].out[0], compute_graph.flexml_layers[94].compute_node[0][0].out[0], compute_graph.flexml_layers[95].compute_node[0][0].out[0], compute_graph.flexml_layers[96].compute_node[0][0].out[0], compute_graph.flexml_layers[97].compute_node[0][0].out[0], compute_graph.flexml_layers[98].compute_node[0][0].out[0], compute_graph.flexml_layers[99].compute_node[0][0].out[0], compute_graph.flexml_layers[100].compute_node[0][0].out[0], compute_graph.flexml_layers[101].compute_node[0][0].out[0], compute_graph.flexml_layers[102].compute_node[0][0].out[0], compute_graph.flexml_layers[103].compute_node[0][0].out[0], compute_graph.flexml_layers[104].compute_node[0][0].out[0], compute_graph.flexml_layers[105].compute_node[0][0].out[0], compute_graph.flexml_layers[106].compute_node[0][0].out[0], compute_graph.flexml_layers[107].compute_node[0][0].out[0], compute_graph.flexml_layers[108].compute_node[0][0].out[0], compute_graph.flexml_layers[109].compute_node[0][0].out[0], compute_graph.flexml_layers[110].compute_node[0][0].out[0], compute_graph.flexml_layers[111].compute_node[0][0].out[0], compute_graph.flexml_layers[112].compute_node[0][0].out[0], compute_graph.flexml_layers[113].compute_node[0][0].out[0], compute_graph.flexml_layers[114].compute_node[0][0].out[0], compute_graph.flexml_layers[115].compute_node[0][0].out[0], compute_graph.flexml_layers[116].compute_node[0][0].out[0], compute_graph.flexml_layers[117].compute_node[0][0].out[0], compute_graph.flexml_layers[118].compute_node[0][0].out[0], compute_graph.flexml_layers[119].compute_node[0][0].out[0], compute_graph.flexml_layers[120].compute_node[0][0].out[0], compute_graph.flexml_layers[121].compute_node[0][0].out[0], compute_graph.flexml_layers[122].compute_node[0][0].out[0], compute_graph.flexml_layers[123].compute_node[0][0].out[0], compute_graph.flexml_layers[124].compute_node[0][0].out[0], compute_graph.flexml_layers[125].compute_node[0][0].out[0], compute_graph.flexml_layers[126].compute_node[0][0].out[0], compute_graph.flexml_layers[127].compute_node[0][0].out[0], compute_graph.flexml_layers[128].compute_node[0][0].out[0], compute_graph.flexml_layers[129].compute_node[0][0].out[0], compute_graph.flexml_layers[130].compute_node[0][0].out[0], compute_graph.flexml_layers[131].compute_node[0][0].out[0], compute_graph.flexml_layers[132].compute_node[0][0].out[0], compute_graph.flexml_layers[133].compute_node[0][0].out[0], compute_graph.flexml_layers[134].compute_node[0][0].out[0], compute_graph.flexml_layers[135].compute_node[0][0].out[0], compute_graph.flexml_layers[136].compute_node[0][0].out[0], compute_graph.flexml_layers[137].compute_node[0][0].out[0], compute_graph.flexml_layers[138].compute_node[0][0].out[0], compute_graph.flexml_layers[139].compute_node[0][0].out[0], compute_graph.flexml_layers[140].compute_node[0][0].out[0], compute_graph.flexml_layers[141].compute_node[0][0].out[0], compute_graph.flexml_layers[142].compute_node[0][0].out[0], compute_graph.flexml_layers[143].compute_node[0][0].out[0], compute_graph.flexml_layers[146].compute_node[0][0].out[0], compute_graph.flexml_layers[147].compute_node[0][0].out[0], compute_graph.flexml_layers[148].compute_node[0][0].out[0], compute_graph.flexml_layers[149].compute_node[0][0].out[0], compute_graph.flexml_layers[150].compute_node[0][0].out[0], compute_graph.flexml_layers[151].compute_node[0][0].out[0], compute_graph.flexml_layers[152].compute_node[0][0].out[0], compute_graph.flexml_layers[153].compute_node[0][0].out[0], compute_graph.flexml_layers[154].compute_node[0][0].out[0], compute_graph.flexml_layers[155].compute_node[0][0].out[0], compute_graph.flexml_layers[156].compute_node[0][0].out[0], compute_graph.flexml_layers[157].compute_node[0][0].out[0], compute_graph.flexml_layers[158].compute_node[0][0].out[0], compute_graph.flexml_layers[159].compute_node[0][0].out[0], compute_graph.flexml_layers[160].compute_node[0][0].out[0], compute_graph.flexml_layers[161].compute_node[0][0].out[0], compute_graph.flexml_layers[162].compute_node[0][0].out[0], compute_graph.flexml_layers[163].compute_node[0][0].out[0], compute_graph.flexml_layers[164].compute_node[0][0].out[0], compute_graph.flexml_layers[165].compute_node[0][0].out[0], compute_graph.flexml_layers[166].compute_node[0][0].out[0], compute_graph.flexml_layers[167].compute_node[0][0].out[0], compute_graph.flexml_layers[168].compute_node[0][0].out[0], compute_graph.flexml_layers[169].compute_node[0][0].out[0], compute_graph.flexml_layers[170].compute_node[0][0].out[0], compute_graph.flexml_layers[171].compute_node[0][0].out[0], compute_graph.flexml_layers[172].compute_node[0][0].out[0], compute_graph.flexml_layers[173].compute_node[0][0].out[0], compute_graph.flexml_layers[174].compute_node[0][0].out[0], compute_graph.flexml_layers[175].compute_node[0][0].out[0], compute_graph.flexml_layers[176].compute_node[0][0].out[0], compute_graph.flexml_layers[177].compute_node[0][0].out[0], compute_graph.flexml_layers[178].compute_node[0][0].out[0], compute_graph.flexml_layers[179].compute_node[0][0].out[0], compute_graph.flexml_layers[180].compute_node[0][0].out[0], compute_graph.flexml_layers[181].compute_node[0][0].out[0], compute_graph.flexml_layers[182].compute_node[0][0].out[0], compute_graph.flexml_layers[183].compute_node[0][0].out[0], compute_graph.flexml_layers[184].compute_node[0][0].out[0], compute_graph.flexml_layers[185].compute_node[0][0].out[0], compute_graph.flexml_layers[186].compute_node[0][0].out[0], compute_graph.flexml_layers[187].compute_node[0][0].out[0], compute_graph.flexml_layers[188].compute_node[0][0].out[0], compute_graph.flexml_layers[189].compute_node[0][0].out[0], compute_graph.flexml_layers[190].compute_node[0][0].out[0], compute_graph.flexml_layers[191].compute_node[0][0].out[0], compute_graph.flexml_layers[192].compute_node[0][0].out[0], compute_graph.flexml_layers[193].compute_node[0][0].out[0], compute_graph.flexml_layers[194].compute_node[0][0].out[0], compute_graph.flexml_layers[195].compute_node[0][0].out[0], compute_graph.flexml_layers[196].compute_node[0][0].out[0], compute_graph.flexml_layers[197].compute_node[0][0].out[0], compute_graph.flexml_layers[198].compute_node[0][0].out[0], compute_graph.flexml_layers[199].compute_node[0][0].out[0], compute_graph.flexml_layers[200].compute_node[0][0].out[0], compute_graph.flexml_layers[203].compute_node[0][0].out[0], compute_graph.flexml_layers[204].compute_node[0][0].out[0], compute_graph.flexml_layers[205].compute_node[0][0].out[0], compute_graph.flexml_layers[206].compute_node[0][0].out[0], compute_graph.flexml_layers[207].compute_node[0][0].out[0], compute_graph.flexml_layers[208].compute_node[0][0].out[0], compute_graph.flexml_layers[209].compute_node[0][0].out[0], compute_graph.flexml_layers[210].compute_node[0][0].out[0], compute_graph.flexml_layers[211].compute_node[0][0].out[0], compute_graph.flexml_layers[212].compute_node[0][0].out[0], compute_graph.flexml_layers[213].compute_node[0][0].out[0], compute_graph.flexml_layers[214].compute_node[0][0].out[0], compute_graph.flexml_layers[215].compute_node[0][0].out[0], compute_graph.flexml_layers[216].compute_node[0][0].out[0], compute_graph.flexml_layers[217].compute_node[0][0].out[0], compute_graph.flexml_layers[218].compute_node[0][0].out[0], compute_graph.flexml_layers[219].compute_node[0][0].out[0], compute_graph.flexml_layers[220].compute_node[0][0].out[0], compute_graph.flexml_layers[221].compute_node[0][0].out[0], compute_graph.flexml_layers[222].compute_node[0][0].out[0], compute_graph.flexml_layers[223].compute_node[0][0].out[0], compute_graph.flexml_layers[224].compute_node[0][0].out[0], compute_graph.flexml_layers[225].compute_node[0][0].out[0], compute_graph.flexml_layers[226].compute_node[0][0].out[0], compute_graph.flexml_layers[227].compute_node[0][0].out[0], compute_graph.flexml_layers[228].compute_node[0][0].out[0], compute_graph.flexml_layers[229].compute_node[0][0].out[0], compute_graph.flexml_layers[230].compute_node[0][0].out[0], compute_graph.flexml_layers[231].compute_node[0][0].out[0], compute_graph.flexml_layers[232].compute_node[0][0].out[0], compute_graph.flexml_layers[233].compute_node[0][0].out[0], compute_graph.flexml_layers[234].compute_node[0][0].out[0], compute_graph.flexml_layers[235].compute_node[0][0].out[0], compute_graph.flexml_layers[236].compute_node[0][0].out[0]) + +Column 0Row 0Channel 1 + +i4_pi1(compute_graph.flexml_layers[36].compute_node[0][0].in[1], compute_graph.templated_graph_260.mk[0][0].in[1], compute_graph.templated_graph_268.mk[0][0].in[1], compute_graph.templated_graph_273.mk[0][0].in[1], compute_graph.flexml_layers[3].compute_node[0][0].in[1], compute_graph.flexml_layers[8].compute_node[0][0].in[1], compute_graph.flexml_layers[9].compute_node[0][0].in[1], compute_graph.flexml_layers[10].compute_node[0][0].in[1], compute_graph.flexml_layers[11].compute_node[0][0].in[1], compute_graph.flexml_layers[12].compute_node[0][0].in[1], compute_graph.flexml_layers[13].compute_node[0][0].in[1], compute_graph.flexml_layers[14].compute_node[0][0].in[1], compute_graph.flexml_layers[15].compute_node[0][0].in[1], compute_graph.flexml_layers[16].compute_node[0][0].in[1], compute_graph.flexml_layers[17].compute_node[0][0].in[1], compute_graph.flexml_layers[19].compute_node[0][0].in[1], compute_graph.flexml_layers[20].compute_node[0][0].in[1], compute_graph.flexml_layers[25].compute_node[0][0].in[1], compute_graph.flexml_layers[26].compute_node[0][0].in[1], compute_graph.flexml_layers[27].compute_node[0][0].in[1], compute_graph.flexml_layers[29].compute_node[0][0].in[1], compute_graph.flexml_layers[30].compute_node[0][0].in[1], compute_graph.flexml_layers[35].compute_node[0][0].in[1], compute_graph.flexml_layers[144].compute_node[0][0].in[1], compute_graph.flexml_layers[37].compute_node[0][0].in[1], compute_graph.flexml_layers[39].compute_node[0][0].in[1], compute_graph.flexml_layers[40].compute_node[0][0].in[1], compute_graph.flexml_layers[45].compute_node[0][0].in[1], compute_graph.flexml_layers[46].compute_node[0][0].in[1], compute_graph.flexml_layers[51].compute_node[0][0].in[1], compute_graph.flexml_layers[56].compute_node[0][0].in[1], compute_graph.flexml_layers[57].compute_node[0][0].in[1], compute_graph.flexml_layers[62].compute_node[0][0].in[1], compute_graph.flexml_layers[201].compute_node[0][0].in[1], compute_graph.flexml_layers[202].compute_node[0][0].in[1], compute_graph.flexml_layers[67].compute_node[0][0].in[1], compute_graph.flexml_layers[68].compute_node[0][0].in[1], compute_graph.flexml_layers[73].compute_node[0][0].in[1], compute_graph.flexml_layers[78].compute_node[0][0].in[1], compute_graph.flexml_layers[79].compute_node[0][0].in[1], compute_graph.flexml_layers[84].compute_node[0][0].in[1], compute_graph.flexml_layers[89].compute_node[0][0].in[1], compute_graph.flexml_layers[90].compute_node[0][0].in[1], compute_graph.flexml_layers[95].compute_node[0][0].in[1], compute_graph.flexml_layers[101].compute_node[0][0].in[1], compute_graph.flexml_layers[102].compute_node[0][0].in[1], compute_graph.flexml_layers[107].compute_node[0][0].in[1], compute_graph.flexml_layers[108].compute_node[0][0].in[1], compute_graph.flexml_layers[113].compute_node[0][0].in[1], compute_graph.flexml_layers[119].compute_node[0][0].in[1], compute_graph.flexml_layers[120].compute_node[0][0].in[1], compute_graph.flexml_layers[125].compute_node[0][0].in[1], compute_graph.flexml_layers[126].compute_node[0][0].in[1], compute_graph.flexml_layers[131].compute_node[0][0].in[1], compute_graph.flexml_layers[137].compute_node[0][0].in[1], compute_graph.flexml_layers[138].compute_node[0][0].in[1], compute_graph.flexml_layers[143].compute_node[0][0].in[1], compute_graph.flexml_layers[149].compute_node[0][0].in[1], compute_graph.flexml_layers[155].compute_node[0][0].in[1], compute_graph.flexml_layers[156].compute_node[0][0].in[1], compute_graph.flexml_layers[161].compute_node[0][0].in[1], compute_graph.flexml_layers[162].compute_node[0][0].in[1], compute_graph.flexml_layers[167].compute_node[0][0].in[1], compute_graph.flexml_layers[173].compute_node[0][0].in[1], compute_graph.flexml_layers[174].compute_node[0][0].in[1], compute_graph.flexml_layers[179].compute_node[0][0].in[1], compute_graph.flexml_layers[180].compute_node[0][0].in[1], compute_graph.flexml_layers[186].compute_node[0][0].in[1], compute_graph.flexml_layers[188].compute_node[0][0].in[1], compute_graph.flexml_layers[189].compute_node[0][0].in[1], compute_graph.flexml_layers[194].compute_node[0][0].in[1], compute_graph.flexml_layers[207].compute_node[0][0].in[1], compute_graph.flexml_layers[211].compute_node[0][0].in[1], compute_graph.flexml_layers[212].compute_node[0][0].in[1], compute_graph.flexml_layers[217].compute_node[0][0].in[1], compute_graph.flexml_layers[221].compute_node[0][0].in[1], compute_graph.flexml_layers[222].compute_node[0][0].in[1], compute_graph.flexml_layers[227].compute_node[0][0].in[1], compute_graph.flexml_layers[231].compute_node[0][0].in[1], compute_graph.flexml_layers[232].compute_node[0][0].in[1], compute_graph.flexml_layers[233].compute_node[0][0].in[1]) + +Column 0Row 1Channel 0 + +i5_pi0(compute_graph.templated_graph_4.trans_comp_nd[0][1].in[0], compute_graph.flexml_layers[36].compute_node[0][1].in[0], compute_graph.templated_graph_231.mk[0][1].in[0], compute_graph.templated_graph_254.mk[0][1].in[0], compute_graph.templated_graph_258.compute_node[0][1].in[0], compute_graph.templated_graph_259.compute_node[0][1].in[0], compute_graph.templated_graph_260.mk[0][1].in[0], compute_graph.templated_graph_263.compute_node[0][1].in[0], compute_graph.templated_graph_266.compute_node[0][1].in[0], compute_graph.templated_graph_268.mk[0][1].in[0], compute_graph.templated_graph_273.mk[0][1].in[0], compute_graph.templated_graph_274.mk[0][1].in[0], compute_graph.flexml_layers[63].compute_node[0][1].in[0], compute_graph.flexml_layers[64].compute_node[0][1].in[0], compute_graph.templated_graph_293.mk[0][1].in[0], compute_graph.templated_graph_298.compute_node[0][1].in[0], compute_graph.templated_graph_300.compute_node[0][1].in[0], compute_graph.flexml_layers[0].compute_node[0][1].in[0], compute_graph.flexml_layers[1].compute_node[0][1].in[0], compute_graph.flexml_layers[1].compute_node[0][1].in[2], compute_graph.flexml_layers[2].compute_node[0][1].in[0], compute_graph.flexml_layers[2].compute_node[0][1].in[2], compute_graph.flexml_layers[3].compute_node[0][1].in[0], compute_graph.flexml_layers[4].compute_node[0][1].in[0], compute_graph.flexml_layers[5].compute_node[0][1].in[0], compute_graph.flexml_layers[6].compute_node[0][1].in[0], compute_graph.flexml_layers[7].compute_node[0][1].in[0], compute_graph.flexml_layers[7].compute_node[0][1].in[2], compute_graph.flexml_layers[8].compute_node[0][1].in[0], compute_graph.flexml_layers[9].compute_node[0][1].in[0], compute_graph.flexml_layers[9].compute_node[0][1].in[2], compute_graph.flexml_layers[10].compute_node[0][1].in[0], compute_graph.flexml_layers[11].compute_node[0][1].in[0], compute_graph.flexml_layers[12].compute_node[0][1].in[0], compute_graph.flexml_layers[13].compute_node[0][1].in[0], compute_graph.flexml_layers[14].compute_node[0][1].in[0], compute_graph.flexml_layers[15].compute_node[0][1].in[0], compute_graph.flexml_layers[15].compute_node[0][1].in[2], compute_graph.flexml_layers[16].compute_node[0][1].in[0], compute_graph.flexml_layers[17].compute_node[0][1].in[0], compute_graph.flexml_layers[18].compute_node[0][1].in[0], compute_graph.flexml_layers[19].compute_node[0][1].in[0], compute_graph.flexml_layers[20].compute_node[0][1].in[0], compute_graph.flexml_layers[21].compute_node[0][1].in[0], compute_graph.flexml_layers[22].compute_node[0][1].in[0], compute_graph.flexml_layers[23].compute_node[0][1].in[0], compute_graph.flexml_layers[24].compute_node[0][1].in[0], compute_graph.flexml_layers[24].compute_node[0][1].in[2], compute_graph.flexml_layers[25].compute_node[0][1].in[0], compute_graph.flexml_layers[26].compute_node[0][1].in[0], compute_graph.flexml_layers[27].compute_node[0][1].in[0], compute_graph.flexml_layers[28].compute_node[0][1].in[0], compute_graph.flexml_layers[29].compute_node[0][1].in[0], compute_graph.flexml_layers[30].compute_node[0][1].in[0], compute_graph.flexml_layers[31].compute_node[0][1].in[0], compute_graph.flexml_layers[32].compute_node[0][1].in[0], compute_graph.flexml_layers[33].compute_node[0][1].in[0], compute_graph.flexml_layers[34].compute_node[0][1].in[0], compute_graph.flexml_layers[34].compute_node[0][1].in[2], compute_graph.flexml_layers[35].compute_node[0][1].in[0], compute_graph.flexml_layers[35].compute_node[0][1].in[2], compute_graph.flexml_layers[144].compute_node[0][1].in[0], compute_graph.flexml_layers[145].compute_node[0][1].in[0], compute_graph.flexml_layers[37].compute_node[0][1].in[0], compute_graph.flexml_layers[38].compute_node[0][1].in[0], compute_graph.flexml_layers[39].compute_node[0][1].in[0], compute_graph.flexml_layers[40].compute_node[0][1].in[0], compute_graph.flexml_layers[41].compute_node[0][1].in[0], compute_graph.flexml_layers[42].compute_node[0][1].in[0], compute_graph.flexml_layers[43].compute_node[0][1].in[0], compute_graph.flexml_layers[44].compute_node[0][1].in[0], compute_graph.flexml_layers[44].compute_node[0][1].in[2], compute_graph.flexml_layers[45].compute_node[0][1].in[0], compute_graph.flexml_layers[45].compute_node[0][1].in[2], compute_graph.flexml_layers[46].compute_node[0][1].in[0], compute_graph.flexml_layers[47].compute_node[0][1].in[0], compute_graph.flexml_layers[48].compute_node[0][1].in[0], compute_graph.flexml_layers[49].compute_node[0][1].in[0], compute_graph.flexml_layers[50].compute_node[0][1].in[0], compute_graph.flexml_layers[50].compute_node[0][1].in[2], compute_graph.flexml_layers[51].compute_node[0][1].in[0], compute_graph.flexml_layers[52].compute_node[0][1].in[0], compute_graph.flexml_layers[53].compute_node[0][1].in[0], compute_graph.flexml_layers[54].compute_node[0][1].in[0], compute_graph.flexml_layers[55].compute_node[0][1].in[0], compute_graph.flexml_layers[55].compute_node[0][1].in[2], compute_graph.flexml_layers[56].compute_node[0][1].in[0], compute_graph.flexml_layers[57].compute_node[0][1].in[0], compute_graph.flexml_layers[58].compute_node[0][1].in[0], compute_graph.flexml_layers[59].compute_node[0][1].in[0], compute_graph.flexml_layers[60].compute_node[0][1].in[0], compute_graph.flexml_layers[61].compute_node[0][1].in[0], compute_graph.flexml_layers[61].compute_node[0][1].in[2], compute_graph.flexml_layers[62].compute_node[0][1].in[0], compute_graph.flexml_layers[201].compute_node[0][1].in[0], compute_graph.flexml_layers[202].compute_node[0][1].in[0], compute_graph.flexml_layers[65].compute_node[0][1].in[0], compute_graph.flexml_layers[66].compute_node[0][1].in[0], compute_graph.flexml_layers[66].compute_node[0][1].in[2], compute_graph.flexml_layers[67].compute_node[0][1].in[0], compute_graph.flexml_layers[67].compute_node[0][1].in[2], compute_graph.flexml_layers[68].compute_node[0][1].in[0], compute_graph.flexml_layers[69].compute_node[0][1].in[0], compute_graph.flexml_layers[70].compute_node[0][1].in[0], compute_graph.flexml_layers[71].compute_node[0][1].in[0], compute_graph.flexml_layers[72].compute_node[0][1].in[0], compute_graph.flexml_layers[72].compute_node[0][1].in[2], compute_graph.flexml_layers[73].compute_node[0][1].in[0], compute_graph.flexml_layers[74].compute_node[0][1].in[0], compute_graph.flexml_layers[75].compute_node[0][1].in[0], compute_graph.flexml_layers[76].compute_node[0][1].in[0], compute_graph.flexml_layers[77].compute_node[0][1].in[0], compute_graph.flexml_layers[77].compute_node[0][1].in[2], compute_graph.flexml_layers[78].compute_node[0][1].in[0], compute_graph.flexml_layers[78].compute_node[0][1].in[2], compute_graph.flexml_layers[79].compute_node[0][1].in[0], compute_graph.flexml_layers[80].compute_node[0][1].in[0], compute_graph.flexml_layers[81].compute_node[0][1].in[0], compute_graph.flexml_layers[82].compute_node[0][1].in[0], compute_graph.flexml_layers[83].compute_node[0][1].in[0], compute_graph.flexml_layers[83].compute_node[0][1].in[2], compute_graph.flexml_layers[84].compute_node[0][1].in[0], compute_graph.flexml_layers[85].compute_node[0][1].in[0], compute_graph.flexml_layers[86].compute_node[0][1].in[0], compute_graph.flexml_layers[87].compute_node[0][1].in[0], compute_graph.flexml_layers[88].compute_node[0][1].in[0], compute_graph.flexml_layers[88].compute_node[0][1].in[2], compute_graph.flexml_layers[89].compute_node[0][1].in[0], compute_graph.flexml_layers[89].compute_node[0][1].in[2], compute_graph.flexml_layers[90].compute_node[0][1].in[0], compute_graph.flexml_layers[91].compute_node[0][1].in[0], compute_graph.flexml_layers[92].compute_node[0][1].in[0], compute_graph.flexml_layers[93].compute_node[0][1].in[0], compute_graph.flexml_layers[94].compute_node[0][1].in[0], compute_graph.flexml_layers[94].compute_node[0][1].in[2], compute_graph.flexml_layers[95].compute_node[0][1].in[0], compute_graph.flexml_layers[96].compute_node[0][1].in[0], compute_graph.flexml_layers[97].compute_node[0][1].in[0], compute_graph.flexml_layers[98].compute_node[0][1].in[0], compute_graph.flexml_layers[99].compute_node[0][1].in[0], compute_graph.flexml_layers[99].compute_node[0][1].in[2], compute_graph.flexml_layers[100].compute_node[0][1].in[0], compute_graph.flexml_layers[101].compute_node[0][1].in[0], compute_graph.flexml_layers[102].compute_node[0][1].in[0], compute_graph.flexml_layers[103].compute_node[0][1].in[0], compute_graph.flexml_layers[104].compute_node[0][1].in[0], compute_graph.flexml_layers[105].compute_node[0][1].in[0], compute_graph.flexml_layers[106].compute_node[0][1].in[0], compute_graph.flexml_layers[106].compute_node[0][1].in[2], compute_graph.flexml_layers[107].compute_node[0][1].in[0], compute_graph.flexml_layers[108].compute_node[0][1].in[0], compute_graph.flexml_layers[109].compute_node[0][1].in[0], compute_graph.flexml_layers[110].compute_node[0][1].in[0], compute_graph.flexml_layers[111].compute_node[0][1].in[0], compute_graph.flexml_layers[112].compute_node[0][1].in[0], compute_graph.flexml_layers[112].compute_node[0][1].in[2], compute_graph.flexml_layers[113].compute_node[0][1].in[0], compute_graph.flexml_layers[114].compute_node[0][1].in[0], compute_graph.flexml_layers[115].compute_node[0][1].in[0], compute_graph.flexml_layers[116].compute_node[0][1].in[0], compute_graph.flexml_layers[117].compute_node[0][1].in[0], compute_graph.flexml_layers[117].compute_node[0][1].in[2], compute_graph.flexml_layers[118].compute_node[0][1].in[0], compute_graph.flexml_layers[119].compute_node[0][1].in[0], compute_graph.flexml_layers[120].compute_node[0][1].in[0], compute_graph.flexml_layers[121].compute_node[0][1].in[0], compute_graph.flexml_layers[122].compute_node[0][1].in[0], compute_graph.flexml_layers[123].compute_node[0][1].in[0], compute_graph.flexml_layers[124].compute_node[0][1].in[0], compute_graph.flexml_layers[124].compute_node[0][1].in[2], compute_graph.flexml_layers[125].compute_node[0][1].in[0], compute_graph.flexml_layers[125].compute_node[0][1].in[2], compute_graph.flexml_layers[126].compute_node[0][1].in[0], compute_graph.flexml_layers[127].compute_node[0][1].in[0], compute_graph.flexml_layers[128].compute_node[0][1].in[0], compute_graph.flexml_layers[129].compute_node[0][1].in[0], compute_graph.flexml_layers[130].compute_node[0][1].in[0], compute_graph.flexml_layers[130].compute_node[0][1].in[2], compute_graph.flexml_layers[131].compute_node[0][1].in[0], compute_graph.flexml_layers[132].compute_node[0][1].in[0], compute_graph.flexml_layers[133].compute_node[0][1].in[0], compute_graph.flexml_layers[134].compute_node[0][1].in[0], compute_graph.flexml_layers[135].compute_node[0][1].in[0], compute_graph.flexml_layers[135].compute_node[0][1].in[2], compute_graph.flexml_layers[136].compute_node[0][1].in[0], compute_graph.flexml_layers[137].compute_node[0][1].in[0], compute_graph.flexml_layers[138].compute_node[0][1].in[0], compute_graph.flexml_layers[139].compute_node[0][1].in[0], compute_graph.flexml_layers[140].compute_node[0][1].in[0], compute_graph.flexml_layers[141].compute_node[0][1].in[0], compute_graph.flexml_layers[142].compute_node[0][1].in[0], compute_graph.flexml_layers[142].compute_node[0][1].in[2], compute_graph.flexml_layers[143].compute_node[0][1].in[0], compute_graph.flexml_layers[146].compute_node[0][1].in[0], compute_graph.flexml_layers[147].compute_node[0][1].in[0], compute_graph.flexml_layers[148].compute_node[0][1].in[0], compute_graph.flexml_layers[148].compute_node[0][1].in[2], compute_graph.flexml_layers[149].compute_node[0][1].in[0], compute_graph.flexml_layers[150].compute_node[0][1].in[0], compute_graph.flexml_layers[151].compute_node[0][1].in[0], compute_graph.flexml_layers[152].compute_node[0][1].in[0], compute_graph.flexml_layers[153].compute_node[0][1].in[0], compute_graph.flexml_layers[153].compute_node[0][1].in[2], compute_graph.flexml_layers[154].compute_node[0][1].in[0], compute_graph.flexml_layers[155].compute_node[0][1].in[0], compute_graph.flexml_layers[156].compute_node[0][1].in[0], compute_graph.flexml_layers[157].compute_node[0][1].in[0], compute_graph.flexml_layers[158].compute_node[0][1].in[0], compute_graph.flexml_layers[159].compute_node[0][1].in[0], compute_graph.flexml_layers[160].compute_node[0][1].in[0], compute_graph.flexml_layers[160].compute_node[0][1].in[2], compute_graph.flexml_layers[161].compute_node[0][1].in[0], compute_graph.flexml_layers[161].compute_node[0][1].in[2], compute_graph.flexml_layers[162].compute_node[0][1].in[0], compute_graph.flexml_layers[163].compute_node[0][1].in[0], compute_graph.flexml_layers[164].compute_node[0][1].in[0], compute_graph.flexml_layers[165].compute_node[0][1].in[0], compute_graph.flexml_layers[166].compute_node[0][1].in[0], compute_graph.flexml_layers[166].compute_node[0][1].in[2], compute_graph.flexml_layers[167].compute_node[0][1].in[0], compute_graph.flexml_layers[168].compute_node[0][1].in[0], compute_graph.flexml_layers[169].compute_node[0][1].in[0], compute_graph.flexml_layers[170].compute_node[0][1].in[0], compute_graph.flexml_layers[171].compute_node[0][1].in[0], compute_graph.flexml_layers[171].compute_node[0][1].in[2], compute_graph.flexml_layers[172].compute_node[0][1].in[0], compute_graph.flexml_layers[173].compute_node[0][1].in[0], compute_graph.flexml_layers[174].compute_node[0][1].in[0], compute_graph.flexml_layers[175].compute_node[0][1].in[0], compute_graph.flexml_layers[176].compute_node[0][1].in[0], compute_graph.flexml_layers[177].compute_node[0][1].in[0], compute_graph.flexml_layers[178].compute_node[0][1].in[0], compute_graph.flexml_layers[178].compute_node[0][1].in[2], compute_graph.flexml_layers[179].compute_node[0][1].in[0], compute_graph.flexml_layers[179].compute_node[0][1].in[2], compute_graph.flexml_layers[180].compute_node[0][1].in[0], compute_graph.flexml_layers[181].compute_node[0][1].in[0], compute_graph.flexml_layers[182].compute_node[0][1].in[0], compute_graph.flexml_layers[183].compute_node[0][1].in[0], compute_graph.flexml_layers[184].compute_node[0][1].in[0], compute_graph.flexml_layers[184].compute_node[0][1].in[2], compute_graph.flexml_layers[185].compute_node[0][1].in[0], compute_graph.flexml_layers[186].compute_node[0][1].in[0], compute_graph.flexml_layers[187].compute_node[0][1].in[0], compute_graph.flexml_layers[188].compute_node[0][1].in[0], compute_graph.flexml_layers[188].compute_node[0][1].in[2], compute_graph.flexml_layers[189].compute_node[0][1].in[0], compute_graph.flexml_layers[190].compute_node[0][1].in[0], compute_graph.flexml_layers[191].compute_node[0][1].in[0], compute_graph.flexml_layers[191].compute_node[0][1].in[2], compute_graph.flexml_layers[192].compute_node[0][1].in[0], compute_graph.flexml_layers[192].compute_node[0][1].in[2], compute_graph.flexml_layers[193].compute_node[0][1].in[0], compute_graph.flexml_layers[193].compute_node[0][1].in[2], compute_graph.flexml_layers[194].compute_node[0][1].in[0], compute_graph.flexml_layers[195].compute_node[0][1].in[0], compute_graph.flexml_layers[196].compute_node[0][1].in[0], compute_graph.flexml_layers[196].compute_node[0][1].in[2], compute_graph.flexml_layers[197].compute_node[0][1].in[0], compute_graph.flexml_layers[197].compute_node[0][1].in[2], compute_graph.flexml_layers[198].compute_node[0][1].in[0], compute_graph.flexml_layers[199].compute_node[0][1].in[0], compute_graph.flexml_layers[200].compute_node[0][1].in[0], compute_graph.flexml_layers[203].compute_node[0][1].in[0], compute_graph.flexml_layers[204].compute_node[0][1].in[0], compute_graph.flexml_layers[204].compute_node[0][1].in[2], compute_graph.flexml_layers[205].compute_node[0][1].in[0], compute_graph.flexml_layers[205].compute_node[0][1].in[2], compute_graph.flexml_layers[206].compute_node[0][1].in[0], compute_graph.flexml_layers[206].compute_node[0][1].in[2], compute_graph.flexml_layers[207].compute_node[0][1].in[0], compute_graph.flexml_layers[208].compute_node[0][1].in[0], compute_graph.flexml_layers[209].compute_node[0][1].in[0], compute_graph.flexml_layers[209].compute_node[0][1].in[2], compute_graph.flexml_layers[210].compute_node[0][1].in[0], compute_graph.flexml_layers[210].compute_node[0][1].in[2], compute_graph.flexml_layers[211].compute_node[0][1].in[0], compute_graph.flexml_layers[212].compute_node[0][1].in[0], compute_graph.flexml_layers[213].compute_node[0][1].in[0], compute_graph.flexml_layers[214].compute_node[0][1].in[0], compute_graph.flexml_layers[214].compute_node[0][1].in[2], compute_graph.flexml_layers[215].compute_node[0][1].in[0], compute_graph.flexml_layers[215].compute_node[0][1].in[2], compute_graph.flexml_layers[216].compute_node[0][1].in[0], compute_graph.flexml_layers[216].compute_node[0][1].in[2], compute_graph.flexml_layers[217].compute_node[0][1].in[0], compute_graph.flexml_layers[218].compute_node[0][1].in[0], compute_graph.flexml_layers[219].compute_node[0][1].in[0], compute_graph.flexml_layers[219].compute_node[0][1].in[2], compute_graph.flexml_layers[220].compute_node[0][1].in[0], compute_graph.flexml_layers[220].compute_node[0][1].in[2], compute_graph.flexml_layers[221].compute_node[0][1].in[0], compute_graph.flexml_layers[222].compute_node[0][1].in[0], compute_graph.flexml_layers[223].compute_node[0][1].in[0], compute_graph.flexml_layers[224].compute_node[0][1].in[0], compute_graph.flexml_layers[224].compute_node[0][1].in[2], compute_graph.flexml_layers[225].compute_node[0][1].in[0], compute_graph.flexml_layers[225].compute_node[0][1].in[2], compute_graph.flexml_layers[226].compute_node[0][1].in[0], compute_graph.flexml_layers[226].compute_node[0][1].in[2], compute_graph.flexml_layers[227].compute_node[0][1].in[0], compute_graph.flexml_layers[228].compute_node[0][1].in[0], compute_graph.flexml_layers[229].compute_node[0][1].in[0], compute_graph.flexml_layers[229].compute_node[0][1].in[2], compute_graph.flexml_layers[230].compute_node[0][1].in[0], compute_graph.flexml_layers[230].compute_node[0][1].in[2], compute_graph.flexml_layers[231].compute_node[0][1].in[0], compute_graph.flexml_layers[232].compute_node[0][1].in[0], compute_graph.flexml_layers[233].compute_node[0][1].in[0], compute_graph.flexml_layers[234].compute_node[0][1].in[0], compute_graph.flexml_layers[235].compute_node[0][1].in[0], compute_graph.flexml_layers[235].compute_node[0][1].in[2], compute_graph.flexml_layers[236].compute_node[0][1].in[0]) + +Column 0Row 1Channel 0 + +i5_po0(compute_graph.templated_graph_4.trans_comp_nd[0][1].out[0], compute_graph.flexml_layers[36].compute_node[0][1].out[0], compute_graph.templated_graph_231.mk[0][1].out[0], compute_graph.templated_graph_254.mk[0][1].out[0], compute_graph.templated_graph_258.compute_node[0][1].out[0], compute_graph.templated_graph_259.compute_node[0][1].out[0], compute_graph.templated_graph_260.mk[0][1].out[0], compute_graph.templated_graph_263.compute_node[0][1].out[0], compute_graph.templated_graph_266.compute_node[0][1].out[0], compute_graph.templated_graph_268.mk[0][1].out[0], compute_graph.templated_graph_273.mk[0][1].out[0], compute_graph.templated_graph_274.mk[0][1].out[0], compute_graph.flexml_layers[63].compute_node[0][1].out[0], compute_graph.flexml_layers[64].compute_node[0][1].out[0], compute_graph.templated_graph_293.mk[0][1].out[0], compute_graph.templated_graph_298.compute_node[0][1].out[0], compute_graph.templated_graph_300.compute_node[0][1].out[0], compute_graph.flexml_layers[0].compute_node[0][1].out[0], compute_graph.flexml_layers[1].compute_node[0][1].out[0], compute_graph.flexml_layers[2].compute_node[0][1].out[0], compute_graph.flexml_layers[3].compute_node[0][1].out[0], compute_graph.flexml_layers[4].compute_node[0][1].out[0], compute_graph.flexml_layers[5].compute_node[0][1].out[0], compute_graph.flexml_layers[6].compute_node[0][1].out[0], compute_graph.flexml_layers[7].compute_node[0][1].out[0], compute_graph.flexml_layers[8].compute_node[0][1].out[0], compute_graph.flexml_layers[9].compute_node[0][1].out[0], compute_graph.flexml_layers[10].compute_node[0][1].out[0], compute_graph.flexml_layers[11].compute_node[0][1].out[0], compute_graph.flexml_layers[12].compute_node[0][1].out[0], compute_graph.flexml_layers[13].compute_node[0][1].out[0], compute_graph.flexml_layers[14].compute_node[0][1].out[0], compute_graph.flexml_layers[15].compute_node[0][1].out[0], compute_graph.flexml_layers[16].compute_node[0][1].out[0], compute_graph.flexml_layers[17].compute_node[0][1].out[0], compute_graph.flexml_layers[18].compute_node[0][1].out[0], compute_graph.flexml_layers[19].compute_node[0][1].out[0], compute_graph.flexml_layers[20].compute_node[0][1].out[0], compute_graph.flexml_layers[21].compute_node[0][1].out[0], compute_graph.flexml_layers[22].compute_node[0][1].out[0], compute_graph.flexml_layers[23].compute_node[0][1].out[0], compute_graph.flexml_layers[24].compute_node[0][1].out[0], compute_graph.flexml_layers[25].compute_node[0][1].out[0], compute_graph.flexml_layers[26].compute_node[0][1].out[0], compute_graph.flexml_layers[27].compute_node[0][1].out[0], compute_graph.flexml_layers[28].compute_node[0][1].out[0], compute_graph.flexml_layers[29].compute_node[0][1].out[0], compute_graph.flexml_layers[30].compute_node[0][1].out[0], compute_graph.flexml_layers[31].compute_node[0][1].out[0], compute_graph.flexml_layers[32].compute_node[0][1].out[0], compute_graph.flexml_layers[33].compute_node[0][1].out[0], compute_graph.flexml_layers[34].compute_node[0][1].out[0], compute_graph.flexml_layers[35].compute_node[0][1].out[0], compute_graph.flexml_layers[144].compute_node[0][1].out[0], compute_graph.flexml_layers[145].compute_node[0][1].out[0], compute_graph.flexml_layers[37].compute_node[0][1].out[0], compute_graph.flexml_layers[38].compute_node[0][1].out[0], compute_graph.flexml_layers[39].compute_node[0][1].out[0], compute_graph.flexml_layers[40].compute_node[0][1].out[0], compute_graph.flexml_layers[41].compute_node[0][1].out[0], compute_graph.flexml_layers[42].compute_node[0][1].out[0], compute_graph.flexml_layers[43].compute_node[0][1].out[0], compute_graph.flexml_layers[44].compute_node[0][1].out[0], compute_graph.flexml_layers[45].compute_node[0][1].out[0], compute_graph.flexml_layers[46].compute_node[0][1].out[0], compute_graph.flexml_layers[47].compute_node[0][1].out[0], compute_graph.flexml_layers[48].compute_node[0][1].out[0], compute_graph.flexml_layers[49].compute_node[0][1].out[0], compute_graph.flexml_layers[50].compute_node[0][1].out[0], compute_graph.flexml_layers[51].compute_node[0][1].out[0], compute_graph.flexml_layers[52].compute_node[0][1].out[0], compute_graph.flexml_layers[53].compute_node[0][1].out[0], compute_graph.flexml_layers[54].compute_node[0][1].out[0], compute_graph.flexml_layers[55].compute_node[0][1].out[0], compute_graph.flexml_layers[56].compute_node[0][1].out[0], compute_graph.flexml_layers[57].compute_node[0][1].out[0], compute_graph.flexml_layers[58].compute_node[0][1].out[0], compute_graph.flexml_layers[59].compute_node[0][1].out[0], compute_graph.flexml_layers[60].compute_node[0][1].out[0], compute_graph.flexml_layers[61].compute_node[0][1].out[0], compute_graph.flexml_layers[62].compute_node[0][1].out[0], compute_graph.flexml_layers[201].compute_node[0][1].out[0], compute_graph.flexml_layers[202].compute_node[0][1].out[0], compute_graph.flexml_layers[65].compute_node[0][1].out[0], compute_graph.flexml_layers[66].compute_node[0][1].out[0], compute_graph.flexml_layers[67].compute_node[0][1].out[0], compute_graph.flexml_layers[68].compute_node[0][1].out[0], compute_graph.flexml_layers[69].compute_node[0][1].out[0], compute_graph.flexml_layers[70].compute_node[0][1].out[0], compute_graph.flexml_layers[71].compute_node[0][1].out[0], compute_graph.flexml_layers[72].compute_node[0][1].out[0], compute_graph.flexml_layers[73].compute_node[0][1].out[0], compute_graph.flexml_layers[74].compute_node[0][1].out[0], compute_graph.flexml_layers[75].compute_node[0][1].out[0], compute_graph.flexml_layers[76].compute_node[0][1].out[0], compute_graph.flexml_layers[77].compute_node[0][1].out[0], compute_graph.flexml_layers[78].compute_node[0][1].out[0], compute_graph.flexml_layers[79].compute_node[0][1].out[0], compute_graph.flexml_layers[80].compute_node[0][1].out[0], compute_graph.flexml_layers[81].compute_node[0][1].out[0], compute_graph.flexml_layers[82].compute_node[0][1].out[0], compute_graph.flexml_layers[83].compute_node[0][1].out[0], compute_graph.flexml_layers[84].compute_node[0][1].out[0], compute_graph.flexml_layers[85].compute_node[0][1].out[0], compute_graph.flexml_layers[86].compute_node[0][1].out[0], compute_graph.flexml_layers[87].compute_node[0][1].out[0], compute_graph.flexml_layers[88].compute_node[0][1].out[0], compute_graph.flexml_layers[89].compute_node[0][1].out[0], compute_graph.flexml_layers[90].compute_node[0][1].out[0], compute_graph.flexml_layers[91].compute_node[0][1].out[0], compute_graph.flexml_layers[92].compute_node[0][1].out[0], compute_graph.flexml_layers[93].compute_node[0][1].out[0], compute_graph.flexml_layers[94].compute_node[0][1].out[0], compute_graph.flexml_layers[95].compute_node[0][1].out[0], compute_graph.flexml_layers[96].compute_node[0][1].out[0], compute_graph.flexml_layers[97].compute_node[0][1].out[0], compute_graph.flexml_layers[98].compute_node[0][1].out[0], compute_graph.flexml_layers[99].compute_node[0][1].out[0], compute_graph.flexml_layers[100].compute_node[0][1].out[0], compute_graph.flexml_layers[101].compute_node[0][1].out[0], compute_graph.flexml_layers[102].compute_node[0][1].out[0], compute_graph.flexml_layers[103].compute_node[0][1].out[0], compute_graph.flexml_layers[104].compute_node[0][1].out[0], compute_graph.flexml_layers[105].compute_node[0][1].out[0], compute_graph.flexml_layers[106].compute_node[0][1].out[0], compute_graph.flexml_layers[107].compute_node[0][1].out[0], compute_graph.flexml_layers[108].compute_node[0][1].out[0], compute_graph.flexml_layers[109].compute_node[0][1].out[0], compute_graph.flexml_layers[110].compute_node[0][1].out[0], compute_graph.flexml_layers[111].compute_node[0][1].out[0], compute_graph.flexml_layers[112].compute_node[0][1].out[0], compute_graph.flexml_layers[113].compute_node[0][1].out[0], compute_graph.flexml_layers[114].compute_node[0][1].out[0], compute_graph.flexml_layers[115].compute_node[0][1].out[0], compute_graph.flexml_layers[116].compute_node[0][1].out[0], compute_graph.flexml_layers[117].compute_node[0][1].out[0], compute_graph.flexml_layers[118].compute_node[0][1].out[0], compute_graph.flexml_layers[119].compute_node[0][1].out[0], compute_graph.flexml_layers[120].compute_node[0][1].out[0], compute_graph.flexml_layers[121].compute_node[0][1].out[0], compute_graph.flexml_layers[122].compute_node[0][1].out[0], compute_graph.flexml_layers[123].compute_node[0][1].out[0], compute_graph.flexml_layers[124].compute_node[0][1].out[0], compute_graph.flexml_layers[125].compute_node[0][1].out[0], compute_graph.flexml_layers[126].compute_node[0][1].out[0], compute_graph.flexml_layers[127].compute_node[0][1].out[0], compute_graph.flexml_layers[128].compute_node[0][1].out[0], compute_graph.flexml_layers[129].compute_node[0][1].out[0], compute_graph.flexml_layers[130].compute_node[0][1].out[0], compute_graph.flexml_layers[131].compute_node[0][1].out[0], compute_graph.flexml_layers[132].compute_node[0][1].out[0], compute_graph.flexml_layers[133].compute_node[0][1].out[0], compute_graph.flexml_layers[134].compute_node[0][1].out[0], compute_graph.flexml_layers[135].compute_node[0][1].out[0], compute_graph.flexml_layers[136].compute_node[0][1].out[0], compute_graph.flexml_layers[137].compute_node[0][1].out[0], compute_graph.flexml_layers[138].compute_node[0][1].out[0], compute_graph.flexml_layers[139].compute_node[0][1].out[0], compute_graph.flexml_layers[140].compute_node[0][1].out[0], compute_graph.flexml_layers[141].compute_node[0][1].out[0], compute_graph.flexml_layers[142].compute_node[0][1].out[0], compute_graph.flexml_layers[143].compute_node[0][1].out[0], compute_graph.flexml_layers[146].compute_node[0][1].out[0], compute_graph.flexml_layers[147].compute_node[0][1].out[0], compute_graph.flexml_layers[148].compute_node[0][1].out[0], compute_graph.flexml_layers[149].compute_node[0][1].out[0], compute_graph.flexml_layers[150].compute_node[0][1].out[0], compute_graph.flexml_layers[151].compute_node[0][1].out[0], compute_graph.flexml_layers[152].compute_node[0][1].out[0], compute_graph.flexml_layers[153].compute_node[0][1].out[0], compute_graph.flexml_layers[154].compute_node[0][1].out[0], compute_graph.flexml_layers[155].compute_node[0][1].out[0], compute_graph.flexml_layers[156].compute_node[0][1].out[0], compute_graph.flexml_layers[157].compute_node[0][1].out[0], compute_graph.flexml_layers[158].compute_node[0][1].out[0], compute_graph.flexml_layers[159].compute_node[0][1].out[0], compute_graph.flexml_layers[160].compute_node[0][1].out[0], compute_graph.flexml_layers[161].compute_node[0][1].out[0], compute_graph.flexml_layers[162].compute_node[0][1].out[0], compute_graph.flexml_layers[163].compute_node[0][1].out[0], compute_graph.flexml_layers[164].compute_node[0][1].out[0], compute_graph.flexml_layers[165].compute_node[0][1].out[0], compute_graph.flexml_layers[166].compute_node[0][1].out[0], compute_graph.flexml_layers[167].compute_node[0][1].out[0], compute_graph.flexml_layers[168].compute_node[0][1].out[0], compute_graph.flexml_layers[169].compute_node[0][1].out[0], compute_graph.flexml_layers[170].compute_node[0][1].out[0], compute_graph.flexml_layers[171].compute_node[0][1].out[0], compute_graph.flexml_layers[172].compute_node[0][1].out[0], compute_graph.flexml_layers[173].compute_node[0][1].out[0], compute_graph.flexml_layers[174].compute_node[0][1].out[0], compute_graph.flexml_layers[175].compute_node[0][1].out[0], compute_graph.flexml_layers[176].compute_node[0][1].out[0], compute_graph.flexml_layers[177].compute_node[0][1].out[0], compute_graph.flexml_layers[178].compute_node[0][1].out[0], compute_graph.flexml_layers[179].compute_node[0][1].out[0], compute_graph.flexml_layers[180].compute_node[0][1].out[0], compute_graph.flexml_layers[181].compute_node[0][1].out[0], compute_graph.flexml_layers[182].compute_node[0][1].out[0], compute_graph.flexml_layers[183].compute_node[0][1].out[0], compute_graph.flexml_layers[184].compute_node[0][1].out[0], compute_graph.flexml_layers[185].compute_node[0][1].out[0], compute_graph.flexml_layers[186].compute_node[0][1].out[0], compute_graph.flexml_layers[187].compute_node[0][1].out[0], compute_graph.flexml_layers[188].compute_node[0][1].out[0], compute_graph.flexml_layers[189].compute_node[0][1].out[0], compute_graph.flexml_layers[190].compute_node[0][1].out[0], compute_graph.flexml_layers[191].compute_node[0][1].out[0], compute_graph.flexml_layers[192].compute_node[0][1].out[0], compute_graph.flexml_layers[193].compute_node[0][1].out[0], compute_graph.flexml_layers[194].compute_node[0][1].out[0], compute_graph.flexml_layers[195].compute_node[0][1].out[0], compute_graph.flexml_layers[196].compute_node[0][1].out[0], compute_graph.flexml_layers[197].compute_node[0][1].out[0], compute_graph.flexml_layers[198].compute_node[0][1].out[0], compute_graph.flexml_layers[199].compute_node[0][1].out[0], compute_graph.flexml_layers[200].compute_node[0][1].out[0], compute_graph.flexml_layers[203].compute_node[0][1].out[0], compute_graph.flexml_layers[204].compute_node[0][1].out[0], compute_graph.flexml_layers[205].compute_node[0][1].out[0], compute_graph.flexml_layers[206].compute_node[0][1].out[0], compute_graph.flexml_layers[207].compute_node[0][1].out[0], compute_graph.flexml_layers[208].compute_node[0][1].out[0], compute_graph.flexml_layers[209].compute_node[0][1].out[0], compute_graph.flexml_layers[210].compute_node[0][1].out[0], compute_graph.flexml_layers[211].compute_node[0][1].out[0], compute_graph.flexml_layers[212].compute_node[0][1].out[0], compute_graph.flexml_layers[213].compute_node[0][1].out[0], compute_graph.flexml_layers[214].compute_node[0][1].out[0], compute_graph.flexml_layers[215].compute_node[0][1].out[0], compute_graph.flexml_layers[216].compute_node[0][1].out[0], compute_graph.flexml_layers[217].compute_node[0][1].out[0], compute_graph.flexml_layers[218].compute_node[0][1].out[0], compute_graph.flexml_layers[219].compute_node[0][1].out[0], compute_graph.flexml_layers[220].compute_node[0][1].out[0], compute_graph.flexml_layers[221].compute_node[0][1].out[0], compute_graph.flexml_layers[222].compute_node[0][1].out[0], compute_graph.flexml_layers[223].compute_node[0][1].out[0], compute_graph.flexml_layers[224].compute_node[0][1].out[0], compute_graph.flexml_layers[225].compute_node[0][1].out[0], compute_graph.flexml_layers[226].compute_node[0][1].out[0], compute_graph.flexml_layers[227].compute_node[0][1].out[0], compute_graph.flexml_layers[228].compute_node[0][1].out[0], compute_graph.flexml_layers[229].compute_node[0][1].out[0], compute_graph.flexml_layers[230].compute_node[0][1].out[0], compute_graph.flexml_layers[231].compute_node[0][1].out[0], compute_graph.flexml_layers[232].compute_node[0][1].out[0], compute_graph.flexml_layers[233].compute_node[0][1].out[0], compute_graph.flexml_layers[234].compute_node[0][1].out[0], compute_graph.flexml_layers[235].compute_node[0][1].out[0], compute_graph.flexml_layers[236].compute_node[0][1].out[0]) + +Column 0Row 1Channel 1 + +i5_pi1(compute_graph.flexml_layers[36].compute_node[0][1].in[1], compute_graph.templated_graph_260.mk[0][1].in[1], compute_graph.templated_graph_268.mk[0][1].in[1], compute_graph.templated_graph_273.mk[0][1].in[1], compute_graph.flexml_layers[3].compute_node[0][1].in[1], compute_graph.flexml_layers[8].compute_node[0][1].in[1], compute_graph.flexml_layers[9].compute_node[0][1].in[1], compute_graph.flexml_layers[10].compute_node[0][1].in[1], compute_graph.flexml_layers[11].compute_node[0][1].in[1], compute_graph.flexml_layers[12].compute_node[0][1].in[1], compute_graph.flexml_layers[13].compute_node[0][1].in[1], compute_graph.flexml_layers[14].compute_node[0][1].in[1], compute_graph.flexml_layers[15].compute_node[0][1].in[1], compute_graph.flexml_layers[16].compute_node[0][1].in[1], compute_graph.flexml_layers[17].compute_node[0][1].in[1], compute_graph.flexml_layers[19].compute_node[0][1].in[1], compute_graph.flexml_layers[20].compute_node[0][1].in[1], compute_graph.flexml_layers[25].compute_node[0][1].in[1], compute_graph.flexml_layers[26].compute_node[0][1].in[1], compute_graph.flexml_layers[27].compute_node[0][1].in[1], compute_graph.flexml_layers[29].compute_node[0][1].in[1], compute_graph.flexml_layers[30].compute_node[0][1].in[1], compute_graph.flexml_layers[35].compute_node[0][1].in[1], compute_graph.flexml_layers[144].compute_node[0][1].in[1], compute_graph.flexml_layers[37].compute_node[0][1].in[1], compute_graph.flexml_layers[39].compute_node[0][1].in[1], compute_graph.flexml_layers[40].compute_node[0][1].in[1], compute_graph.flexml_layers[45].compute_node[0][1].in[1], compute_graph.flexml_layers[46].compute_node[0][1].in[1], compute_graph.flexml_layers[51].compute_node[0][1].in[1], compute_graph.flexml_layers[56].compute_node[0][1].in[1], compute_graph.flexml_layers[57].compute_node[0][1].in[1], compute_graph.flexml_layers[62].compute_node[0][1].in[1], compute_graph.flexml_layers[201].compute_node[0][1].in[1], compute_graph.flexml_layers[202].compute_node[0][1].in[1], compute_graph.flexml_layers[67].compute_node[0][1].in[1], compute_graph.flexml_layers[68].compute_node[0][1].in[1], compute_graph.flexml_layers[73].compute_node[0][1].in[1], compute_graph.flexml_layers[78].compute_node[0][1].in[1], compute_graph.flexml_layers[79].compute_node[0][1].in[1], compute_graph.flexml_layers[84].compute_node[0][1].in[1], compute_graph.flexml_layers[89].compute_node[0][1].in[1], compute_graph.flexml_layers[90].compute_node[0][1].in[1], compute_graph.flexml_layers[95].compute_node[0][1].in[1], compute_graph.flexml_layers[101].compute_node[0][1].in[1], compute_graph.flexml_layers[102].compute_node[0][1].in[1], compute_graph.flexml_layers[107].compute_node[0][1].in[1], compute_graph.flexml_layers[108].compute_node[0][1].in[1], compute_graph.flexml_layers[113].compute_node[0][1].in[1], compute_graph.flexml_layers[119].compute_node[0][1].in[1], compute_graph.flexml_layers[120].compute_node[0][1].in[1], compute_graph.flexml_layers[125].compute_node[0][1].in[1], compute_graph.flexml_layers[126].compute_node[0][1].in[1], compute_graph.flexml_layers[131].compute_node[0][1].in[1], compute_graph.flexml_layers[137].compute_node[0][1].in[1], compute_graph.flexml_layers[138].compute_node[0][1].in[1], compute_graph.flexml_layers[143].compute_node[0][1].in[1], compute_graph.flexml_layers[149].compute_node[0][1].in[1], compute_graph.flexml_layers[155].compute_node[0][1].in[1], compute_graph.flexml_layers[156].compute_node[0][1].in[1], compute_graph.flexml_layers[161].compute_node[0][1].in[1], compute_graph.flexml_layers[162].compute_node[0][1].in[1], compute_graph.flexml_layers[167].compute_node[0][1].in[1], compute_graph.flexml_layers[173].compute_node[0][1].in[1], compute_graph.flexml_layers[174].compute_node[0][1].in[1], compute_graph.flexml_layers[179].compute_node[0][1].in[1], compute_graph.flexml_layers[180].compute_node[0][1].in[1], compute_graph.flexml_layers[186].compute_node[0][1].in[1], compute_graph.flexml_layers[188].compute_node[0][1].in[1], compute_graph.flexml_layers[189].compute_node[0][1].in[1], compute_graph.flexml_layers[194].compute_node[0][1].in[1], compute_graph.flexml_layers[207].compute_node[0][1].in[1], compute_graph.flexml_layers[211].compute_node[0][1].in[1], compute_graph.flexml_layers[212].compute_node[0][1].in[1], compute_graph.flexml_layers[217].compute_node[0][1].in[1], compute_graph.flexml_layers[221].compute_node[0][1].in[1], compute_graph.flexml_layers[222].compute_node[0][1].in[1], compute_graph.flexml_layers[227].compute_node[0][1].in[1], compute_graph.flexml_layers[231].compute_node[0][1].in[1], compute_graph.flexml_layers[232].compute_node[0][1].in[1], compute_graph.flexml_layers[233].compute_node[0][1].in[1]) + +Column 0Row 2Channel 0 + +i6_pi0(compute_graph.templated_graph_4.trans_comp_nd[0][2].in[0], compute_graph.flexml_layers[36].compute_node[0][2].in[0], compute_graph.templated_graph_231.mk[0][2].in[0], compute_graph.templated_graph_254.mk[0][2].in[0], compute_graph.templated_graph_258.compute_node[0][2].in[0], compute_graph.templated_graph_259.compute_node[0][2].in[0], compute_graph.templated_graph_260.mk[0][2].in[0], compute_graph.templated_graph_263.compute_node[0][2].in[0], compute_graph.templated_graph_266.compute_node[0][2].in[0], compute_graph.templated_graph_268.mk[0][2].in[0], compute_graph.templated_graph_273.mk[0][2].in[0], compute_graph.templated_graph_274.mk[0][2].in[0], compute_graph.flexml_layers[63].compute_node[0][2].in[0], compute_graph.flexml_layers[64].compute_node[0][2].in[0], compute_graph.templated_graph_293.mk[0][2].in[0], compute_graph.templated_graph_298.compute_node[0][2].in[0], compute_graph.templated_graph_300.compute_node[0][2].in[0], compute_graph.flexml_layers[0].compute_node[0][2].in[0], compute_graph.flexml_layers[1].compute_node[0][2].in[0], compute_graph.flexml_layers[1].compute_node[0][2].in[2], compute_graph.flexml_layers[2].compute_node[0][2].in[0], compute_graph.flexml_layers[2].compute_node[0][2].in[2], compute_graph.flexml_layers[3].compute_node[0][2].in[0], compute_graph.flexml_layers[4].compute_node[0][2].in[0], compute_graph.flexml_layers[5].compute_node[0][2].in[0], compute_graph.flexml_layers[6].compute_node[0][2].in[0], compute_graph.flexml_layers[7].compute_node[0][2].in[0], compute_graph.flexml_layers[7].compute_node[0][2].in[2], compute_graph.flexml_layers[8].compute_node[0][2].in[0], compute_graph.flexml_layers[9].compute_node[0][2].in[0], compute_graph.flexml_layers[9].compute_node[0][2].in[2], compute_graph.flexml_layers[10].compute_node[0][2].in[0], compute_graph.flexml_layers[11].compute_node[0][2].in[0], compute_graph.flexml_layers[12].compute_node[0][2].in[0], compute_graph.flexml_layers[13].compute_node[0][2].in[0], compute_graph.flexml_layers[14].compute_node[0][2].in[0], compute_graph.flexml_layers[15].compute_node[0][2].in[0], compute_graph.flexml_layers[15].compute_node[0][2].in[2], compute_graph.flexml_layers[16].compute_node[0][2].in[0], compute_graph.flexml_layers[17].compute_node[0][2].in[0], compute_graph.flexml_layers[18].compute_node[0][2].in[0], compute_graph.flexml_layers[19].compute_node[0][2].in[0], compute_graph.flexml_layers[20].compute_node[0][2].in[0], compute_graph.flexml_layers[21].compute_node[0][2].in[0], compute_graph.flexml_layers[22].compute_node[0][2].in[0], compute_graph.flexml_layers[23].compute_node[0][2].in[0], compute_graph.flexml_layers[24].compute_node[0][2].in[0], compute_graph.flexml_layers[24].compute_node[0][2].in[2], compute_graph.flexml_layers[25].compute_node[0][2].in[0], compute_graph.flexml_layers[26].compute_node[0][2].in[0], compute_graph.flexml_layers[27].compute_node[0][2].in[0], compute_graph.flexml_layers[28].compute_node[0][2].in[0], compute_graph.flexml_layers[29].compute_node[0][2].in[0], compute_graph.flexml_layers[30].compute_node[0][2].in[0], compute_graph.flexml_layers[31].compute_node[0][2].in[0], compute_graph.flexml_layers[32].compute_node[0][2].in[0], compute_graph.flexml_layers[33].compute_node[0][2].in[0], compute_graph.flexml_layers[34].compute_node[0][2].in[0], compute_graph.flexml_layers[34].compute_node[0][2].in[2], compute_graph.flexml_layers[35].compute_node[0][2].in[0], compute_graph.flexml_layers[35].compute_node[0][2].in[2], compute_graph.flexml_layers[144].compute_node[0][2].in[0], compute_graph.flexml_layers[145].compute_node[0][2].in[0], compute_graph.flexml_layers[37].compute_node[0][2].in[0], compute_graph.flexml_layers[38].compute_node[0][2].in[0], compute_graph.flexml_layers[39].compute_node[0][2].in[0], compute_graph.flexml_layers[40].compute_node[0][2].in[0], compute_graph.flexml_layers[41].compute_node[0][2].in[0], compute_graph.flexml_layers[42].compute_node[0][2].in[0], compute_graph.flexml_layers[43].compute_node[0][2].in[0], compute_graph.flexml_layers[44].compute_node[0][2].in[0], compute_graph.flexml_layers[44].compute_node[0][2].in[2], compute_graph.flexml_layers[45].compute_node[0][2].in[0], compute_graph.flexml_layers[45].compute_node[0][2].in[2], compute_graph.flexml_layers[46].compute_node[0][2].in[0], compute_graph.flexml_layers[47].compute_node[0][2].in[0], compute_graph.flexml_layers[48].compute_node[0][2].in[0], compute_graph.flexml_layers[49].compute_node[0][2].in[0], compute_graph.flexml_layers[50].compute_node[0][2].in[0], compute_graph.flexml_layers[50].compute_node[0][2].in[2], compute_graph.flexml_layers[51].compute_node[0][2].in[0], compute_graph.flexml_layers[52].compute_node[0][2].in[0], compute_graph.flexml_layers[53].compute_node[0][2].in[0], compute_graph.flexml_layers[54].compute_node[0][2].in[0], compute_graph.flexml_layers[55].compute_node[0][2].in[0], compute_graph.flexml_layers[55].compute_node[0][2].in[2], compute_graph.flexml_layers[56].compute_node[0][2].in[0], compute_graph.flexml_layers[57].compute_node[0][2].in[0], compute_graph.flexml_layers[58].compute_node[0][2].in[0], compute_graph.flexml_layers[59].compute_node[0][2].in[0], compute_graph.flexml_layers[60].compute_node[0][2].in[0], compute_graph.flexml_layers[61].compute_node[0][2].in[0], compute_graph.flexml_layers[61].compute_node[0][2].in[2], compute_graph.flexml_layers[62].compute_node[0][2].in[0], compute_graph.flexml_layers[201].compute_node[0][2].in[0], compute_graph.flexml_layers[202].compute_node[0][2].in[0], compute_graph.flexml_layers[65].compute_node[0][2].in[0], compute_graph.flexml_layers[66].compute_node[0][2].in[0], compute_graph.flexml_layers[66].compute_node[0][2].in[2], compute_graph.flexml_layers[67].compute_node[0][2].in[0], compute_graph.flexml_layers[67].compute_node[0][2].in[2], compute_graph.flexml_layers[68].compute_node[0][2].in[0], compute_graph.flexml_layers[69].compute_node[0][2].in[0], compute_graph.flexml_layers[70].compute_node[0][2].in[0], compute_graph.flexml_layers[71].compute_node[0][2].in[0], compute_graph.flexml_layers[72].compute_node[0][2].in[0], compute_graph.flexml_layers[72].compute_node[0][2].in[2], compute_graph.flexml_layers[73].compute_node[0][2].in[0], compute_graph.flexml_layers[74].compute_node[0][2].in[0], compute_graph.flexml_layers[75].compute_node[0][2].in[0], compute_graph.flexml_layers[76].compute_node[0][2].in[0], compute_graph.flexml_layers[77].compute_node[0][2].in[0], compute_graph.flexml_layers[77].compute_node[0][2].in[2], compute_graph.flexml_layers[78].compute_node[0][2].in[0], compute_graph.flexml_layers[78].compute_node[0][2].in[2], compute_graph.flexml_layers[79].compute_node[0][2].in[0], compute_graph.flexml_layers[80].compute_node[0][2].in[0], compute_graph.flexml_layers[81].compute_node[0][2].in[0], compute_graph.flexml_layers[82].compute_node[0][2].in[0], compute_graph.flexml_layers[83].compute_node[0][2].in[0], compute_graph.flexml_layers[83].compute_node[0][2].in[2], compute_graph.flexml_layers[84].compute_node[0][2].in[0], compute_graph.flexml_layers[85].compute_node[0][2].in[0], compute_graph.flexml_layers[86].compute_node[0][2].in[0], compute_graph.flexml_layers[87].compute_node[0][2].in[0], compute_graph.flexml_layers[88].compute_node[0][2].in[0], compute_graph.flexml_layers[88].compute_node[0][2].in[2], compute_graph.flexml_layers[89].compute_node[0][2].in[0], compute_graph.flexml_layers[89].compute_node[0][2].in[2], compute_graph.flexml_layers[90].compute_node[0][2].in[0], compute_graph.flexml_layers[91].compute_node[0][2].in[0], compute_graph.flexml_layers[92].compute_node[0][2].in[0], compute_graph.flexml_layers[93].compute_node[0][2].in[0], compute_graph.flexml_layers[94].compute_node[0][2].in[0], compute_graph.flexml_layers[94].compute_node[0][2].in[2], compute_graph.flexml_layers[95].compute_node[0][2].in[0], compute_graph.flexml_layers[96].compute_node[0][2].in[0], compute_graph.flexml_layers[97].compute_node[0][2].in[0], compute_graph.flexml_layers[98].compute_node[0][2].in[0], compute_graph.flexml_layers[99].compute_node[0][2].in[0], compute_graph.flexml_layers[99].compute_node[0][2].in[2], compute_graph.flexml_layers[100].compute_node[0][2].in[0], compute_graph.flexml_layers[101].compute_node[0][2].in[0], compute_graph.flexml_layers[102].compute_node[0][2].in[0], compute_graph.flexml_layers[103].compute_node[0][2].in[0], compute_graph.flexml_layers[104].compute_node[0][2].in[0], compute_graph.flexml_layers[105].compute_node[0][2].in[0], compute_graph.flexml_layers[106].compute_node[0][2].in[0], compute_graph.flexml_layers[106].compute_node[0][2].in[2], compute_graph.flexml_layers[107].compute_node[0][2].in[0], compute_graph.flexml_layers[108].compute_node[0][2].in[0], compute_graph.flexml_layers[109].compute_node[0][2].in[0], compute_graph.flexml_layers[110].compute_node[0][2].in[0], compute_graph.flexml_layers[111].compute_node[0][2].in[0], compute_graph.flexml_layers[112].compute_node[0][2].in[0], compute_graph.flexml_layers[112].compute_node[0][2].in[2], compute_graph.flexml_layers[113].compute_node[0][2].in[0], compute_graph.flexml_layers[114].compute_node[0][2].in[0], compute_graph.flexml_layers[115].compute_node[0][2].in[0], compute_graph.flexml_layers[116].compute_node[0][2].in[0], compute_graph.flexml_layers[117].compute_node[0][2].in[0], compute_graph.flexml_layers[117].compute_node[0][2].in[2], compute_graph.flexml_layers[118].compute_node[0][2].in[0], compute_graph.flexml_layers[119].compute_node[0][2].in[0], compute_graph.flexml_layers[120].compute_node[0][2].in[0], compute_graph.flexml_layers[121].compute_node[0][2].in[0], compute_graph.flexml_layers[122].compute_node[0][2].in[0], compute_graph.flexml_layers[123].compute_node[0][2].in[0], compute_graph.flexml_layers[124].compute_node[0][2].in[0], compute_graph.flexml_layers[124].compute_node[0][2].in[2], compute_graph.flexml_layers[125].compute_node[0][2].in[0], compute_graph.flexml_layers[125].compute_node[0][2].in[2], compute_graph.flexml_layers[126].compute_node[0][2].in[0], compute_graph.flexml_layers[127].compute_node[0][2].in[0], compute_graph.flexml_layers[128].compute_node[0][2].in[0], compute_graph.flexml_layers[129].compute_node[0][2].in[0], compute_graph.flexml_layers[130].compute_node[0][2].in[0], compute_graph.flexml_layers[130].compute_node[0][2].in[2], compute_graph.flexml_layers[131].compute_node[0][2].in[0], compute_graph.flexml_layers[132].compute_node[0][2].in[0], compute_graph.flexml_layers[133].compute_node[0][2].in[0], compute_graph.flexml_layers[134].compute_node[0][2].in[0], compute_graph.flexml_layers[135].compute_node[0][2].in[0], compute_graph.flexml_layers[135].compute_node[0][2].in[2], compute_graph.flexml_layers[136].compute_node[0][2].in[0], compute_graph.flexml_layers[137].compute_node[0][2].in[0], compute_graph.flexml_layers[138].compute_node[0][2].in[0], compute_graph.flexml_layers[139].compute_node[0][2].in[0], compute_graph.flexml_layers[140].compute_node[0][2].in[0], compute_graph.flexml_layers[141].compute_node[0][2].in[0], compute_graph.flexml_layers[142].compute_node[0][2].in[0], compute_graph.flexml_layers[142].compute_node[0][2].in[2], compute_graph.flexml_layers[143].compute_node[0][2].in[0], compute_graph.flexml_layers[146].compute_node[0][2].in[0], compute_graph.flexml_layers[147].compute_node[0][2].in[0], compute_graph.flexml_layers[148].compute_node[0][2].in[0], compute_graph.flexml_layers[148].compute_node[0][2].in[2], compute_graph.flexml_layers[149].compute_node[0][2].in[0], compute_graph.flexml_layers[150].compute_node[0][2].in[0], compute_graph.flexml_layers[151].compute_node[0][2].in[0], compute_graph.flexml_layers[152].compute_node[0][2].in[0], compute_graph.flexml_layers[153].compute_node[0][2].in[0], compute_graph.flexml_layers[153].compute_node[0][2].in[2], compute_graph.flexml_layers[154].compute_node[0][2].in[0], compute_graph.flexml_layers[155].compute_node[0][2].in[0], compute_graph.flexml_layers[156].compute_node[0][2].in[0], compute_graph.flexml_layers[157].compute_node[0][2].in[0], compute_graph.flexml_layers[158].compute_node[0][2].in[0], compute_graph.flexml_layers[159].compute_node[0][2].in[0], compute_graph.flexml_layers[160].compute_node[0][2].in[0], compute_graph.flexml_layers[160].compute_node[0][2].in[2], compute_graph.flexml_layers[161].compute_node[0][2].in[0], compute_graph.flexml_layers[161].compute_node[0][2].in[2], compute_graph.flexml_layers[162].compute_node[0][2].in[0], compute_graph.flexml_layers[163].compute_node[0][2].in[0], compute_graph.flexml_layers[164].compute_node[0][2].in[0], compute_graph.flexml_layers[165].compute_node[0][2].in[0], compute_graph.flexml_layers[166].compute_node[0][2].in[0], compute_graph.flexml_layers[166].compute_node[0][2].in[2], compute_graph.flexml_layers[167].compute_node[0][2].in[0], compute_graph.flexml_layers[168].compute_node[0][2].in[0], compute_graph.flexml_layers[169].compute_node[0][2].in[0], compute_graph.flexml_layers[170].compute_node[0][2].in[0], compute_graph.flexml_layers[171].compute_node[0][2].in[0], compute_graph.flexml_layers[171].compute_node[0][2].in[2], compute_graph.flexml_layers[172].compute_node[0][2].in[0], compute_graph.flexml_layers[173].compute_node[0][2].in[0], compute_graph.flexml_layers[174].compute_node[0][2].in[0], compute_graph.flexml_layers[175].compute_node[0][2].in[0], compute_graph.flexml_layers[176].compute_node[0][2].in[0], compute_graph.flexml_layers[177].compute_node[0][2].in[0], compute_graph.flexml_layers[178].compute_node[0][2].in[0], compute_graph.flexml_layers[178].compute_node[0][2].in[2], compute_graph.flexml_layers[179].compute_node[0][2].in[0], compute_graph.flexml_layers[179].compute_node[0][2].in[2], compute_graph.flexml_layers[180].compute_node[0][2].in[0], compute_graph.flexml_layers[181].compute_node[0][2].in[0], compute_graph.flexml_layers[182].compute_node[0][2].in[0], compute_graph.flexml_layers[183].compute_node[0][2].in[0], compute_graph.flexml_layers[184].compute_node[0][2].in[0], compute_graph.flexml_layers[184].compute_node[0][2].in[2], compute_graph.flexml_layers[185].compute_node[0][2].in[0], compute_graph.flexml_layers[186].compute_node[0][2].in[0], compute_graph.flexml_layers[187].compute_node[0][2].in[0], compute_graph.flexml_layers[188].compute_node[0][2].in[0], compute_graph.flexml_layers[188].compute_node[0][2].in[2], compute_graph.flexml_layers[189].compute_node[0][2].in[0], compute_graph.flexml_layers[190].compute_node[0][2].in[0], compute_graph.flexml_layers[191].compute_node[0][2].in[0], compute_graph.flexml_layers[191].compute_node[0][2].in[2], compute_graph.flexml_layers[192].compute_node[0][2].in[0], compute_graph.flexml_layers[192].compute_node[0][2].in[2], compute_graph.flexml_layers[193].compute_node[0][2].in[0], compute_graph.flexml_layers[193].compute_node[0][2].in[2], compute_graph.flexml_layers[194].compute_node[0][2].in[0], compute_graph.flexml_layers[195].compute_node[0][2].in[0], compute_graph.flexml_layers[196].compute_node[0][2].in[0], compute_graph.flexml_layers[196].compute_node[0][2].in[2], compute_graph.flexml_layers[197].compute_node[0][2].in[0], compute_graph.flexml_layers[197].compute_node[0][2].in[2], compute_graph.flexml_layers[198].compute_node[0][2].in[0], compute_graph.flexml_layers[199].compute_node[0][2].in[0], compute_graph.flexml_layers[200].compute_node[0][2].in[0], compute_graph.flexml_layers[203].compute_node[0][2].in[0], compute_graph.flexml_layers[204].compute_node[0][2].in[0], compute_graph.flexml_layers[204].compute_node[0][2].in[2], compute_graph.flexml_layers[205].compute_node[0][2].in[0], compute_graph.flexml_layers[205].compute_node[0][2].in[2], compute_graph.flexml_layers[206].compute_node[0][2].in[0], compute_graph.flexml_layers[206].compute_node[0][2].in[2], compute_graph.flexml_layers[207].compute_node[0][2].in[0], compute_graph.flexml_layers[208].compute_node[0][2].in[0], compute_graph.flexml_layers[209].compute_node[0][2].in[0], compute_graph.flexml_layers[209].compute_node[0][2].in[2], compute_graph.flexml_layers[210].compute_node[0][2].in[0], compute_graph.flexml_layers[210].compute_node[0][2].in[2], compute_graph.flexml_layers[211].compute_node[0][2].in[0], compute_graph.flexml_layers[212].compute_node[0][2].in[0], compute_graph.flexml_layers[213].compute_node[0][2].in[0], compute_graph.flexml_layers[214].compute_node[0][2].in[0], compute_graph.flexml_layers[214].compute_node[0][2].in[2], compute_graph.flexml_layers[215].compute_node[0][2].in[0], compute_graph.flexml_layers[215].compute_node[0][2].in[2], compute_graph.flexml_layers[216].compute_node[0][2].in[0], compute_graph.flexml_layers[216].compute_node[0][2].in[2], compute_graph.flexml_layers[217].compute_node[0][2].in[0], compute_graph.flexml_layers[218].compute_node[0][2].in[0], compute_graph.flexml_layers[219].compute_node[0][2].in[0], compute_graph.flexml_layers[219].compute_node[0][2].in[2], compute_graph.flexml_layers[220].compute_node[0][2].in[0], compute_graph.flexml_layers[220].compute_node[0][2].in[2], compute_graph.flexml_layers[221].compute_node[0][2].in[0], compute_graph.flexml_layers[222].compute_node[0][2].in[0], compute_graph.flexml_layers[223].compute_node[0][2].in[0], compute_graph.flexml_layers[224].compute_node[0][2].in[0], compute_graph.flexml_layers[224].compute_node[0][2].in[2], compute_graph.flexml_layers[225].compute_node[0][2].in[0], compute_graph.flexml_layers[225].compute_node[0][2].in[2], compute_graph.flexml_layers[226].compute_node[0][2].in[0], compute_graph.flexml_layers[226].compute_node[0][2].in[2], compute_graph.flexml_layers[227].compute_node[0][2].in[0], compute_graph.flexml_layers[228].compute_node[0][2].in[0], compute_graph.flexml_layers[229].compute_node[0][2].in[0], compute_graph.flexml_layers[229].compute_node[0][2].in[2], compute_graph.flexml_layers[230].compute_node[0][2].in[0], compute_graph.flexml_layers[230].compute_node[0][2].in[2], compute_graph.flexml_layers[231].compute_node[0][2].in[0], compute_graph.flexml_layers[232].compute_node[0][2].in[0], compute_graph.flexml_layers[233].compute_node[0][2].in[0], compute_graph.flexml_layers[234].compute_node[0][2].in[0], compute_graph.flexml_layers[235].compute_node[0][2].in[0], compute_graph.flexml_layers[235].compute_node[0][2].in[2], compute_graph.flexml_layers[236].compute_node[0][2].in[0]) + +Column 0Row 2Channel 0 + +i6_po0(compute_graph.templated_graph_4.trans_comp_nd[0][2].out[0], compute_graph.flexml_layers[36].compute_node[0][2].out[0], compute_graph.templated_graph_231.mk[0][2].out[0], compute_graph.templated_graph_254.mk[0][2].out[0], compute_graph.templated_graph_258.compute_node[0][2].out[0], compute_graph.templated_graph_259.compute_node[0][2].out[0], compute_graph.templated_graph_260.mk[0][2].out[0], compute_graph.templated_graph_263.compute_node[0][2].out[0], compute_graph.templated_graph_266.compute_node[0][2].out[0], compute_graph.templated_graph_268.mk[0][2].out[0], compute_graph.templated_graph_273.mk[0][2].out[0], compute_graph.templated_graph_274.mk[0][2].out[0], compute_graph.flexml_layers[63].compute_node[0][2].out[0], compute_graph.flexml_layers[64].compute_node[0][2].out[0], compute_graph.templated_graph_293.mk[0][2].out[0], compute_graph.templated_graph_298.compute_node[0][2].out[0], compute_graph.templated_graph_300.compute_node[0][2].out[0], compute_graph.flexml_layers[0].compute_node[0][2].out[0], compute_graph.flexml_layers[1].compute_node[0][2].out[0], compute_graph.flexml_layers[2].compute_node[0][2].out[0], compute_graph.flexml_layers[3].compute_node[0][2].out[0], compute_graph.flexml_layers[4].compute_node[0][2].out[0], compute_graph.flexml_layers[5].compute_node[0][2].out[0], compute_graph.flexml_layers[6].compute_node[0][2].out[0], compute_graph.flexml_layers[7].compute_node[0][2].out[0], compute_graph.flexml_layers[8].compute_node[0][2].out[0], compute_graph.flexml_layers[9].compute_node[0][2].out[0], compute_graph.flexml_layers[10].compute_node[0][2].out[0], compute_graph.flexml_layers[11].compute_node[0][2].out[0], compute_graph.flexml_layers[12].compute_node[0][2].out[0], compute_graph.flexml_layers[13].compute_node[0][2].out[0], compute_graph.flexml_layers[14].compute_node[0][2].out[0], compute_graph.flexml_layers[15].compute_node[0][2].out[0], compute_graph.flexml_layers[16].compute_node[0][2].out[0], compute_graph.flexml_layers[17].compute_node[0][2].out[0], compute_graph.flexml_layers[18].compute_node[0][2].out[0], compute_graph.flexml_layers[19].compute_node[0][2].out[0], compute_graph.flexml_layers[20].compute_node[0][2].out[0], compute_graph.flexml_layers[21].compute_node[0][2].out[0], compute_graph.flexml_layers[22].compute_node[0][2].out[0], compute_graph.flexml_layers[23].compute_node[0][2].out[0], compute_graph.flexml_layers[24].compute_node[0][2].out[0], compute_graph.flexml_layers[25].compute_node[0][2].out[0], compute_graph.flexml_layers[26].compute_node[0][2].out[0], compute_graph.flexml_layers[27].compute_node[0][2].out[0], compute_graph.flexml_layers[28].compute_node[0][2].out[0], compute_graph.flexml_layers[29].compute_node[0][2].out[0], compute_graph.flexml_layers[30].compute_node[0][2].out[0], compute_graph.flexml_layers[31].compute_node[0][2].out[0], compute_graph.flexml_layers[32].compute_node[0][2].out[0], compute_graph.flexml_layers[33].compute_node[0][2].out[0], compute_graph.flexml_layers[34].compute_node[0][2].out[0], compute_graph.flexml_layers[35].compute_node[0][2].out[0], compute_graph.flexml_layers[144].compute_node[0][2].out[0], compute_graph.flexml_layers[145].compute_node[0][2].out[0], compute_graph.flexml_layers[37].compute_node[0][2].out[0], compute_graph.flexml_layers[38].compute_node[0][2].out[0], compute_graph.flexml_layers[39].compute_node[0][2].out[0], compute_graph.flexml_layers[40].compute_node[0][2].out[0], compute_graph.flexml_layers[41].compute_node[0][2].out[0], compute_graph.flexml_layers[42].compute_node[0][2].out[0], compute_graph.flexml_layers[43].compute_node[0][2].out[0], compute_graph.flexml_layers[44].compute_node[0][2].out[0], compute_graph.flexml_layers[45].compute_node[0][2].out[0], compute_graph.flexml_layers[46].compute_node[0][2].out[0], compute_graph.flexml_layers[47].compute_node[0][2].out[0], compute_graph.flexml_layers[48].compute_node[0][2].out[0], compute_graph.flexml_layers[49].compute_node[0][2].out[0], compute_graph.flexml_layers[50].compute_node[0][2].out[0], compute_graph.flexml_layers[51].compute_node[0][2].out[0], compute_graph.flexml_layers[52].compute_node[0][2].out[0], compute_graph.flexml_layers[53].compute_node[0][2].out[0], compute_graph.flexml_layers[54].compute_node[0][2].out[0], compute_graph.flexml_layers[55].compute_node[0][2].out[0], compute_graph.flexml_layers[56].compute_node[0][2].out[0], compute_graph.flexml_layers[57].compute_node[0][2].out[0], compute_graph.flexml_layers[58].compute_node[0][2].out[0], compute_graph.flexml_layers[59].compute_node[0][2].out[0], compute_graph.flexml_layers[60].compute_node[0][2].out[0], compute_graph.flexml_layers[61].compute_node[0][2].out[0], compute_graph.flexml_layers[62].compute_node[0][2].out[0], compute_graph.flexml_layers[201].compute_node[0][2].out[0], compute_graph.flexml_layers[202].compute_node[0][2].out[0], compute_graph.flexml_layers[65].compute_node[0][2].out[0], compute_graph.flexml_layers[66].compute_node[0][2].out[0], compute_graph.flexml_layers[67].compute_node[0][2].out[0], compute_graph.flexml_layers[68].compute_node[0][2].out[0], compute_graph.flexml_layers[69].compute_node[0][2].out[0], compute_graph.flexml_layers[70].compute_node[0][2].out[0], compute_graph.flexml_layers[71].compute_node[0][2].out[0], compute_graph.flexml_layers[72].compute_node[0][2].out[0], compute_graph.flexml_layers[73].compute_node[0][2].out[0], compute_graph.flexml_layers[74].compute_node[0][2].out[0], compute_graph.flexml_layers[75].compute_node[0][2].out[0], compute_graph.flexml_layers[76].compute_node[0][2].out[0], compute_graph.flexml_layers[77].compute_node[0][2].out[0], compute_graph.flexml_layers[78].compute_node[0][2].out[0], compute_graph.flexml_layers[79].compute_node[0][2].out[0], compute_graph.flexml_layers[80].compute_node[0][2].out[0], compute_graph.flexml_layers[81].compute_node[0][2].out[0], compute_graph.flexml_layers[82].compute_node[0][2].out[0], compute_graph.flexml_layers[83].compute_node[0][2].out[0], compute_graph.flexml_layers[84].compute_node[0][2].out[0], compute_graph.flexml_layers[85].compute_node[0][2].out[0], compute_graph.flexml_layers[86].compute_node[0][2].out[0], compute_graph.flexml_layers[87].compute_node[0][2].out[0], compute_graph.flexml_layers[88].compute_node[0][2].out[0], compute_graph.flexml_layers[89].compute_node[0][2].out[0], compute_graph.flexml_layers[90].compute_node[0][2].out[0], compute_graph.flexml_layers[91].compute_node[0][2].out[0], compute_graph.flexml_layers[92].compute_node[0][2].out[0], compute_graph.flexml_layers[93].compute_node[0][2].out[0], compute_graph.flexml_layers[94].compute_node[0][2].out[0], compute_graph.flexml_layers[95].compute_node[0][2].out[0], compute_graph.flexml_layers[96].compute_node[0][2].out[0], compute_graph.flexml_layers[97].compute_node[0][2].out[0], compute_graph.flexml_layers[98].compute_node[0][2].out[0], compute_graph.flexml_layers[99].compute_node[0][2].out[0], compute_graph.flexml_layers[100].compute_node[0][2].out[0], compute_graph.flexml_layers[101].compute_node[0][2].out[0], compute_graph.flexml_layers[102].compute_node[0][2].out[0], compute_graph.flexml_layers[103].compute_node[0][2].out[0], compute_graph.flexml_layers[104].compute_node[0][2].out[0], compute_graph.flexml_layers[105].compute_node[0][2].out[0], compute_graph.flexml_layers[106].compute_node[0][2].out[0], compute_graph.flexml_layers[107].compute_node[0][2].out[0], compute_graph.flexml_layers[108].compute_node[0][2].out[0], compute_graph.flexml_layers[109].compute_node[0][2].out[0], compute_graph.flexml_layers[110].compute_node[0][2].out[0], compute_graph.flexml_layers[111].compute_node[0][2].out[0], compute_graph.flexml_layers[112].compute_node[0][2].out[0], compute_graph.flexml_layers[113].compute_node[0][2].out[0], compute_graph.flexml_layers[114].compute_node[0][2].out[0], compute_graph.flexml_layers[115].compute_node[0][2].out[0], compute_graph.flexml_layers[116].compute_node[0][2].out[0], compute_graph.flexml_layers[117].compute_node[0][2].out[0], compute_graph.flexml_layers[118].compute_node[0][2].out[0], compute_graph.flexml_layers[119].compute_node[0][2].out[0], compute_graph.flexml_layers[120].compute_node[0][2].out[0], compute_graph.flexml_layers[121].compute_node[0][2].out[0], compute_graph.flexml_layers[122].compute_node[0][2].out[0], compute_graph.flexml_layers[123].compute_node[0][2].out[0], compute_graph.flexml_layers[124].compute_node[0][2].out[0], compute_graph.flexml_layers[125].compute_node[0][2].out[0], compute_graph.flexml_layers[126].compute_node[0][2].out[0], compute_graph.flexml_layers[127].compute_node[0][2].out[0], compute_graph.flexml_layers[128].compute_node[0][2].out[0], compute_graph.flexml_layers[129].compute_node[0][2].out[0], compute_graph.flexml_layers[130].compute_node[0][2].out[0], compute_graph.flexml_layers[131].compute_node[0][2].out[0], compute_graph.flexml_layers[132].compute_node[0][2].out[0], compute_graph.flexml_layers[133].compute_node[0][2].out[0], compute_graph.flexml_layers[134].compute_node[0][2].out[0], compute_graph.flexml_layers[135].compute_node[0][2].out[0], compute_graph.flexml_layers[136].compute_node[0][2].out[0], compute_graph.flexml_layers[137].compute_node[0][2].out[0], compute_graph.flexml_layers[138].compute_node[0][2].out[0], compute_graph.flexml_layers[139].compute_node[0][2].out[0], compute_graph.flexml_layers[140].compute_node[0][2].out[0], compute_graph.flexml_layers[141].compute_node[0][2].out[0], compute_graph.flexml_layers[142].compute_node[0][2].out[0], compute_graph.flexml_layers[143].compute_node[0][2].out[0], compute_graph.flexml_layers[146].compute_node[0][2].out[0], compute_graph.flexml_layers[147].compute_node[0][2].out[0], compute_graph.flexml_layers[148].compute_node[0][2].out[0], compute_graph.flexml_layers[149].compute_node[0][2].out[0], compute_graph.flexml_layers[150].compute_node[0][2].out[0], compute_graph.flexml_layers[151].compute_node[0][2].out[0], compute_graph.flexml_layers[152].compute_node[0][2].out[0], compute_graph.flexml_layers[153].compute_node[0][2].out[0], compute_graph.flexml_layers[154].compute_node[0][2].out[0], compute_graph.flexml_layers[155].compute_node[0][2].out[0], compute_graph.flexml_layers[156].compute_node[0][2].out[0], compute_graph.flexml_layers[157].compute_node[0][2].out[0], compute_graph.flexml_layers[158].compute_node[0][2].out[0], compute_graph.flexml_layers[159].compute_node[0][2].out[0], compute_graph.flexml_layers[160].compute_node[0][2].out[0], compute_graph.flexml_layers[161].compute_node[0][2].out[0], compute_graph.flexml_layers[162].compute_node[0][2].out[0], compute_graph.flexml_layers[163].compute_node[0][2].out[0], compute_graph.flexml_layers[164].compute_node[0][2].out[0], compute_graph.flexml_layers[165].compute_node[0][2].out[0], compute_graph.flexml_layers[166].compute_node[0][2].out[0], compute_graph.flexml_layers[167].compute_node[0][2].out[0], compute_graph.flexml_layers[168].compute_node[0][2].out[0], compute_graph.flexml_layers[169].compute_node[0][2].out[0], compute_graph.flexml_layers[170].compute_node[0][2].out[0], compute_graph.flexml_layers[171].compute_node[0][2].out[0], compute_graph.flexml_layers[172].compute_node[0][2].out[0], compute_graph.flexml_layers[173].compute_node[0][2].out[0], compute_graph.flexml_layers[174].compute_node[0][2].out[0], compute_graph.flexml_layers[175].compute_node[0][2].out[0], compute_graph.flexml_layers[176].compute_node[0][2].out[0], compute_graph.flexml_layers[177].compute_node[0][2].out[0], compute_graph.flexml_layers[178].compute_node[0][2].out[0], compute_graph.flexml_layers[179].compute_node[0][2].out[0], compute_graph.flexml_layers[180].compute_node[0][2].out[0], compute_graph.flexml_layers[181].compute_node[0][2].out[0], compute_graph.flexml_layers[182].compute_node[0][2].out[0], compute_graph.flexml_layers[183].compute_node[0][2].out[0], compute_graph.flexml_layers[184].compute_node[0][2].out[0], compute_graph.flexml_layers[185].compute_node[0][2].out[0], compute_graph.flexml_layers[186].compute_node[0][2].out[0], compute_graph.flexml_layers[187].compute_node[0][2].out[0], compute_graph.flexml_layers[188].compute_node[0][2].out[0], compute_graph.flexml_layers[189].compute_node[0][2].out[0], compute_graph.flexml_layers[190].compute_node[0][2].out[0], compute_graph.flexml_layers[191].compute_node[0][2].out[0], compute_graph.flexml_layers[192].compute_node[0][2].out[0], compute_graph.flexml_layers[193].compute_node[0][2].out[0], compute_graph.flexml_layers[194].compute_node[0][2].out[0], compute_graph.flexml_layers[195].compute_node[0][2].out[0], compute_graph.flexml_layers[196].compute_node[0][2].out[0], compute_graph.flexml_layers[197].compute_node[0][2].out[0], compute_graph.flexml_layers[198].compute_node[0][2].out[0], compute_graph.flexml_layers[199].compute_node[0][2].out[0], compute_graph.flexml_layers[200].compute_node[0][2].out[0], compute_graph.flexml_layers[203].compute_node[0][2].out[0], compute_graph.flexml_layers[204].compute_node[0][2].out[0], compute_graph.flexml_layers[205].compute_node[0][2].out[0], compute_graph.flexml_layers[206].compute_node[0][2].out[0], compute_graph.flexml_layers[207].compute_node[0][2].out[0], compute_graph.flexml_layers[208].compute_node[0][2].out[0], compute_graph.flexml_layers[209].compute_node[0][2].out[0], compute_graph.flexml_layers[210].compute_node[0][2].out[0], compute_graph.flexml_layers[211].compute_node[0][2].out[0], compute_graph.flexml_layers[212].compute_node[0][2].out[0], compute_graph.flexml_layers[213].compute_node[0][2].out[0], compute_graph.flexml_layers[214].compute_node[0][2].out[0], compute_graph.flexml_layers[215].compute_node[0][2].out[0], compute_graph.flexml_layers[216].compute_node[0][2].out[0], compute_graph.flexml_layers[217].compute_node[0][2].out[0], compute_graph.flexml_layers[218].compute_node[0][2].out[0], compute_graph.flexml_layers[219].compute_node[0][2].out[0], compute_graph.flexml_layers[220].compute_node[0][2].out[0], compute_graph.flexml_layers[221].compute_node[0][2].out[0], compute_graph.flexml_layers[222].compute_node[0][2].out[0], compute_graph.flexml_layers[223].compute_node[0][2].out[0], compute_graph.flexml_layers[224].compute_node[0][2].out[0], compute_graph.flexml_layers[225].compute_node[0][2].out[0], compute_graph.flexml_layers[226].compute_node[0][2].out[0], compute_graph.flexml_layers[227].compute_node[0][2].out[0], compute_graph.flexml_layers[228].compute_node[0][2].out[0], compute_graph.flexml_layers[229].compute_node[0][2].out[0], compute_graph.flexml_layers[230].compute_node[0][2].out[0], compute_graph.flexml_layers[231].compute_node[0][2].out[0], compute_graph.flexml_layers[232].compute_node[0][2].out[0], compute_graph.flexml_layers[233].compute_node[0][2].out[0], compute_graph.flexml_layers[234].compute_node[0][2].out[0], compute_graph.flexml_layers[235].compute_node[0][2].out[0], compute_graph.flexml_layers[236].compute_node[0][2].out[0]) + +Column 0Row 2Channel 1 + +i6_pi1(compute_graph.flexml_layers[36].compute_node[0][2].in[1], compute_graph.templated_graph_260.mk[0][2].in[1], compute_graph.templated_graph_268.mk[0][2].in[1], compute_graph.templated_graph_273.mk[0][2].in[1], compute_graph.flexml_layers[3].compute_node[0][2].in[1], compute_graph.flexml_layers[8].compute_node[0][2].in[1], compute_graph.flexml_layers[9].compute_node[0][2].in[1], compute_graph.flexml_layers[10].compute_node[0][2].in[1], compute_graph.flexml_layers[11].compute_node[0][2].in[1], compute_graph.flexml_layers[12].compute_node[0][2].in[1], compute_graph.flexml_layers[13].compute_node[0][2].in[1], compute_graph.flexml_layers[14].compute_node[0][2].in[1], compute_graph.flexml_layers[15].compute_node[0][2].in[1], compute_graph.flexml_layers[16].compute_node[0][2].in[1], compute_graph.flexml_layers[17].compute_node[0][2].in[1], compute_graph.flexml_layers[19].compute_node[0][2].in[1], compute_graph.flexml_layers[20].compute_node[0][2].in[1], compute_graph.flexml_layers[25].compute_node[0][2].in[1], compute_graph.flexml_layers[26].compute_node[0][2].in[1], compute_graph.flexml_layers[27].compute_node[0][2].in[1], compute_graph.flexml_layers[29].compute_node[0][2].in[1], compute_graph.flexml_layers[30].compute_node[0][2].in[1], compute_graph.flexml_layers[35].compute_node[0][2].in[1], compute_graph.flexml_layers[144].compute_node[0][2].in[1], compute_graph.flexml_layers[37].compute_node[0][2].in[1], compute_graph.flexml_layers[39].compute_node[0][2].in[1], compute_graph.flexml_layers[40].compute_node[0][2].in[1], compute_graph.flexml_layers[45].compute_node[0][2].in[1], compute_graph.flexml_layers[46].compute_node[0][2].in[1], compute_graph.flexml_layers[51].compute_node[0][2].in[1], compute_graph.flexml_layers[56].compute_node[0][2].in[1], compute_graph.flexml_layers[57].compute_node[0][2].in[1], compute_graph.flexml_layers[62].compute_node[0][2].in[1], compute_graph.flexml_layers[201].compute_node[0][2].in[1], compute_graph.flexml_layers[202].compute_node[0][2].in[1], compute_graph.flexml_layers[67].compute_node[0][2].in[1], compute_graph.flexml_layers[68].compute_node[0][2].in[1], compute_graph.flexml_layers[73].compute_node[0][2].in[1], compute_graph.flexml_layers[78].compute_node[0][2].in[1], compute_graph.flexml_layers[79].compute_node[0][2].in[1], compute_graph.flexml_layers[84].compute_node[0][2].in[1], compute_graph.flexml_layers[89].compute_node[0][2].in[1], compute_graph.flexml_layers[90].compute_node[0][2].in[1], compute_graph.flexml_layers[95].compute_node[0][2].in[1], compute_graph.flexml_layers[101].compute_node[0][2].in[1], compute_graph.flexml_layers[102].compute_node[0][2].in[1], compute_graph.flexml_layers[107].compute_node[0][2].in[1], compute_graph.flexml_layers[108].compute_node[0][2].in[1], compute_graph.flexml_layers[113].compute_node[0][2].in[1], compute_graph.flexml_layers[119].compute_node[0][2].in[1], compute_graph.flexml_layers[120].compute_node[0][2].in[1], compute_graph.flexml_layers[125].compute_node[0][2].in[1], compute_graph.flexml_layers[126].compute_node[0][2].in[1], compute_graph.flexml_layers[131].compute_node[0][2].in[1], compute_graph.flexml_layers[137].compute_node[0][2].in[1], compute_graph.flexml_layers[138].compute_node[0][2].in[1], compute_graph.flexml_layers[143].compute_node[0][2].in[1], compute_graph.flexml_layers[149].compute_node[0][2].in[1], compute_graph.flexml_layers[155].compute_node[0][2].in[1], compute_graph.flexml_layers[156].compute_node[0][2].in[1], compute_graph.flexml_layers[161].compute_node[0][2].in[1], compute_graph.flexml_layers[162].compute_node[0][2].in[1], compute_graph.flexml_layers[167].compute_node[0][2].in[1], compute_graph.flexml_layers[173].compute_node[0][2].in[1], compute_graph.flexml_layers[174].compute_node[0][2].in[1], compute_graph.flexml_layers[179].compute_node[0][2].in[1], compute_graph.flexml_layers[180].compute_node[0][2].in[1], compute_graph.flexml_layers[186].compute_node[0][2].in[1], compute_graph.flexml_layers[188].compute_node[0][2].in[1], compute_graph.flexml_layers[189].compute_node[0][2].in[1], compute_graph.flexml_layers[194].compute_node[0][2].in[1], compute_graph.flexml_layers[207].compute_node[0][2].in[1], compute_graph.flexml_layers[211].compute_node[0][2].in[1], compute_graph.flexml_layers[212].compute_node[0][2].in[1], compute_graph.flexml_layers[217].compute_node[0][2].in[1], compute_graph.flexml_layers[221].compute_node[0][2].in[1], compute_graph.flexml_layers[222].compute_node[0][2].in[1], compute_graph.flexml_layers[227].compute_node[0][2].in[1], compute_graph.flexml_layers[231].compute_node[0][2].in[1], compute_graph.flexml_layers[232].compute_node[0][2].in[1], compute_graph.flexml_layers[233].compute_node[0][2].in[1]) + +Column 0Row 3Channel 0 + +i7_pi0(compute_graph.templated_graph_4.trans_comp_nd[0][3].in[0], compute_graph.flexml_layers[36].compute_node[0][3].in[0], compute_graph.templated_graph_231.mk[0][3].in[0], compute_graph.templated_graph_254.mk[0][3].in[0], compute_graph.templated_graph_258.compute_node[0][3].in[0], compute_graph.templated_graph_259.compute_node[0][3].in[0], compute_graph.templated_graph_260.mk[0][3].in[0], compute_graph.templated_graph_263.compute_node[0][3].in[0], compute_graph.templated_graph_266.compute_node[0][3].in[0], compute_graph.templated_graph_268.mk[0][3].in[0], compute_graph.templated_graph_273.mk[0][3].in[0], compute_graph.templated_graph_274.mk[0][3].in[0], compute_graph.flexml_layers[63].compute_node[0][3].in[0], compute_graph.flexml_layers[64].compute_node[0][3].in[0], compute_graph.templated_graph_293.mk[0][3].in[0], compute_graph.templated_graph_298.compute_node[0][3].in[0], compute_graph.templated_graph_300.compute_node[0][3].in[0], compute_graph.flexml_layers[0].compute_node[0][3].in[0], compute_graph.flexml_layers[1].compute_node[0][3].in[0], compute_graph.flexml_layers[1].compute_node[0][3].in[2], compute_graph.flexml_layers[2].compute_node[0][3].in[0], compute_graph.flexml_layers[2].compute_node[0][3].in[2], compute_graph.flexml_layers[3].compute_node[0][3].in[0], compute_graph.flexml_layers[4].compute_node[0][3].in[0], compute_graph.flexml_layers[5].compute_node[0][3].in[0], compute_graph.flexml_layers[6].compute_node[0][3].in[0], compute_graph.flexml_layers[7].compute_node[0][3].in[0], compute_graph.flexml_layers[7].compute_node[0][3].in[2], compute_graph.flexml_layers[8].compute_node[0][3].in[0], compute_graph.flexml_layers[9].compute_node[0][3].in[0], compute_graph.flexml_layers[9].compute_node[0][3].in[2], compute_graph.flexml_layers[10].compute_node[0][3].in[0], compute_graph.flexml_layers[11].compute_node[0][3].in[0], compute_graph.flexml_layers[12].compute_node[0][3].in[0], compute_graph.flexml_layers[13].compute_node[0][3].in[0], compute_graph.flexml_layers[14].compute_node[0][3].in[0], compute_graph.flexml_layers[15].compute_node[0][3].in[0], compute_graph.flexml_layers[15].compute_node[0][3].in[2], compute_graph.flexml_layers[16].compute_node[0][3].in[0], compute_graph.flexml_layers[17].compute_node[0][3].in[0], compute_graph.flexml_layers[18].compute_node[0][3].in[0], compute_graph.flexml_layers[19].compute_node[0][3].in[0], compute_graph.flexml_layers[20].compute_node[0][3].in[0], compute_graph.flexml_layers[21].compute_node[0][3].in[0], compute_graph.flexml_layers[22].compute_node[0][3].in[0], compute_graph.flexml_layers[23].compute_node[0][3].in[0], compute_graph.flexml_layers[24].compute_node[0][3].in[0], compute_graph.flexml_layers[24].compute_node[0][3].in[2], compute_graph.flexml_layers[25].compute_node[0][3].in[0], compute_graph.flexml_layers[26].compute_node[0][3].in[0], compute_graph.flexml_layers[27].compute_node[0][3].in[0], compute_graph.flexml_layers[28].compute_node[0][3].in[0], compute_graph.flexml_layers[29].compute_node[0][3].in[0], compute_graph.flexml_layers[30].compute_node[0][3].in[0], compute_graph.flexml_layers[31].compute_node[0][3].in[0], compute_graph.flexml_layers[32].compute_node[0][3].in[0], compute_graph.flexml_layers[33].compute_node[0][3].in[0], compute_graph.flexml_layers[34].compute_node[0][3].in[0], compute_graph.flexml_layers[34].compute_node[0][3].in[2], compute_graph.flexml_layers[35].compute_node[0][3].in[0], compute_graph.flexml_layers[35].compute_node[0][3].in[2], compute_graph.flexml_layers[144].compute_node[0][3].in[0], compute_graph.flexml_layers[145].compute_node[0][3].in[0], compute_graph.flexml_layers[37].compute_node[0][3].in[0], compute_graph.flexml_layers[38].compute_node[0][3].in[0], compute_graph.flexml_layers[39].compute_node[0][3].in[0], compute_graph.flexml_layers[40].compute_node[0][3].in[0], compute_graph.flexml_layers[41].compute_node[0][3].in[0], compute_graph.flexml_layers[42].compute_node[0][3].in[0], compute_graph.flexml_layers[43].compute_node[0][3].in[0], compute_graph.flexml_layers[44].compute_node[0][3].in[0], compute_graph.flexml_layers[44].compute_node[0][3].in[2], compute_graph.flexml_layers[45].compute_node[0][3].in[0], compute_graph.flexml_layers[45].compute_node[0][3].in[2], compute_graph.flexml_layers[46].compute_node[0][3].in[0], compute_graph.flexml_layers[47].compute_node[0][3].in[0], compute_graph.flexml_layers[48].compute_node[0][3].in[0], compute_graph.flexml_layers[49].compute_node[0][3].in[0], compute_graph.flexml_layers[50].compute_node[0][3].in[0], compute_graph.flexml_layers[50].compute_node[0][3].in[2], compute_graph.flexml_layers[51].compute_node[0][3].in[0], compute_graph.flexml_layers[52].compute_node[0][3].in[0], compute_graph.flexml_layers[53].compute_node[0][3].in[0], compute_graph.flexml_layers[54].compute_node[0][3].in[0], compute_graph.flexml_layers[55].compute_node[0][3].in[0], compute_graph.flexml_layers[55].compute_node[0][3].in[2], compute_graph.flexml_layers[56].compute_node[0][3].in[0], compute_graph.flexml_layers[57].compute_node[0][3].in[0], compute_graph.flexml_layers[58].compute_node[0][3].in[0], compute_graph.flexml_layers[59].compute_node[0][3].in[0], compute_graph.flexml_layers[60].compute_node[0][3].in[0], compute_graph.flexml_layers[61].compute_node[0][3].in[0], compute_graph.flexml_layers[61].compute_node[0][3].in[2], compute_graph.flexml_layers[62].compute_node[0][3].in[0], compute_graph.flexml_layers[201].compute_node[0][3].in[0], compute_graph.flexml_layers[202].compute_node[0][3].in[0], compute_graph.flexml_layers[65].compute_node[0][3].in[0], compute_graph.flexml_layers[66].compute_node[0][3].in[0], compute_graph.flexml_layers[66].compute_node[0][3].in[2], compute_graph.flexml_layers[67].compute_node[0][3].in[0], compute_graph.flexml_layers[67].compute_node[0][3].in[2], compute_graph.flexml_layers[68].compute_node[0][3].in[0], compute_graph.flexml_layers[69].compute_node[0][3].in[0], compute_graph.flexml_layers[70].compute_node[0][3].in[0], compute_graph.flexml_layers[71].compute_node[0][3].in[0], compute_graph.flexml_layers[72].compute_node[0][3].in[0], compute_graph.flexml_layers[72].compute_node[0][3].in[2], compute_graph.flexml_layers[73].compute_node[0][3].in[0], compute_graph.flexml_layers[74].compute_node[0][3].in[0], compute_graph.flexml_layers[75].compute_node[0][3].in[0], compute_graph.flexml_layers[76].compute_node[0][3].in[0], compute_graph.flexml_layers[77].compute_node[0][3].in[0], compute_graph.flexml_layers[77].compute_node[0][3].in[2], compute_graph.flexml_layers[78].compute_node[0][3].in[0], compute_graph.flexml_layers[78].compute_node[0][3].in[2], compute_graph.flexml_layers[79].compute_node[0][3].in[0], compute_graph.flexml_layers[80].compute_node[0][3].in[0], compute_graph.flexml_layers[81].compute_node[0][3].in[0], compute_graph.flexml_layers[82].compute_node[0][3].in[0], compute_graph.flexml_layers[83].compute_node[0][3].in[0], compute_graph.flexml_layers[83].compute_node[0][3].in[2], compute_graph.flexml_layers[84].compute_node[0][3].in[0], compute_graph.flexml_layers[85].compute_node[0][3].in[0], compute_graph.flexml_layers[86].compute_node[0][3].in[0], compute_graph.flexml_layers[87].compute_node[0][3].in[0], compute_graph.flexml_layers[88].compute_node[0][3].in[0], compute_graph.flexml_layers[88].compute_node[0][3].in[2], compute_graph.flexml_layers[89].compute_node[0][3].in[0], compute_graph.flexml_layers[89].compute_node[0][3].in[2], compute_graph.flexml_layers[90].compute_node[0][3].in[0], compute_graph.flexml_layers[91].compute_node[0][3].in[0], compute_graph.flexml_layers[92].compute_node[0][3].in[0], compute_graph.flexml_layers[93].compute_node[0][3].in[0], compute_graph.flexml_layers[94].compute_node[0][3].in[0], compute_graph.flexml_layers[94].compute_node[0][3].in[2], compute_graph.flexml_layers[95].compute_node[0][3].in[0], compute_graph.flexml_layers[96].compute_node[0][3].in[0], compute_graph.flexml_layers[97].compute_node[0][3].in[0], compute_graph.flexml_layers[98].compute_node[0][3].in[0], compute_graph.flexml_layers[99].compute_node[0][3].in[0], compute_graph.flexml_layers[99].compute_node[0][3].in[2], compute_graph.flexml_layers[100].compute_node[0][3].in[0], compute_graph.flexml_layers[101].compute_node[0][3].in[0], compute_graph.flexml_layers[102].compute_node[0][3].in[0], compute_graph.flexml_layers[103].compute_node[0][3].in[0], compute_graph.flexml_layers[104].compute_node[0][3].in[0], compute_graph.flexml_layers[105].compute_node[0][3].in[0], compute_graph.flexml_layers[106].compute_node[0][3].in[0], compute_graph.flexml_layers[106].compute_node[0][3].in[2], compute_graph.flexml_layers[107].compute_node[0][3].in[0], compute_graph.flexml_layers[108].compute_node[0][3].in[0], compute_graph.flexml_layers[109].compute_node[0][3].in[0], compute_graph.flexml_layers[110].compute_node[0][3].in[0], compute_graph.flexml_layers[111].compute_node[0][3].in[0], compute_graph.flexml_layers[112].compute_node[0][3].in[0], compute_graph.flexml_layers[112].compute_node[0][3].in[2], compute_graph.flexml_layers[113].compute_node[0][3].in[0], compute_graph.flexml_layers[114].compute_node[0][3].in[0], compute_graph.flexml_layers[115].compute_node[0][3].in[0], compute_graph.flexml_layers[116].compute_node[0][3].in[0], compute_graph.flexml_layers[117].compute_node[0][3].in[0], compute_graph.flexml_layers[117].compute_node[0][3].in[2], compute_graph.flexml_layers[118].compute_node[0][3].in[0], compute_graph.flexml_layers[119].compute_node[0][3].in[0], compute_graph.flexml_layers[120].compute_node[0][3].in[0], compute_graph.flexml_layers[121].compute_node[0][3].in[0], compute_graph.flexml_layers[122].compute_node[0][3].in[0], compute_graph.flexml_layers[123].compute_node[0][3].in[0], compute_graph.flexml_layers[124].compute_node[0][3].in[0], compute_graph.flexml_layers[124].compute_node[0][3].in[2], compute_graph.flexml_layers[125].compute_node[0][3].in[0], compute_graph.flexml_layers[125].compute_node[0][3].in[2], compute_graph.flexml_layers[126].compute_node[0][3].in[0], compute_graph.flexml_layers[127].compute_node[0][3].in[0], compute_graph.flexml_layers[128].compute_node[0][3].in[0], compute_graph.flexml_layers[129].compute_node[0][3].in[0], compute_graph.flexml_layers[130].compute_node[0][3].in[0], compute_graph.flexml_layers[130].compute_node[0][3].in[2], compute_graph.flexml_layers[131].compute_node[0][3].in[0], compute_graph.flexml_layers[132].compute_node[0][3].in[0], compute_graph.flexml_layers[133].compute_node[0][3].in[0], compute_graph.flexml_layers[134].compute_node[0][3].in[0], compute_graph.flexml_layers[135].compute_node[0][3].in[0], compute_graph.flexml_layers[135].compute_node[0][3].in[2], compute_graph.flexml_layers[136].compute_node[0][3].in[0], compute_graph.flexml_layers[137].compute_node[0][3].in[0], compute_graph.flexml_layers[138].compute_node[0][3].in[0], compute_graph.flexml_layers[139].compute_node[0][3].in[0], compute_graph.flexml_layers[140].compute_node[0][3].in[0], compute_graph.flexml_layers[141].compute_node[0][3].in[0], compute_graph.flexml_layers[142].compute_node[0][3].in[0], compute_graph.flexml_layers[142].compute_node[0][3].in[2], compute_graph.flexml_layers[143].compute_node[0][3].in[0], compute_graph.flexml_layers[146].compute_node[0][3].in[0], compute_graph.flexml_layers[147].compute_node[0][3].in[0], compute_graph.flexml_layers[148].compute_node[0][3].in[0], compute_graph.flexml_layers[148].compute_node[0][3].in[2], compute_graph.flexml_layers[149].compute_node[0][3].in[0], compute_graph.flexml_layers[150].compute_node[0][3].in[0], compute_graph.flexml_layers[151].compute_node[0][3].in[0], compute_graph.flexml_layers[152].compute_node[0][3].in[0], compute_graph.flexml_layers[153].compute_node[0][3].in[0], compute_graph.flexml_layers[153].compute_node[0][3].in[2], compute_graph.flexml_layers[154].compute_node[0][3].in[0], compute_graph.flexml_layers[155].compute_node[0][3].in[0], compute_graph.flexml_layers[156].compute_node[0][3].in[0], compute_graph.flexml_layers[157].compute_node[0][3].in[0], compute_graph.flexml_layers[158].compute_node[0][3].in[0], compute_graph.flexml_layers[159].compute_node[0][3].in[0], compute_graph.flexml_layers[160].compute_node[0][3].in[0], compute_graph.flexml_layers[160].compute_node[0][3].in[2], compute_graph.flexml_layers[161].compute_node[0][3].in[0], compute_graph.flexml_layers[161].compute_node[0][3].in[2], compute_graph.flexml_layers[162].compute_node[0][3].in[0], compute_graph.flexml_layers[163].compute_node[0][3].in[0], compute_graph.flexml_layers[164].compute_node[0][3].in[0], compute_graph.flexml_layers[165].compute_node[0][3].in[0], compute_graph.flexml_layers[166].compute_node[0][3].in[0], compute_graph.flexml_layers[166].compute_node[0][3].in[2], compute_graph.flexml_layers[167].compute_node[0][3].in[0], compute_graph.flexml_layers[168].compute_node[0][3].in[0], compute_graph.flexml_layers[169].compute_node[0][3].in[0], compute_graph.flexml_layers[170].compute_node[0][3].in[0], compute_graph.flexml_layers[171].compute_node[0][3].in[0], compute_graph.flexml_layers[171].compute_node[0][3].in[2], compute_graph.flexml_layers[172].compute_node[0][3].in[0], compute_graph.flexml_layers[173].compute_node[0][3].in[0], compute_graph.flexml_layers[174].compute_node[0][3].in[0], compute_graph.flexml_layers[175].compute_node[0][3].in[0], compute_graph.flexml_layers[176].compute_node[0][3].in[0], compute_graph.flexml_layers[177].compute_node[0][3].in[0], compute_graph.flexml_layers[178].compute_node[0][3].in[0], compute_graph.flexml_layers[178].compute_node[0][3].in[2], compute_graph.flexml_layers[179].compute_node[0][3].in[0], compute_graph.flexml_layers[179].compute_node[0][3].in[2], compute_graph.flexml_layers[180].compute_node[0][3].in[0], compute_graph.flexml_layers[181].compute_node[0][3].in[0], compute_graph.flexml_layers[182].compute_node[0][3].in[0], compute_graph.flexml_layers[183].compute_node[0][3].in[0], compute_graph.flexml_layers[184].compute_node[0][3].in[0], compute_graph.flexml_layers[184].compute_node[0][3].in[2], compute_graph.flexml_layers[185].compute_node[0][3].in[0], compute_graph.flexml_layers[186].compute_node[0][3].in[0], compute_graph.flexml_layers[187].compute_node[0][3].in[0], compute_graph.flexml_layers[188].compute_node[0][3].in[0], compute_graph.flexml_layers[188].compute_node[0][3].in[2], compute_graph.flexml_layers[189].compute_node[0][3].in[0], compute_graph.flexml_layers[190].compute_node[0][3].in[0], compute_graph.flexml_layers[191].compute_node[0][3].in[0], compute_graph.flexml_layers[191].compute_node[0][3].in[2], compute_graph.flexml_layers[192].compute_node[0][3].in[0], compute_graph.flexml_layers[192].compute_node[0][3].in[2], compute_graph.flexml_layers[193].compute_node[0][3].in[0], compute_graph.flexml_layers[193].compute_node[0][3].in[2], compute_graph.flexml_layers[194].compute_node[0][3].in[0], compute_graph.flexml_layers[195].compute_node[0][3].in[0], compute_graph.flexml_layers[196].compute_node[0][3].in[0], compute_graph.flexml_layers[196].compute_node[0][3].in[2], compute_graph.flexml_layers[197].compute_node[0][3].in[0], compute_graph.flexml_layers[197].compute_node[0][3].in[2], compute_graph.flexml_layers[198].compute_node[0][3].in[0], compute_graph.flexml_layers[199].compute_node[0][3].in[0], compute_graph.flexml_layers[200].compute_node[0][3].in[0], compute_graph.flexml_layers[203].compute_node[0][3].in[0], compute_graph.flexml_layers[204].compute_node[0][3].in[0], compute_graph.flexml_layers[204].compute_node[0][3].in[2], compute_graph.flexml_layers[205].compute_node[0][3].in[0], compute_graph.flexml_layers[205].compute_node[0][3].in[2], compute_graph.flexml_layers[206].compute_node[0][3].in[0], compute_graph.flexml_layers[206].compute_node[0][3].in[2], compute_graph.flexml_layers[207].compute_node[0][3].in[0], compute_graph.flexml_layers[208].compute_node[0][3].in[0], compute_graph.flexml_layers[209].compute_node[0][3].in[0], compute_graph.flexml_layers[209].compute_node[0][3].in[2], compute_graph.flexml_layers[210].compute_node[0][3].in[0], compute_graph.flexml_layers[210].compute_node[0][3].in[2], compute_graph.flexml_layers[211].compute_node[0][3].in[0], compute_graph.flexml_layers[212].compute_node[0][3].in[0], compute_graph.flexml_layers[213].compute_node[0][3].in[0], compute_graph.flexml_layers[214].compute_node[0][3].in[0], compute_graph.flexml_layers[214].compute_node[0][3].in[2], compute_graph.flexml_layers[215].compute_node[0][3].in[0], compute_graph.flexml_layers[215].compute_node[0][3].in[2], compute_graph.flexml_layers[216].compute_node[0][3].in[0], compute_graph.flexml_layers[216].compute_node[0][3].in[2], compute_graph.flexml_layers[217].compute_node[0][3].in[0], compute_graph.flexml_layers[218].compute_node[0][3].in[0], compute_graph.flexml_layers[219].compute_node[0][3].in[0], compute_graph.flexml_layers[219].compute_node[0][3].in[2], compute_graph.flexml_layers[220].compute_node[0][3].in[0], compute_graph.flexml_layers[220].compute_node[0][3].in[2], compute_graph.flexml_layers[221].compute_node[0][3].in[0], compute_graph.flexml_layers[222].compute_node[0][3].in[0], compute_graph.flexml_layers[223].compute_node[0][3].in[0], compute_graph.flexml_layers[224].compute_node[0][3].in[0], compute_graph.flexml_layers[224].compute_node[0][3].in[2], compute_graph.flexml_layers[225].compute_node[0][3].in[0], compute_graph.flexml_layers[225].compute_node[0][3].in[2], compute_graph.flexml_layers[226].compute_node[0][3].in[0], compute_graph.flexml_layers[226].compute_node[0][3].in[2], compute_graph.flexml_layers[227].compute_node[0][3].in[0], compute_graph.flexml_layers[228].compute_node[0][3].in[0], compute_graph.flexml_layers[229].compute_node[0][3].in[0], compute_graph.flexml_layers[229].compute_node[0][3].in[2], compute_graph.flexml_layers[230].compute_node[0][3].in[0], compute_graph.flexml_layers[230].compute_node[0][3].in[2], compute_graph.flexml_layers[231].compute_node[0][3].in[0], compute_graph.flexml_layers[232].compute_node[0][3].in[0], compute_graph.flexml_layers[233].compute_node[0][3].in[0], compute_graph.flexml_layers[234].compute_node[0][3].in[0], compute_graph.flexml_layers[235].compute_node[0][3].in[0], compute_graph.flexml_layers[235].compute_node[0][3].in[2], compute_graph.flexml_layers[236].compute_node[0][3].in[0]) + +Column 0Row 3Channel 0 + +i7_po0(compute_graph.templated_graph_4.trans_comp_nd[0][3].out[0], compute_graph.flexml_layers[36].compute_node[0][3].out[0], compute_graph.templated_graph_231.mk[0][3].out[0], compute_graph.templated_graph_254.mk[0][3].out[0], compute_graph.templated_graph_258.compute_node[0][3].out[0], compute_graph.templated_graph_259.compute_node[0][3].out[0], compute_graph.templated_graph_260.mk[0][3].out[0], compute_graph.templated_graph_263.compute_node[0][3].out[0], compute_graph.templated_graph_266.compute_node[0][3].out[0], compute_graph.templated_graph_268.mk[0][3].out[0], compute_graph.templated_graph_273.mk[0][3].out[0], compute_graph.templated_graph_274.mk[0][3].out[0], compute_graph.flexml_layers[63].compute_node[0][3].out[0], compute_graph.flexml_layers[64].compute_node[0][3].out[0], compute_graph.templated_graph_293.mk[0][3].out[0], compute_graph.templated_graph_298.compute_node[0][3].out[0], compute_graph.templated_graph_300.compute_node[0][3].out[0], compute_graph.flexml_layers[0].compute_node[0][3].out[0], compute_graph.flexml_layers[1].compute_node[0][3].out[0], compute_graph.flexml_layers[2].compute_node[0][3].out[0], compute_graph.flexml_layers[3].compute_node[0][3].out[0], compute_graph.flexml_layers[4].compute_node[0][3].out[0], compute_graph.flexml_layers[5].compute_node[0][3].out[0], compute_graph.flexml_layers[6].compute_node[0][3].out[0], compute_graph.flexml_layers[7].compute_node[0][3].out[0], compute_graph.flexml_layers[8].compute_node[0][3].out[0], compute_graph.flexml_layers[9].compute_node[0][3].out[0], compute_graph.flexml_layers[10].compute_node[0][3].out[0], compute_graph.flexml_layers[11].compute_node[0][3].out[0], compute_graph.flexml_layers[12].compute_node[0][3].out[0], compute_graph.flexml_layers[13].compute_node[0][3].out[0], compute_graph.flexml_layers[14].compute_node[0][3].out[0], compute_graph.flexml_layers[15].compute_node[0][3].out[0], compute_graph.flexml_layers[16].compute_node[0][3].out[0], compute_graph.flexml_layers[17].compute_node[0][3].out[0], compute_graph.flexml_layers[18].compute_node[0][3].out[0], compute_graph.flexml_layers[19].compute_node[0][3].out[0], compute_graph.flexml_layers[20].compute_node[0][3].out[0], compute_graph.flexml_layers[21].compute_node[0][3].out[0], compute_graph.flexml_layers[22].compute_node[0][3].out[0], compute_graph.flexml_layers[23].compute_node[0][3].out[0], compute_graph.flexml_layers[24].compute_node[0][3].out[0], compute_graph.flexml_layers[25].compute_node[0][3].out[0], compute_graph.flexml_layers[26].compute_node[0][3].out[0], compute_graph.flexml_layers[27].compute_node[0][3].out[0], compute_graph.flexml_layers[28].compute_node[0][3].out[0], compute_graph.flexml_layers[29].compute_node[0][3].out[0], compute_graph.flexml_layers[30].compute_node[0][3].out[0], compute_graph.flexml_layers[31].compute_node[0][3].out[0], compute_graph.flexml_layers[32].compute_node[0][3].out[0], compute_graph.flexml_layers[33].compute_node[0][3].out[0], compute_graph.flexml_layers[34].compute_node[0][3].out[0], compute_graph.flexml_layers[35].compute_node[0][3].out[0], compute_graph.flexml_layers[144].compute_node[0][3].out[0], compute_graph.flexml_layers[145].compute_node[0][3].out[0], compute_graph.flexml_layers[37].compute_node[0][3].out[0], compute_graph.flexml_layers[38].compute_node[0][3].out[0], compute_graph.flexml_layers[39].compute_node[0][3].out[0], compute_graph.flexml_layers[40].compute_node[0][3].out[0], compute_graph.flexml_layers[41].compute_node[0][3].out[0], compute_graph.flexml_layers[42].compute_node[0][3].out[0], compute_graph.flexml_layers[43].compute_node[0][3].out[0], compute_graph.flexml_layers[44].compute_node[0][3].out[0], compute_graph.flexml_layers[45].compute_node[0][3].out[0], compute_graph.flexml_layers[46].compute_node[0][3].out[0], compute_graph.flexml_layers[47].compute_node[0][3].out[0], compute_graph.flexml_layers[48].compute_node[0][3].out[0], compute_graph.flexml_layers[49].compute_node[0][3].out[0], compute_graph.flexml_layers[50].compute_node[0][3].out[0], compute_graph.flexml_layers[51].compute_node[0][3].out[0], compute_graph.flexml_layers[52].compute_node[0][3].out[0], compute_graph.flexml_layers[53].compute_node[0][3].out[0], compute_graph.flexml_layers[54].compute_node[0][3].out[0], compute_graph.flexml_layers[55].compute_node[0][3].out[0], compute_graph.flexml_layers[56].compute_node[0][3].out[0], compute_graph.flexml_layers[57].compute_node[0][3].out[0], compute_graph.flexml_layers[58].compute_node[0][3].out[0], compute_graph.flexml_layers[59].compute_node[0][3].out[0], compute_graph.flexml_layers[60].compute_node[0][3].out[0], compute_graph.flexml_layers[61].compute_node[0][3].out[0], compute_graph.flexml_layers[62].compute_node[0][3].out[0], compute_graph.flexml_layers[201].compute_node[0][3].out[0], compute_graph.flexml_layers[202].compute_node[0][3].out[0], compute_graph.flexml_layers[65].compute_node[0][3].out[0], compute_graph.flexml_layers[66].compute_node[0][3].out[0], compute_graph.flexml_layers[67].compute_node[0][3].out[0], compute_graph.flexml_layers[68].compute_node[0][3].out[0], compute_graph.flexml_layers[69].compute_node[0][3].out[0], compute_graph.flexml_layers[70].compute_node[0][3].out[0], compute_graph.flexml_layers[71].compute_node[0][3].out[0], compute_graph.flexml_layers[72].compute_node[0][3].out[0], compute_graph.flexml_layers[73].compute_node[0][3].out[0], compute_graph.flexml_layers[74].compute_node[0][3].out[0], compute_graph.flexml_layers[75].compute_node[0][3].out[0], compute_graph.flexml_layers[76].compute_node[0][3].out[0], compute_graph.flexml_layers[77].compute_node[0][3].out[0], compute_graph.flexml_layers[78].compute_node[0][3].out[0], compute_graph.flexml_layers[79].compute_node[0][3].out[0], compute_graph.flexml_layers[80].compute_node[0][3].out[0], compute_graph.flexml_layers[81].compute_node[0][3].out[0], compute_graph.flexml_layers[82].compute_node[0][3].out[0], compute_graph.flexml_layers[83].compute_node[0][3].out[0], compute_graph.flexml_layers[84].compute_node[0][3].out[0], compute_graph.flexml_layers[85].compute_node[0][3].out[0], compute_graph.flexml_layers[86].compute_node[0][3].out[0], compute_graph.flexml_layers[87].compute_node[0][3].out[0], compute_graph.flexml_layers[88].compute_node[0][3].out[0], compute_graph.flexml_layers[89].compute_node[0][3].out[0], compute_graph.flexml_layers[90].compute_node[0][3].out[0], compute_graph.flexml_layers[91].compute_node[0][3].out[0], compute_graph.flexml_layers[92].compute_node[0][3].out[0], compute_graph.flexml_layers[93].compute_node[0][3].out[0], compute_graph.flexml_layers[94].compute_node[0][3].out[0], compute_graph.flexml_layers[95].compute_node[0][3].out[0], compute_graph.flexml_layers[96].compute_node[0][3].out[0], compute_graph.flexml_layers[97].compute_node[0][3].out[0], compute_graph.flexml_layers[98].compute_node[0][3].out[0], compute_graph.flexml_layers[99].compute_node[0][3].out[0], compute_graph.flexml_layers[100].compute_node[0][3].out[0], compute_graph.flexml_layers[101].compute_node[0][3].out[0], compute_graph.flexml_layers[102].compute_node[0][3].out[0], compute_graph.flexml_layers[103].compute_node[0][3].out[0], compute_graph.flexml_layers[104].compute_node[0][3].out[0], compute_graph.flexml_layers[105].compute_node[0][3].out[0], compute_graph.flexml_layers[106].compute_node[0][3].out[0], compute_graph.flexml_layers[107].compute_node[0][3].out[0], compute_graph.flexml_layers[108].compute_node[0][3].out[0], compute_graph.flexml_layers[109].compute_node[0][3].out[0], compute_graph.flexml_layers[110].compute_node[0][3].out[0], compute_graph.flexml_layers[111].compute_node[0][3].out[0], compute_graph.flexml_layers[112].compute_node[0][3].out[0], compute_graph.flexml_layers[113].compute_node[0][3].out[0], compute_graph.flexml_layers[114].compute_node[0][3].out[0], compute_graph.flexml_layers[115].compute_node[0][3].out[0], compute_graph.flexml_layers[116].compute_node[0][3].out[0], compute_graph.flexml_layers[117].compute_node[0][3].out[0], compute_graph.flexml_layers[118].compute_node[0][3].out[0], compute_graph.flexml_layers[119].compute_node[0][3].out[0], compute_graph.flexml_layers[120].compute_node[0][3].out[0], compute_graph.flexml_layers[121].compute_node[0][3].out[0], compute_graph.flexml_layers[122].compute_node[0][3].out[0], compute_graph.flexml_layers[123].compute_node[0][3].out[0], compute_graph.flexml_layers[124].compute_node[0][3].out[0], compute_graph.flexml_layers[125].compute_node[0][3].out[0], compute_graph.flexml_layers[126].compute_node[0][3].out[0], compute_graph.flexml_layers[127].compute_node[0][3].out[0], compute_graph.flexml_layers[128].compute_node[0][3].out[0], compute_graph.flexml_layers[129].compute_node[0][3].out[0], compute_graph.flexml_layers[130].compute_node[0][3].out[0], compute_graph.flexml_layers[131].compute_node[0][3].out[0], compute_graph.flexml_layers[132].compute_node[0][3].out[0], compute_graph.flexml_layers[133].compute_node[0][3].out[0], compute_graph.flexml_layers[134].compute_node[0][3].out[0], compute_graph.flexml_layers[135].compute_node[0][3].out[0], compute_graph.flexml_layers[136].compute_node[0][3].out[0], compute_graph.flexml_layers[137].compute_node[0][3].out[0], compute_graph.flexml_layers[138].compute_node[0][3].out[0], compute_graph.flexml_layers[139].compute_node[0][3].out[0], compute_graph.flexml_layers[140].compute_node[0][3].out[0], compute_graph.flexml_layers[141].compute_node[0][3].out[0], compute_graph.flexml_layers[142].compute_node[0][3].out[0], compute_graph.flexml_layers[143].compute_node[0][3].out[0], compute_graph.flexml_layers[146].compute_node[0][3].out[0], compute_graph.flexml_layers[147].compute_node[0][3].out[0], compute_graph.flexml_layers[148].compute_node[0][3].out[0], compute_graph.flexml_layers[149].compute_node[0][3].out[0], compute_graph.flexml_layers[150].compute_node[0][3].out[0], compute_graph.flexml_layers[151].compute_node[0][3].out[0], compute_graph.flexml_layers[152].compute_node[0][3].out[0], compute_graph.flexml_layers[153].compute_node[0][3].out[0], compute_graph.flexml_layers[154].compute_node[0][3].out[0], compute_graph.flexml_layers[155].compute_node[0][3].out[0], compute_graph.flexml_layers[156].compute_node[0][3].out[0], compute_graph.flexml_layers[157].compute_node[0][3].out[0], compute_graph.flexml_layers[158].compute_node[0][3].out[0], compute_graph.flexml_layers[159].compute_node[0][3].out[0], compute_graph.flexml_layers[160].compute_node[0][3].out[0], compute_graph.flexml_layers[161].compute_node[0][3].out[0], compute_graph.flexml_layers[162].compute_node[0][3].out[0], compute_graph.flexml_layers[163].compute_node[0][3].out[0], compute_graph.flexml_layers[164].compute_node[0][3].out[0], compute_graph.flexml_layers[165].compute_node[0][3].out[0], compute_graph.flexml_layers[166].compute_node[0][3].out[0], compute_graph.flexml_layers[167].compute_node[0][3].out[0], compute_graph.flexml_layers[168].compute_node[0][3].out[0], compute_graph.flexml_layers[169].compute_node[0][3].out[0], compute_graph.flexml_layers[170].compute_node[0][3].out[0], compute_graph.flexml_layers[171].compute_node[0][3].out[0], compute_graph.flexml_layers[172].compute_node[0][3].out[0], compute_graph.flexml_layers[173].compute_node[0][3].out[0], compute_graph.flexml_layers[174].compute_node[0][3].out[0], compute_graph.flexml_layers[175].compute_node[0][3].out[0], compute_graph.flexml_layers[176].compute_node[0][3].out[0], compute_graph.flexml_layers[177].compute_node[0][3].out[0], compute_graph.flexml_layers[178].compute_node[0][3].out[0], compute_graph.flexml_layers[179].compute_node[0][3].out[0], compute_graph.flexml_layers[180].compute_node[0][3].out[0], compute_graph.flexml_layers[181].compute_node[0][3].out[0], compute_graph.flexml_layers[182].compute_node[0][3].out[0], compute_graph.flexml_layers[183].compute_node[0][3].out[0], compute_graph.flexml_layers[184].compute_node[0][3].out[0], compute_graph.flexml_layers[185].compute_node[0][3].out[0], compute_graph.flexml_layers[186].compute_node[0][3].out[0], compute_graph.flexml_layers[187].compute_node[0][3].out[0], compute_graph.flexml_layers[188].compute_node[0][3].out[0], compute_graph.flexml_layers[189].compute_node[0][3].out[0], compute_graph.flexml_layers[190].compute_node[0][3].out[0], compute_graph.flexml_layers[191].compute_node[0][3].out[0], compute_graph.flexml_layers[192].compute_node[0][3].out[0], compute_graph.flexml_layers[193].compute_node[0][3].out[0], compute_graph.flexml_layers[194].compute_node[0][3].out[0], compute_graph.flexml_layers[195].compute_node[0][3].out[0], compute_graph.flexml_layers[196].compute_node[0][3].out[0], compute_graph.flexml_layers[197].compute_node[0][3].out[0], compute_graph.flexml_layers[198].compute_node[0][3].out[0], compute_graph.flexml_layers[199].compute_node[0][3].out[0], compute_graph.flexml_layers[200].compute_node[0][3].out[0], compute_graph.flexml_layers[203].compute_node[0][3].out[0], compute_graph.flexml_layers[204].compute_node[0][3].out[0], compute_graph.flexml_layers[205].compute_node[0][3].out[0], compute_graph.flexml_layers[206].compute_node[0][3].out[0], compute_graph.flexml_layers[207].compute_node[0][3].out[0], compute_graph.flexml_layers[208].compute_node[0][3].out[0], compute_graph.flexml_layers[209].compute_node[0][3].out[0], compute_graph.flexml_layers[210].compute_node[0][3].out[0], compute_graph.flexml_layers[211].compute_node[0][3].out[0], compute_graph.flexml_layers[212].compute_node[0][3].out[0], compute_graph.flexml_layers[213].compute_node[0][3].out[0], compute_graph.flexml_layers[214].compute_node[0][3].out[0], compute_graph.flexml_layers[215].compute_node[0][3].out[0], compute_graph.flexml_layers[216].compute_node[0][3].out[0], compute_graph.flexml_layers[217].compute_node[0][3].out[0], compute_graph.flexml_layers[218].compute_node[0][3].out[0], compute_graph.flexml_layers[219].compute_node[0][3].out[0], compute_graph.flexml_layers[220].compute_node[0][3].out[0], compute_graph.flexml_layers[221].compute_node[0][3].out[0], compute_graph.flexml_layers[222].compute_node[0][3].out[0], compute_graph.flexml_layers[223].compute_node[0][3].out[0], compute_graph.flexml_layers[224].compute_node[0][3].out[0], compute_graph.flexml_layers[225].compute_node[0][3].out[0], compute_graph.flexml_layers[226].compute_node[0][3].out[0], compute_graph.flexml_layers[227].compute_node[0][3].out[0], compute_graph.flexml_layers[228].compute_node[0][3].out[0], compute_graph.flexml_layers[229].compute_node[0][3].out[0], compute_graph.flexml_layers[230].compute_node[0][3].out[0], compute_graph.flexml_layers[231].compute_node[0][3].out[0], compute_graph.flexml_layers[232].compute_node[0][3].out[0], compute_graph.flexml_layers[233].compute_node[0][3].out[0], compute_graph.flexml_layers[234].compute_node[0][3].out[0], compute_graph.flexml_layers[235].compute_node[0][3].out[0], compute_graph.flexml_layers[236].compute_node[0][3].out[0]) + +Column 0Row 3Channel 1 + +i7_pi1(compute_graph.flexml_layers[36].compute_node[0][3].in[1], compute_graph.templated_graph_260.mk[0][3].in[1], compute_graph.templated_graph_268.mk[0][3].in[1], compute_graph.templated_graph_273.mk[0][3].in[1], compute_graph.flexml_layers[3].compute_node[0][3].in[1], compute_graph.flexml_layers[8].compute_node[0][3].in[1], compute_graph.flexml_layers[9].compute_node[0][3].in[1], compute_graph.flexml_layers[10].compute_node[0][3].in[1], compute_graph.flexml_layers[11].compute_node[0][3].in[1], compute_graph.flexml_layers[12].compute_node[0][3].in[1], compute_graph.flexml_layers[13].compute_node[0][3].in[1], compute_graph.flexml_layers[14].compute_node[0][3].in[1], compute_graph.flexml_layers[15].compute_node[0][3].in[1], compute_graph.flexml_layers[16].compute_node[0][3].in[1], compute_graph.flexml_layers[17].compute_node[0][3].in[1], compute_graph.flexml_layers[19].compute_node[0][3].in[1], compute_graph.flexml_layers[20].compute_node[0][3].in[1], compute_graph.flexml_layers[25].compute_node[0][3].in[1], compute_graph.flexml_layers[26].compute_node[0][3].in[1], compute_graph.flexml_layers[27].compute_node[0][3].in[1], compute_graph.flexml_layers[29].compute_node[0][3].in[1], compute_graph.flexml_layers[30].compute_node[0][3].in[1], compute_graph.flexml_layers[35].compute_node[0][3].in[1], compute_graph.flexml_layers[144].compute_node[0][3].in[1], compute_graph.flexml_layers[37].compute_node[0][3].in[1], compute_graph.flexml_layers[39].compute_node[0][3].in[1], compute_graph.flexml_layers[40].compute_node[0][3].in[1], compute_graph.flexml_layers[45].compute_node[0][3].in[1], compute_graph.flexml_layers[46].compute_node[0][3].in[1], compute_graph.flexml_layers[51].compute_node[0][3].in[1], compute_graph.flexml_layers[56].compute_node[0][3].in[1], compute_graph.flexml_layers[57].compute_node[0][3].in[1], compute_graph.flexml_layers[62].compute_node[0][3].in[1], compute_graph.flexml_layers[201].compute_node[0][3].in[1], compute_graph.flexml_layers[202].compute_node[0][3].in[1], compute_graph.flexml_layers[67].compute_node[0][3].in[1], compute_graph.flexml_layers[68].compute_node[0][3].in[1], compute_graph.flexml_layers[73].compute_node[0][3].in[1], compute_graph.flexml_layers[78].compute_node[0][3].in[1], compute_graph.flexml_layers[79].compute_node[0][3].in[1], compute_graph.flexml_layers[84].compute_node[0][3].in[1], compute_graph.flexml_layers[89].compute_node[0][3].in[1], compute_graph.flexml_layers[90].compute_node[0][3].in[1], compute_graph.flexml_layers[95].compute_node[0][3].in[1], compute_graph.flexml_layers[101].compute_node[0][3].in[1], compute_graph.flexml_layers[102].compute_node[0][3].in[1], compute_graph.flexml_layers[107].compute_node[0][3].in[1], compute_graph.flexml_layers[108].compute_node[0][3].in[1], compute_graph.flexml_layers[113].compute_node[0][3].in[1], compute_graph.flexml_layers[119].compute_node[0][3].in[1], compute_graph.flexml_layers[120].compute_node[0][3].in[1], compute_graph.flexml_layers[125].compute_node[0][3].in[1], compute_graph.flexml_layers[126].compute_node[0][3].in[1], compute_graph.flexml_layers[131].compute_node[0][3].in[1], compute_graph.flexml_layers[137].compute_node[0][3].in[1], compute_graph.flexml_layers[138].compute_node[0][3].in[1], compute_graph.flexml_layers[143].compute_node[0][3].in[1], compute_graph.flexml_layers[149].compute_node[0][3].in[1], compute_graph.flexml_layers[155].compute_node[0][3].in[1], compute_graph.flexml_layers[156].compute_node[0][3].in[1], compute_graph.flexml_layers[161].compute_node[0][3].in[1], compute_graph.flexml_layers[162].compute_node[0][3].in[1], compute_graph.flexml_layers[167].compute_node[0][3].in[1], compute_graph.flexml_layers[173].compute_node[0][3].in[1], compute_graph.flexml_layers[174].compute_node[0][3].in[1], compute_graph.flexml_layers[179].compute_node[0][3].in[1], compute_graph.flexml_layers[180].compute_node[0][3].in[1], compute_graph.flexml_layers[186].compute_node[0][3].in[1], compute_graph.flexml_layers[188].compute_node[0][3].in[1], compute_graph.flexml_layers[189].compute_node[0][3].in[1], compute_graph.flexml_layers[194].compute_node[0][3].in[1], compute_graph.flexml_layers[207].compute_node[0][3].in[1], compute_graph.flexml_layers[211].compute_node[0][3].in[1], compute_graph.flexml_layers[212].compute_node[0][3].in[1], compute_graph.flexml_layers[217].compute_node[0][3].in[1], compute_graph.flexml_layers[221].compute_node[0][3].in[1], compute_graph.flexml_layers[222].compute_node[0][3].in[1], compute_graph.flexml_layers[227].compute_node[0][3].in[1], compute_graph.flexml_layers[231].compute_node[0][3].in[1], compute_graph.flexml_layers[232].compute_node[0][3].in[1], compute_graph.flexml_layers[233].compute_node[0][3].in[1]) + +Column 1Row 0Channel 0 + +i8_pi0(compute_graph.templated_graph_4.trans_comp_nd[1][0].in[0], compute_graph.flexml_layers[36].compute_node[1][0].in[0], compute_graph.templated_graph_231.mk[1][0].in[0], compute_graph.templated_graph_254.mk[1][0].in[0], compute_graph.templated_graph_258.compute_node[1][0].in[0], compute_graph.templated_graph_259.compute_node[1][0].in[0], compute_graph.templated_graph_260.mk[1][0].in[0], compute_graph.templated_graph_263.compute_node[1][0].in[0], compute_graph.templated_graph_266.compute_node[1][0].in[0], compute_graph.templated_graph_268.mk[1][0].in[0], compute_graph.templated_graph_273.mk[1][0].in[0], compute_graph.templated_graph_274.mk[1][0].in[0], compute_graph.flexml_layers[63].compute_node[1][0].in[0], compute_graph.flexml_layers[64].compute_node[1][0].in[0], compute_graph.templated_graph_293.mk[1][0].in[0], compute_graph.templated_graph_298.compute_node[1][0].in[0], compute_graph.templated_graph_300.compute_node[1][0].in[0], compute_graph.flexml_layers[0].compute_node[1][0].in[0], compute_graph.flexml_layers[1].compute_node[1][0].in[0], compute_graph.flexml_layers[1].compute_node[1][0].in[2], compute_graph.flexml_layers[2].compute_node[1][0].in[0], compute_graph.flexml_layers[2].compute_node[1][0].in[2], compute_graph.flexml_layers[3].compute_node[1][0].in[0], compute_graph.flexml_layers[4].compute_node[1][0].in[0], compute_graph.flexml_layers[5].compute_node[1][0].in[0], compute_graph.flexml_layers[6].compute_node[1][0].in[0], compute_graph.flexml_layers[7].compute_node[1][0].in[0], compute_graph.flexml_layers[7].compute_node[1][0].in[2], compute_graph.flexml_layers[8].compute_node[1][0].in[0], compute_graph.flexml_layers[9].compute_node[1][0].in[0], compute_graph.flexml_layers[9].compute_node[1][0].in[2], compute_graph.flexml_layers[10].compute_node[1][0].in[0], compute_graph.flexml_layers[11].compute_node[1][0].in[0], compute_graph.flexml_layers[12].compute_node[1][0].in[0], compute_graph.flexml_layers[13].compute_node[1][0].in[0], compute_graph.flexml_layers[14].compute_node[1][0].in[0], compute_graph.flexml_layers[15].compute_node[1][0].in[0], compute_graph.flexml_layers[15].compute_node[1][0].in[2], compute_graph.flexml_layers[16].compute_node[1][0].in[0], compute_graph.flexml_layers[17].compute_node[1][0].in[0], compute_graph.flexml_layers[18].compute_node[1][0].in[0], compute_graph.flexml_layers[19].compute_node[1][0].in[0], compute_graph.flexml_layers[20].compute_node[1][0].in[0], compute_graph.flexml_layers[21].compute_node[1][0].in[0], compute_graph.flexml_layers[22].compute_node[1][0].in[0], compute_graph.flexml_layers[23].compute_node[1][0].in[0], compute_graph.flexml_layers[24].compute_node[1][0].in[0], compute_graph.flexml_layers[24].compute_node[1][0].in[2], compute_graph.flexml_layers[25].compute_node[1][0].in[0], compute_graph.flexml_layers[26].compute_node[1][0].in[0], compute_graph.flexml_layers[27].compute_node[1][0].in[0], compute_graph.flexml_layers[28].compute_node[1][0].in[0], compute_graph.flexml_layers[29].compute_node[1][0].in[0], compute_graph.flexml_layers[30].compute_node[1][0].in[0], compute_graph.flexml_layers[31].compute_node[1][0].in[0], compute_graph.flexml_layers[32].compute_node[1][0].in[0], compute_graph.flexml_layers[33].compute_node[1][0].in[0], compute_graph.flexml_layers[34].compute_node[1][0].in[0], compute_graph.flexml_layers[34].compute_node[1][0].in[2], compute_graph.flexml_layers[35].compute_node[1][0].in[0], compute_graph.flexml_layers[35].compute_node[1][0].in[2], compute_graph.flexml_layers[144].compute_node[1][0].in[0], compute_graph.flexml_layers[37].compute_node[1][0].in[0], compute_graph.flexml_layers[38].compute_node[1][0].in[0], compute_graph.flexml_layers[39].compute_node[1][0].in[0], compute_graph.flexml_layers[40].compute_node[1][0].in[0], compute_graph.flexml_layers[41].compute_node[1][0].in[0], compute_graph.flexml_layers[42].compute_node[1][0].in[0], compute_graph.flexml_layers[43].compute_node[1][0].in[0], compute_graph.flexml_layers[44].compute_node[1][0].in[0], compute_graph.flexml_layers[44].compute_node[1][0].in[2], compute_graph.flexml_layers[45].compute_node[1][0].in[0], compute_graph.flexml_layers[45].compute_node[1][0].in[2], compute_graph.flexml_layers[46].compute_node[1][0].in[0], compute_graph.flexml_layers[47].compute_node[1][0].in[0], compute_graph.flexml_layers[48].compute_node[1][0].in[0], compute_graph.flexml_layers[49].compute_node[1][0].in[0], compute_graph.flexml_layers[50].compute_node[1][0].in[0], compute_graph.flexml_layers[50].compute_node[1][0].in[2], compute_graph.flexml_layers[51].compute_node[1][0].in[0], compute_graph.flexml_layers[52].compute_node[1][0].in[0], compute_graph.flexml_layers[53].compute_node[1][0].in[0], compute_graph.flexml_layers[54].compute_node[1][0].in[0], compute_graph.flexml_layers[55].compute_node[1][0].in[0], compute_graph.flexml_layers[55].compute_node[1][0].in[2], compute_graph.flexml_layers[56].compute_node[1][0].in[0], compute_graph.flexml_layers[57].compute_node[1][0].in[0], compute_graph.flexml_layers[58].compute_node[1][0].in[0], compute_graph.flexml_layers[59].compute_node[1][0].in[0], compute_graph.flexml_layers[60].compute_node[1][0].in[0], compute_graph.flexml_layers[61].compute_node[1][0].in[0], compute_graph.flexml_layers[61].compute_node[1][0].in[2], compute_graph.flexml_layers[62].compute_node[1][0].in[0], compute_graph.flexml_layers[201].compute_node[1][0].in[0], compute_graph.flexml_layers[202].compute_node[1][0].in[0], compute_graph.flexml_layers[65].compute_node[1][0].in[0], compute_graph.flexml_layers[66].compute_node[1][0].in[0], compute_graph.flexml_layers[66].compute_node[1][0].in[2], compute_graph.flexml_layers[67].compute_node[1][0].in[0], compute_graph.flexml_layers[67].compute_node[1][0].in[2], compute_graph.flexml_layers[68].compute_node[1][0].in[0], compute_graph.flexml_layers[69].compute_node[1][0].in[0], compute_graph.flexml_layers[70].compute_node[1][0].in[0], compute_graph.flexml_layers[71].compute_node[1][0].in[0], compute_graph.flexml_layers[72].compute_node[1][0].in[0], compute_graph.flexml_layers[72].compute_node[1][0].in[2], compute_graph.flexml_layers[73].compute_node[1][0].in[0], compute_graph.flexml_layers[74].compute_node[1][0].in[0], compute_graph.flexml_layers[75].compute_node[1][0].in[0], compute_graph.flexml_layers[76].compute_node[1][0].in[0], compute_graph.flexml_layers[77].compute_node[1][0].in[0], compute_graph.flexml_layers[77].compute_node[1][0].in[2], compute_graph.flexml_layers[78].compute_node[1][0].in[0], compute_graph.flexml_layers[78].compute_node[1][0].in[2], compute_graph.flexml_layers[79].compute_node[1][0].in[0], compute_graph.flexml_layers[80].compute_node[1][0].in[0], compute_graph.flexml_layers[81].compute_node[1][0].in[0], compute_graph.flexml_layers[82].compute_node[1][0].in[0], compute_graph.flexml_layers[83].compute_node[1][0].in[0], compute_graph.flexml_layers[83].compute_node[1][0].in[2], compute_graph.flexml_layers[84].compute_node[1][0].in[0], compute_graph.flexml_layers[85].compute_node[1][0].in[0], compute_graph.flexml_layers[86].compute_node[1][0].in[0], compute_graph.flexml_layers[87].compute_node[1][0].in[0], compute_graph.flexml_layers[88].compute_node[1][0].in[0], compute_graph.flexml_layers[88].compute_node[1][0].in[2], compute_graph.flexml_layers[89].compute_node[1][0].in[0], compute_graph.flexml_layers[89].compute_node[1][0].in[2], compute_graph.flexml_layers[90].compute_node[1][0].in[0], compute_graph.flexml_layers[91].compute_node[1][0].in[0], compute_graph.flexml_layers[92].compute_node[1][0].in[0], compute_graph.flexml_layers[93].compute_node[1][0].in[0], compute_graph.flexml_layers[94].compute_node[1][0].in[0], compute_graph.flexml_layers[94].compute_node[1][0].in[2], compute_graph.flexml_layers[95].compute_node[1][0].in[0], compute_graph.flexml_layers[96].compute_node[1][0].in[0], compute_graph.flexml_layers[97].compute_node[1][0].in[0], compute_graph.flexml_layers[98].compute_node[1][0].in[0], compute_graph.flexml_layers[99].compute_node[1][0].in[0], compute_graph.flexml_layers[99].compute_node[1][0].in[2], compute_graph.flexml_layers[100].compute_node[1][0].in[0], compute_graph.flexml_layers[101].compute_node[1][0].in[0], compute_graph.flexml_layers[102].compute_node[1][0].in[0], compute_graph.flexml_layers[103].compute_node[1][0].in[0], compute_graph.flexml_layers[104].compute_node[1][0].in[0], compute_graph.flexml_layers[105].compute_node[1][0].in[0], compute_graph.flexml_layers[106].compute_node[1][0].in[0], compute_graph.flexml_layers[106].compute_node[1][0].in[2], compute_graph.flexml_layers[107].compute_node[1][0].in[0], compute_graph.flexml_layers[108].compute_node[1][0].in[0], compute_graph.flexml_layers[109].compute_node[1][0].in[0], compute_graph.flexml_layers[110].compute_node[1][0].in[0], compute_graph.flexml_layers[111].compute_node[1][0].in[0], compute_graph.flexml_layers[112].compute_node[1][0].in[0], compute_graph.flexml_layers[112].compute_node[1][0].in[2], compute_graph.flexml_layers[113].compute_node[1][0].in[0], compute_graph.flexml_layers[114].compute_node[1][0].in[0], compute_graph.flexml_layers[115].compute_node[1][0].in[0], compute_graph.flexml_layers[116].compute_node[1][0].in[0], compute_graph.flexml_layers[117].compute_node[1][0].in[0], compute_graph.flexml_layers[117].compute_node[1][0].in[2], compute_graph.flexml_layers[118].compute_node[1][0].in[0], compute_graph.flexml_layers[119].compute_node[1][0].in[0], compute_graph.flexml_layers[120].compute_node[1][0].in[0], compute_graph.flexml_layers[121].compute_node[1][0].in[0], compute_graph.flexml_layers[122].compute_node[1][0].in[0], compute_graph.flexml_layers[123].compute_node[1][0].in[0], compute_graph.flexml_layers[124].compute_node[1][0].in[0], compute_graph.flexml_layers[124].compute_node[1][0].in[2], compute_graph.flexml_layers[125].compute_node[1][0].in[0], compute_graph.flexml_layers[125].compute_node[1][0].in[2], compute_graph.flexml_layers[126].compute_node[1][0].in[0], compute_graph.flexml_layers[127].compute_node[1][0].in[0], compute_graph.flexml_layers[128].compute_node[1][0].in[0], compute_graph.flexml_layers[129].compute_node[1][0].in[0], compute_graph.flexml_layers[130].compute_node[1][0].in[0], compute_graph.flexml_layers[130].compute_node[1][0].in[2], compute_graph.flexml_layers[131].compute_node[1][0].in[0], compute_graph.flexml_layers[132].compute_node[1][0].in[0], compute_graph.flexml_layers[133].compute_node[1][0].in[0], compute_graph.flexml_layers[134].compute_node[1][0].in[0], compute_graph.flexml_layers[135].compute_node[1][0].in[0], compute_graph.flexml_layers[135].compute_node[1][0].in[2], compute_graph.flexml_layers[136].compute_node[1][0].in[0], compute_graph.flexml_layers[137].compute_node[1][0].in[0], compute_graph.flexml_layers[138].compute_node[1][0].in[0], compute_graph.flexml_layers[139].compute_node[1][0].in[0], compute_graph.flexml_layers[140].compute_node[1][0].in[0], compute_graph.flexml_layers[141].compute_node[1][0].in[0], compute_graph.flexml_layers[142].compute_node[1][0].in[0], compute_graph.flexml_layers[142].compute_node[1][0].in[2], compute_graph.flexml_layers[143].compute_node[1][0].in[0], compute_graph.flexml_layers[145].compute_node[1][0].in[0], compute_graph.flexml_layers[146].compute_node[1][0].in[0], compute_graph.flexml_layers[147].compute_node[1][0].in[0], compute_graph.flexml_layers[148].compute_node[1][0].in[0], compute_graph.flexml_layers[148].compute_node[1][0].in[2], compute_graph.flexml_layers[149].compute_node[1][0].in[0], compute_graph.flexml_layers[150].compute_node[1][0].in[0], compute_graph.flexml_layers[151].compute_node[1][0].in[0], compute_graph.flexml_layers[152].compute_node[1][0].in[0], compute_graph.flexml_layers[153].compute_node[1][0].in[0], compute_graph.flexml_layers[153].compute_node[1][0].in[2], compute_graph.flexml_layers[154].compute_node[1][0].in[0], compute_graph.flexml_layers[155].compute_node[1][0].in[0], compute_graph.flexml_layers[156].compute_node[1][0].in[0], compute_graph.flexml_layers[157].compute_node[1][0].in[0], compute_graph.flexml_layers[158].compute_node[1][0].in[0], compute_graph.flexml_layers[159].compute_node[1][0].in[0], compute_graph.flexml_layers[160].compute_node[1][0].in[0], compute_graph.flexml_layers[160].compute_node[1][0].in[2], compute_graph.flexml_layers[161].compute_node[1][0].in[0], compute_graph.flexml_layers[161].compute_node[1][0].in[2], compute_graph.flexml_layers[162].compute_node[1][0].in[0], compute_graph.flexml_layers[163].compute_node[1][0].in[0], compute_graph.flexml_layers[164].compute_node[1][0].in[0], compute_graph.flexml_layers[165].compute_node[1][0].in[0], compute_graph.flexml_layers[166].compute_node[1][0].in[0], compute_graph.flexml_layers[166].compute_node[1][0].in[2], compute_graph.flexml_layers[167].compute_node[1][0].in[0], compute_graph.flexml_layers[168].compute_node[1][0].in[0], compute_graph.flexml_layers[169].compute_node[1][0].in[0], compute_graph.flexml_layers[170].compute_node[1][0].in[0], compute_graph.flexml_layers[171].compute_node[1][0].in[0], compute_graph.flexml_layers[171].compute_node[1][0].in[2], compute_graph.flexml_layers[172].compute_node[1][0].in[0], compute_graph.flexml_layers[173].compute_node[1][0].in[0], compute_graph.flexml_layers[174].compute_node[1][0].in[0], compute_graph.flexml_layers[175].compute_node[1][0].in[0], compute_graph.flexml_layers[176].compute_node[1][0].in[0], compute_graph.flexml_layers[177].compute_node[1][0].in[0], compute_graph.flexml_layers[178].compute_node[1][0].in[0], compute_graph.flexml_layers[178].compute_node[1][0].in[2], compute_graph.flexml_layers[179].compute_node[1][0].in[0], compute_graph.flexml_layers[179].compute_node[1][0].in[2], compute_graph.flexml_layers[180].compute_node[1][0].in[0], compute_graph.flexml_layers[181].compute_node[1][0].in[0], compute_graph.flexml_layers[182].compute_node[1][0].in[0], compute_graph.flexml_layers[183].compute_node[1][0].in[0], compute_graph.flexml_layers[184].compute_node[1][0].in[0], compute_graph.flexml_layers[184].compute_node[1][0].in[2], compute_graph.flexml_layers[185].compute_node[1][0].in[0], compute_graph.flexml_layers[186].compute_node[1][0].in[0], compute_graph.flexml_layers[187].compute_node[1][0].in[0], compute_graph.flexml_layers[188].compute_node[1][0].in[0], compute_graph.flexml_layers[188].compute_node[1][0].in[2], compute_graph.flexml_layers[189].compute_node[1][0].in[0], compute_graph.flexml_layers[190].compute_node[1][0].in[0], compute_graph.flexml_layers[191].compute_node[1][0].in[0], compute_graph.flexml_layers[191].compute_node[1][0].in[2], compute_graph.flexml_layers[192].compute_node[1][0].in[0], compute_graph.flexml_layers[192].compute_node[1][0].in[2], compute_graph.flexml_layers[193].compute_node[1][0].in[0], compute_graph.flexml_layers[193].compute_node[1][0].in[2], compute_graph.flexml_layers[194].compute_node[1][0].in[0], compute_graph.flexml_layers[195].compute_node[1][0].in[0], compute_graph.flexml_layers[196].compute_node[1][0].in[0], compute_graph.flexml_layers[196].compute_node[1][0].in[2], compute_graph.flexml_layers[197].compute_node[1][0].in[0], compute_graph.flexml_layers[197].compute_node[1][0].in[2], compute_graph.flexml_layers[198].compute_node[1][0].in[0], compute_graph.flexml_layers[199].compute_node[1][0].in[0], compute_graph.flexml_layers[200].compute_node[1][0].in[0], compute_graph.flexml_layers[203].compute_node[1][0].in[0], compute_graph.flexml_layers[204].compute_node[1][0].in[0], compute_graph.flexml_layers[204].compute_node[1][0].in[2], compute_graph.flexml_layers[205].compute_node[1][0].in[0], compute_graph.flexml_layers[205].compute_node[1][0].in[2], compute_graph.flexml_layers[206].compute_node[1][0].in[0], compute_graph.flexml_layers[206].compute_node[1][0].in[2], compute_graph.flexml_layers[207].compute_node[1][0].in[0], compute_graph.flexml_layers[208].compute_node[1][0].in[0], compute_graph.flexml_layers[209].compute_node[1][0].in[0], compute_graph.flexml_layers[209].compute_node[1][0].in[2], compute_graph.flexml_layers[210].compute_node[1][0].in[0], compute_graph.flexml_layers[210].compute_node[1][0].in[2], compute_graph.flexml_layers[211].compute_node[1][0].in[0], compute_graph.flexml_layers[212].compute_node[1][0].in[0], compute_graph.flexml_layers[213].compute_node[1][0].in[0], compute_graph.flexml_layers[214].compute_node[1][0].in[0], compute_graph.flexml_layers[214].compute_node[1][0].in[2], compute_graph.flexml_layers[215].compute_node[1][0].in[0], compute_graph.flexml_layers[215].compute_node[1][0].in[2], compute_graph.flexml_layers[216].compute_node[1][0].in[0], compute_graph.flexml_layers[216].compute_node[1][0].in[2], compute_graph.flexml_layers[217].compute_node[1][0].in[0], compute_graph.flexml_layers[218].compute_node[1][0].in[0], compute_graph.flexml_layers[219].compute_node[1][0].in[0], compute_graph.flexml_layers[219].compute_node[1][0].in[2], compute_graph.flexml_layers[220].compute_node[1][0].in[0], compute_graph.flexml_layers[220].compute_node[1][0].in[2], compute_graph.flexml_layers[221].compute_node[1][0].in[0], compute_graph.flexml_layers[222].compute_node[1][0].in[0], compute_graph.flexml_layers[223].compute_node[1][0].in[0], compute_graph.flexml_layers[224].compute_node[1][0].in[0], compute_graph.flexml_layers[224].compute_node[1][0].in[2], compute_graph.flexml_layers[225].compute_node[1][0].in[0], compute_graph.flexml_layers[225].compute_node[1][0].in[2], compute_graph.flexml_layers[226].compute_node[1][0].in[0], compute_graph.flexml_layers[226].compute_node[1][0].in[2], compute_graph.flexml_layers[227].compute_node[1][0].in[0], compute_graph.flexml_layers[228].compute_node[1][0].in[0], compute_graph.flexml_layers[229].compute_node[1][0].in[0], compute_graph.flexml_layers[229].compute_node[1][0].in[2], compute_graph.flexml_layers[230].compute_node[1][0].in[0], compute_graph.flexml_layers[230].compute_node[1][0].in[2], compute_graph.flexml_layers[231].compute_node[1][0].in[0], compute_graph.flexml_layers[232].compute_node[1][0].in[0], compute_graph.flexml_layers[233].compute_node[1][0].in[0], compute_graph.flexml_layers[234].compute_node[1][0].in[0], compute_graph.flexml_layers[235].compute_node[1][0].in[0], compute_graph.flexml_layers[235].compute_node[1][0].in[2], compute_graph.flexml_layers[236].compute_node[1][0].in[0]) + +Column 1Row 0Channel 0 + +i8_po0(compute_graph.templated_graph_4.trans_comp_nd[1][0].out[0], compute_graph.flexml_layers[36].compute_node[1][0].out[0], compute_graph.templated_graph_231.mk[1][0].out[0], compute_graph.templated_graph_254.mk[1][0].out[0], compute_graph.templated_graph_258.compute_node[1][0].out[0], compute_graph.templated_graph_259.compute_node[1][0].out[0], compute_graph.templated_graph_260.mk[1][0].out[0], compute_graph.templated_graph_263.compute_node[1][0].out[0], compute_graph.templated_graph_266.compute_node[1][0].out[0], compute_graph.templated_graph_268.mk[1][0].out[0], compute_graph.templated_graph_273.mk[1][0].out[0], compute_graph.templated_graph_274.mk[1][0].out[0], compute_graph.flexml_layers[63].compute_node[1][0].out[0], compute_graph.templated_graph_293.mk[1][0].out[0], compute_graph.templated_graph_298.compute_node[1][0].out[0], compute_graph.templated_graph_300.compute_node[1][0].out[0], compute_graph.flexml_layers[0].compute_node[1][0].out[0], compute_graph.flexml_layers[1].compute_node[1][0].out[0], compute_graph.flexml_layers[2].compute_node[1][0].out[0], compute_graph.flexml_layers[3].compute_node[1][0].out[0], compute_graph.flexml_layers[4].compute_node[1][0].out[0], compute_graph.flexml_layers[5].compute_node[1][0].out[0], compute_graph.flexml_layers[6].compute_node[1][0].out[0], compute_graph.flexml_layers[7].compute_node[1][0].out[0], compute_graph.flexml_layers[8].compute_node[1][0].out[0], compute_graph.flexml_layers[9].compute_node[1][0].out[0], compute_graph.flexml_layers[10].compute_node[1][0].out[0], compute_graph.flexml_layers[11].compute_node[1][0].out[0], compute_graph.flexml_layers[12].compute_node[1][0].out[0], compute_graph.flexml_layers[13].compute_node[1][0].out[0], compute_graph.flexml_layers[14].compute_node[1][0].out[0], compute_graph.flexml_layers[15].compute_node[1][0].out[0], compute_graph.flexml_layers[16].compute_node[1][0].out[0], compute_graph.flexml_layers[17].compute_node[1][0].out[0], compute_graph.flexml_layers[18].compute_node[1][0].out[0], compute_graph.flexml_layers[19].compute_node[1][0].out[0], compute_graph.flexml_layers[20].compute_node[1][0].out[0], compute_graph.flexml_layers[21].compute_node[1][0].out[0], compute_graph.flexml_layers[22].compute_node[1][0].out[0], compute_graph.flexml_layers[23].compute_node[1][0].out[0], compute_graph.flexml_layers[24].compute_node[1][0].out[0], compute_graph.flexml_layers[25].compute_node[1][0].out[0], compute_graph.flexml_layers[26].compute_node[1][0].out[0], compute_graph.flexml_layers[27].compute_node[1][0].out[0], compute_graph.flexml_layers[28].compute_node[1][0].out[0], compute_graph.flexml_layers[29].compute_node[1][0].out[0], compute_graph.flexml_layers[30].compute_node[1][0].out[0], compute_graph.flexml_layers[31].compute_node[1][0].out[0], compute_graph.flexml_layers[32].compute_node[1][0].out[0], compute_graph.flexml_layers[33].compute_node[1][0].out[0], compute_graph.flexml_layers[34].compute_node[1][0].out[0], compute_graph.flexml_layers[35].compute_node[1][0].out[0], compute_graph.flexml_layers[144].compute_node[1][0].out[0], compute_graph.flexml_layers[37].compute_node[1][0].out[0], compute_graph.flexml_layers[38].compute_node[1][0].out[0], compute_graph.flexml_layers[39].compute_node[1][0].out[0], compute_graph.flexml_layers[40].compute_node[1][0].out[0], compute_graph.flexml_layers[41].compute_node[1][0].out[0], compute_graph.flexml_layers[42].compute_node[1][0].out[0], compute_graph.flexml_layers[43].compute_node[1][0].out[0], compute_graph.flexml_layers[44].compute_node[1][0].out[0], compute_graph.flexml_layers[45].compute_node[1][0].out[0], compute_graph.flexml_layers[46].compute_node[1][0].out[0], compute_graph.flexml_layers[47].compute_node[1][0].out[0], compute_graph.flexml_layers[48].compute_node[1][0].out[0], compute_graph.flexml_layers[49].compute_node[1][0].out[0], compute_graph.flexml_layers[50].compute_node[1][0].out[0], compute_graph.flexml_layers[51].compute_node[1][0].out[0], compute_graph.flexml_layers[52].compute_node[1][0].out[0], compute_graph.flexml_layers[53].compute_node[1][0].out[0], compute_graph.flexml_layers[54].compute_node[1][0].out[0], compute_graph.flexml_layers[55].compute_node[1][0].out[0], compute_graph.flexml_layers[56].compute_node[1][0].out[0], compute_graph.flexml_layers[57].compute_node[1][0].out[0], compute_graph.flexml_layers[58].compute_node[1][0].out[0], compute_graph.flexml_layers[59].compute_node[1][0].out[0], compute_graph.flexml_layers[60].compute_node[1][0].out[0], compute_graph.flexml_layers[61].compute_node[1][0].out[0], compute_graph.flexml_layers[62].compute_node[1][0].out[0], compute_graph.flexml_layers[201].compute_node[1][0].out[0], compute_graph.flexml_layers[202].compute_node[1][0].out[0], compute_graph.flexml_layers[64].compute_node[1][0].out[0], compute_graph.flexml_layers[65].compute_node[1][0].out[0], compute_graph.flexml_layers[66].compute_node[1][0].out[0], compute_graph.flexml_layers[67].compute_node[1][0].out[0], compute_graph.flexml_layers[68].compute_node[1][0].out[0], compute_graph.flexml_layers[69].compute_node[1][0].out[0], compute_graph.flexml_layers[70].compute_node[1][0].out[0], compute_graph.flexml_layers[71].compute_node[1][0].out[0], compute_graph.flexml_layers[72].compute_node[1][0].out[0], compute_graph.flexml_layers[73].compute_node[1][0].out[0], compute_graph.flexml_layers[74].compute_node[1][0].out[0], compute_graph.flexml_layers[75].compute_node[1][0].out[0], compute_graph.flexml_layers[76].compute_node[1][0].out[0], compute_graph.flexml_layers[77].compute_node[1][0].out[0], compute_graph.flexml_layers[78].compute_node[1][0].out[0], compute_graph.flexml_layers[79].compute_node[1][0].out[0], compute_graph.flexml_layers[80].compute_node[1][0].out[0], compute_graph.flexml_layers[81].compute_node[1][0].out[0], compute_graph.flexml_layers[82].compute_node[1][0].out[0], compute_graph.flexml_layers[83].compute_node[1][0].out[0], compute_graph.flexml_layers[84].compute_node[1][0].out[0], compute_graph.flexml_layers[85].compute_node[1][0].out[0], compute_graph.flexml_layers[86].compute_node[1][0].out[0], compute_graph.flexml_layers[87].compute_node[1][0].out[0], compute_graph.flexml_layers[88].compute_node[1][0].out[0], compute_graph.flexml_layers[89].compute_node[1][0].out[0], compute_graph.flexml_layers[90].compute_node[1][0].out[0], compute_graph.flexml_layers[91].compute_node[1][0].out[0], compute_graph.flexml_layers[92].compute_node[1][0].out[0], compute_graph.flexml_layers[93].compute_node[1][0].out[0], compute_graph.flexml_layers[94].compute_node[1][0].out[0], compute_graph.flexml_layers[95].compute_node[1][0].out[0], compute_graph.flexml_layers[96].compute_node[1][0].out[0], compute_graph.flexml_layers[97].compute_node[1][0].out[0], compute_graph.flexml_layers[98].compute_node[1][0].out[0], compute_graph.flexml_layers[99].compute_node[1][0].out[0], compute_graph.flexml_layers[100].compute_node[1][0].out[0], compute_graph.flexml_layers[101].compute_node[1][0].out[0], compute_graph.flexml_layers[102].compute_node[1][0].out[0], compute_graph.flexml_layers[103].compute_node[1][0].out[0], compute_graph.flexml_layers[104].compute_node[1][0].out[0], compute_graph.flexml_layers[105].compute_node[1][0].out[0], compute_graph.flexml_layers[106].compute_node[1][0].out[0], compute_graph.flexml_layers[107].compute_node[1][0].out[0], compute_graph.flexml_layers[108].compute_node[1][0].out[0], compute_graph.flexml_layers[109].compute_node[1][0].out[0], compute_graph.flexml_layers[110].compute_node[1][0].out[0], compute_graph.flexml_layers[111].compute_node[1][0].out[0], compute_graph.flexml_layers[112].compute_node[1][0].out[0], compute_graph.flexml_layers[113].compute_node[1][0].out[0], compute_graph.flexml_layers[114].compute_node[1][0].out[0], compute_graph.flexml_layers[115].compute_node[1][0].out[0], compute_graph.flexml_layers[116].compute_node[1][0].out[0], compute_graph.flexml_layers[117].compute_node[1][0].out[0], compute_graph.flexml_layers[118].compute_node[1][0].out[0], compute_graph.flexml_layers[119].compute_node[1][0].out[0], compute_graph.flexml_layers[120].compute_node[1][0].out[0], compute_graph.flexml_layers[121].compute_node[1][0].out[0], compute_graph.flexml_layers[122].compute_node[1][0].out[0], compute_graph.flexml_layers[123].compute_node[1][0].out[0], compute_graph.flexml_layers[124].compute_node[1][0].out[0], compute_graph.flexml_layers[125].compute_node[1][0].out[0], compute_graph.flexml_layers[126].compute_node[1][0].out[0], compute_graph.flexml_layers[127].compute_node[1][0].out[0], compute_graph.flexml_layers[128].compute_node[1][0].out[0], compute_graph.flexml_layers[129].compute_node[1][0].out[0], compute_graph.flexml_layers[130].compute_node[1][0].out[0], compute_graph.flexml_layers[131].compute_node[1][0].out[0], compute_graph.flexml_layers[132].compute_node[1][0].out[0], compute_graph.flexml_layers[133].compute_node[1][0].out[0], compute_graph.flexml_layers[134].compute_node[1][0].out[0], compute_graph.flexml_layers[135].compute_node[1][0].out[0], compute_graph.flexml_layers[136].compute_node[1][0].out[0], compute_graph.flexml_layers[137].compute_node[1][0].out[0], compute_graph.flexml_layers[138].compute_node[1][0].out[0], compute_graph.flexml_layers[139].compute_node[1][0].out[0], compute_graph.flexml_layers[140].compute_node[1][0].out[0], compute_graph.flexml_layers[141].compute_node[1][0].out[0], compute_graph.flexml_layers[142].compute_node[1][0].out[0], compute_graph.flexml_layers[143].compute_node[1][0].out[0], compute_graph.flexml_layers[145].compute_node[1][0].out[0], compute_graph.flexml_layers[146].compute_node[1][0].out[0], compute_graph.flexml_layers[147].compute_node[1][0].out[0], compute_graph.flexml_layers[148].compute_node[1][0].out[0], compute_graph.flexml_layers[149].compute_node[1][0].out[0], compute_graph.flexml_layers[150].compute_node[1][0].out[0], compute_graph.flexml_layers[151].compute_node[1][0].out[0], compute_graph.flexml_layers[152].compute_node[1][0].out[0], compute_graph.flexml_layers[153].compute_node[1][0].out[0], compute_graph.flexml_layers[154].compute_node[1][0].out[0], compute_graph.flexml_layers[155].compute_node[1][0].out[0], compute_graph.flexml_layers[156].compute_node[1][0].out[0], compute_graph.flexml_layers[157].compute_node[1][0].out[0], compute_graph.flexml_layers[158].compute_node[1][0].out[0], compute_graph.flexml_layers[159].compute_node[1][0].out[0], compute_graph.flexml_layers[160].compute_node[1][0].out[0], compute_graph.flexml_layers[161].compute_node[1][0].out[0], compute_graph.flexml_layers[162].compute_node[1][0].out[0], compute_graph.flexml_layers[163].compute_node[1][0].out[0], compute_graph.flexml_layers[164].compute_node[1][0].out[0], compute_graph.flexml_layers[165].compute_node[1][0].out[0], compute_graph.flexml_layers[166].compute_node[1][0].out[0], compute_graph.flexml_layers[167].compute_node[1][0].out[0], compute_graph.flexml_layers[168].compute_node[1][0].out[0], compute_graph.flexml_layers[169].compute_node[1][0].out[0], compute_graph.flexml_layers[170].compute_node[1][0].out[0], compute_graph.flexml_layers[171].compute_node[1][0].out[0], compute_graph.flexml_layers[172].compute_node[1][0].out[0], compute_graph.flexml_layers[173].compute_node[1][0].out[0], compute_graph.flexml_layers[174].compute_node[1][0].out[0], compute_graph.flexml_layers[175].compute_node[1][0].out[0], compute_graph.flexml_layers[176].compute_node[1][0].out[0], compute_graph.flexml_layers[177].compute_node[1][0].out[0], compute_graph.flexml_layers[178].compute_node[1][0].out[0], compute_graph.flexml_layers[179].compute_node[1][0].out[0], compute_graph.flexml_layers[180].compute_node[1][0].out[0], compute_graph.flexml_layers[181].compute_node[1][0].out[0], compute_graph.flexml_layers[182].compute_node[1][0].out[0], compute_graph.flexml_layers[183].compute_node[1][0].out[0], compute_graph.flexml_layers[184].compute_node[1][0].out[0], compute_graph.flexml_layers[185].compute_node[1][0].out[0], compute_graph.flexml_layers[186].compute_node[1][0].out[0], compute_graph.flexml_layers[187].compute_node[1][0].out[0], compute_graph.flexml_layers[188].compute_node[1][0].out[0], compute_graph.flexml_layers[189].compute_node[1][0].out[0], compute_graph.flexml_layers[190].compute_node[1][0].out[0], compute_graph.flexml_layers[191].compute_node[1][0].out[0], compute_graph.flexml_layers[192].compute_node[1][0].out[0], compute_graph.flexml_layers[193].compute_node[1][0].out[0], compute_graph.flexml_layers[194].compute_node[1][0].out[0], compute_graph.flexml_layers[195].compute_node[1][0].out[0], compute_graph.flexml_layers[196].compute_node[1][0].out[0], compute_graph.flexml_layers[197].compute_node[1][0].out[0], compute_graph.flexml_layers[198].compute_node[1][0].out[0], compute_graph.flexml_layers[199].compute_node[1][0].out[0], compute_graph.flexml_layers[200].compute_node[1][0].out[0], compute_graph.flexml_layers[203].compute_node[1][0].out[0], compute_graph.flexml_layers[204].compute_node[1][0].out[0], compute_graph.flexml_layers[205].compute_node[1][0].out[0], compute_graph.flexml_layers[206].compute_node[1][0].out[0], compute_graph.flexml_layers[207].compute_node[1][0].out[0], compute_graph.flexml_layers[208].compute_node[1][0].out[0], compute_graph.flexml_layers[209].compute_node[1][0].out[0], compute_graph.flexml_layers[210].compute_node[1][0].out[0], compute_graph.flexml_layers[211].compute_node[1][0].out[0], compute_graph.flexml_layers[212].compute_node[1][0].out[0], compute_graph.flexml_layers[213].compute_node[1][0].out[0], compute_graph.flexml_layers[214].compute_node[1][0].out[0], compute_graph.flexml_layers[215].compute_node[1][0].out[0], compute_graph.flexml_layers[216].compute_node[1][0].out[0], compute_graph.flexml_layers[217].compute_node[1][0].out[0], compute_graph.flexml_layers[218].compute_node[1][0].out[0], compute_graph.flexml_layers[219].compute_node[1][0].out[0], compute_graph.flexml_layers[220].compute_node[1][0].out[0], compute_graph.flexml_layers[221].compute_node[1][0].out[0], compute_graph.flexml_layers[222].compute_node[1][0].out[0], compute_graph.flexml_layers[223].compute_node[1][0].out[0], compute_graph.flexml_layers[224].compute_node[1][0].out[0], compute_graph.flexml_layers[225].compute_node[1][0].out[0], compute_graph.flexml_layers[226].compute_node[1][0].out[0], compute_graph.flexml_layers[227].compute_node[1][0].out[0], compute_graph.flexml_layers[228].compute_node[1][0].out[0], compute_graph.flexml_layers[229].compute_node[1][0].out[0], compute_graph.flexml_layers[230].compute_node[1][0].out[0], compute_graph.flexml_layers[231].compute_node[1][0].out[0], compute_graph.flexml_layers[232].compute_node[1][0].out[0], compute_graph.flexml_layers[233].compute_node[1][0].out[0], compute_graph.flexml_layers[234].compute_node[1][0].out[0], compute_graph.flexml_layers[235].compute_node[1][0].out[0], compute_graph.flexml_layers[236].compute_node[1][0].out[0]) + +Column 1Row 0Channel 1 + +i8_pi1(compute_graph.flexml_layers[36].compute_node[1][0].in[1], compute_graph.templated_graph_260.mk[1][0].in[1], compute_graph.templated_graph_268.mk[1][0].in[1], compute_graph.templated_graph_273.mk[1][0].in[1], compute_graph.flexml_layers[3].compute_node[1][0].in[1], compute_graph.flexml_layers[8].compute_node[1][0].in[1], compute_graph.flexml_layers[9].compute_node[1][0].in[1], compute_graph.flexml_layers[10].compute_node[1][0].in[1], compute_graph.flexml_layers[11].compute_node[1][0].in[1], compute_graph.flexml_layers[12].compute_node[1][0].in[1], compute_graph.flexml_layers[13].compute_node[1][0].in[1], compute_graph.flexml_layers[14].compute_node[1][0].in[1], compute_graph.flexml_layers[15].compute_node[1][0].in[1], compute_graph.flexml_layers[16].compute_node[1][0].in[1], compute_graph.flexml_layers[17].compute_node[1][0].in[1], compute_graph.flexml_layers[19].compute_node[1][0].in[1], compute_graph.flexml_layers[20].compute_node[1][0].in[1], compute_graph.flexml_layers[25].compute_node[1][0].in[1], compute_graph.flexml_layers[26].compute_node[1][0].in[1], compute_graph.flexml_layers[27].compute_node[1][0].in[1], compute_graph.flexml_layers[29].compute_node[1][0].in[1], compute_graph.flexml_layers[30].compute_node[1][0].in[1], compute_graph.flexml_layers[35].compute_node[1][0].in[1], compute_graph.flexml_layers[144].compute_node[1][0].in[1], compute_graph.flexml_layers[37].compute_node[1][0].in[1], compute_graph.flexml_layers[39].compute_node[1][0].in[1], compute_graph.flexml_layers[40].compute_node[1][0].in[1], compute_graph.flexml_layers[45].compute_node[1][0].in[1], compute_graph.flexml_layers[46].compute_node[1][0].in[1], compute_graph.flexml_layers[51].compute_node[1][0].in[1], compute_graph.flexml_layers[56].compute_node[1][0].in[1], compute_graph.flexml_layers[57].compute_node[1][0].in[1], compute_graph.flexml_layers[62].compute_node[1][0].in[1], compute_graph.flexml_layers[201].compute_node[1][0].in[1], compute_graph.flexml_layers[202].compute_node[1][0].in[1], compute_graph.flexml_layers[67].compute_node[1][0].in[1], compute_graph.flexml_layers[68].compute_node[1][0].in[1], compute_graph.flexml_layers[73].compute_node[1][0].in[1], compute_graph.flexml_layers[78].compute_node[1][0].in[1], compute_graph.flexml_layers[79].compute_node[1][0].in[1], compute_graph.flexml_layers[84].compute_node[1][0].in[1], compute_graph.flexml_layers[89].compute_node[1][0].in[1], compute_graph.flexml_layers[90].compute_node[1][0].in[1], compute_graph.flexml_layers[95].compute_node[1][0].in[1], compute_graph.flexml_layers[101].compute_node[1][0].in[1], compute_graph.flexml_layers[102].compute_node[1][0].in[1], compute_graph.flexml_layers[107].compute_node[1][0].in[1], compute_graph.flexml_layers[108].compute_node[1][0].in[1], compute_graph.flexml_layers[113].compute_node[1][0].in[1], compute_graph.flexml_layers[119].compute_node[1][0].in[1], compute_graph.flexml_layers[120].compute_node[1][0].in[1], compute_graph.flexml_layers[125].compute_node[1][0].in[1], compute_graph.flexml_layers[126].compute_node[1][0].in[1], compute_graph.flexml_layers[131].compute_node[1][0].in[1], compute_graph.flexml_layers[137].compute_node[1][0].in[1], compute_graph.flexml_layers[138].compute_node[1][0].in[1], compute_graph.flexml_layers[143].compute_node[1][0].in[1], compute_graph.flexml_layers[149].compute_node[1][0].in[1], compute_graph.flexml_layers[155].compute_node[1][0].in[1], compute_graph.flexml_layers[156].compute_node[1][0].in[1], compute_graph.flexml_layers[161].compute_node[1][0].in[1], compute_graph.flexml_layers[162].compute_node[1][0].in[1], compute_graph.flexml_layers[167].compute_node[1][0].in[1], compute_graph.flexml_layers[173].compute_node[1][0].in[1], compute_graph.flexml_layers[174].compute_node[1][0].in[1], compute_graph.flexml_layers[179].compute_node[1][0].in[1], compute_graph.flexml_layers[180].compute_node[1][0].in[1], compute_graph.flexml_layers[186].compute_node[1][0].in[1], compute_graph.flexml_layers[188].compute_node[1][0].in[1], compute_graph.flexml_layers[189].compute_node[1][0].in[1], compute_graph.flexml_layers[194].compute_node[1][0].in[1], compute_graph.flexml_layers[207].compute_node[1][0].in[1], compute_graph.flexml_layers[211].compute_node[1][0].in[1], compute_graph.flexml_layers[212].compute_node[1][0].in[1], compute_graph.flexml_layers[217].compute_node[1][0].in[1], compute_graph.flexml_layers[221].compute_node[1][0].in[1], compute_graph.flexml_layers[222].compute_node[1][0].in[1], compute_graph.flexml_layers[227].compute_node[1][0].in[1], compute_graph.flexml_layers[231].compute_node[1][0].in[1], compute_graph.flexml_layers[232].compute_node[1][0].in[1], compute_graph.flexml_layers[233].compute_node[1][0].in[1]) + +Column 1Row 1Channel 0 + +i9_pi0(compute_graph.templated_graph_4.trans_comp_nd[1][1].in[0], compute_graph.flexml_layers[36].compute_node[1][1].in[0], compute_graph.templated_graph_231.mk[1][1].in[0], compute_graph.templated_graph_254.mk[1][1].in[0], compute_graph.templated_graph_258.compute_node[1][1].in[0], compute_graph.templated_graph_259.compute_node[1][1].in[0], compute_graph.templated_graph_260.mk[1][1].in[0], compute_graph.templated_graph_263.compute_node[1][1].in[0], compute_graph.templated_graph_266.compute_node[1][1].in[0], compute_graph.templated_graph_268.mk[1][1].in[0], compute_graph.templated_graph_273.mk[1][1].in[0], compute_graph.templated_graph_274.mk[1][1].in[0], compute_graph.flexml_layers[63].compute_node[1][1].in[0], compute_graph.templated_graph_293.mk[1][1].in[0], compute_graph.templated_graph_298.compute_node[1][1].in[0], compute_graph.templated_graph_300.compute_node[1][1].in[0], compute_graph.flexml_layers[0].compute_node[1][1].in[0], compute_graph.flexml_layers[1].compute_node[1][1].in[0], compute_graph.flexml_layers[1].compute_node[1][1].in[2], compute_graph.flexml_layers[2].compute_node[1][1].in[0], compute_graph.flexml_layers[2].compute_node[1][1].in[2], compute_graph.flexml_layers[3].compute_node[1][1].in[0], compute_graph.flexml_layers[4].compute_node[1][1].in[0], compute_graph.flexml_layers[5].compute_node[1][1].in[0], compute_graph.flexml_layers[6].compute_node[1][1].in[0], compute_graph.flexml_layers[7].compute_node[1][1].in[0], compute_graph.flexml_layers[7].compute_node[1][1].in[2], compute_graph.flexml_layers[8].compute_node[1][1].in[0], compute_graph.flexml_layers[9].compute_node[1][1].in[0], compute_graph.flexml_layers[9].compute_node[1][1].in[2], compute_graph.flexml_layers[10].compute_node[1][1].in[0], compute_graph.flexml_layers[11].compute_node[1][1].in[0], compute_graph.flexml_layers[12].compute_node[1][1].in[0], compute_graph.flexml_layers[13].compute_node[1][1].in[0], compute_graph.flexml_layers[14].compute_node[1][1].in[0], compute_graph.flexml_layers[15].compute_node[1][1].in[0], compute_graph.flexml_layers[15].compute_node[1][1].in[2], compute_graph.flexml_layers[16].compute_node[1][1].in[0], compute_graph.flexml_layers[17].compute_node[1][1].in[0], compute_graph.flexml_layers[18].compute_node[1][1].in[0], compute_graph.flexml_layers[19].compute_node[1][1].in[0], compute_graph.flexml_layers[20].compute_node[1][1].in[0], compute_graph.flexml_layers[21].compute_node[1][1].in[0], compute_graph.flexml_layers[22].compute_node[1][1].in[0], compute_graph.flexml_layers[23].compute_node[1][1].in[0], compute_graph.flexml_layers[24].compute_node[1][1].in[0], compute_graph.flexml_layers[24].compute_node[1][1].in[2], compute_graph.flexml_layers[25].compute_node[1][1].in[0], compute_graph.flexml_layers[26].compute_node[1][1].in[0], compute_graph.flexml_layers[27].compute_node[1][1].in[0], compute_graph.flexml_layers[28].compute_node[1][1].in[0], compute_graph.flexml_layers[29].compute_node[1][1].in[0], compute_graph.flexml_layers[30].compute_node[1][1].in[0], compute_graph.flexml_layers[31].compute_node[1][1].in[0], compute_graph.flexml_layers[32].compute_node[1][1].in[0], compute_graph.flexml_layers[33].compute_node[1][1].in[0], compute_graph.flexml_layers[34].compute_node[1][1].in[0], compute_graph.flexml_layers[34].compute_node[1][1].in[2], compute_graph.flexml_layers[35].compute_node[1][1].in[0], compute_graph.flexml_layers[35].compute_node[1][1].in[2], compute_graph.flexml_layers[144].compute_node[1][1].in[0], compute_graph.flexml_layers[37].compute_node[1][1].in[0], compute_graph.flexml_layers[38].compute_node[1][1].in[0], compute_graph.flexml_layers[39].compute_node[1][1].in[0], compute_graph.flexml_layers[40].compute_node[1][1].in[0], compute_graph.flexml_layers[41].compute_node[1][1].in[0], compute_graph.flexml_layers[42].compute_node[1][1].in[0], compute_graph.flexml_layers[43].compute_node[1][1].in[0], compute_graph.flexml_layers[44].compute_node[1][1].in[0], compute_graph.flexml_layers[44].compute_node[1][1].in[2], compute_graph.flexml_layers[45].compute_node[1][1].in[0], compute_graph.flexml_layers[45].compute_node[1][1].in[2], compute_graph.flexml_layers[46].compute_node[1][1].in[0], compute_graph.flexml_layers[47].compute_node[1][1].in[0], compute_graph.flexml_layers[48].compute_node[1][1].in[0], compute_graph.flexml_layers[49].compute_node[1][1].in[0], compute_graph.flexml_layers[50].compute_node[1][1].in[0], compute_graph.flexml_layers[50].compute_node[1][1].in[2], compute_graph.flexml_layers[51].compute_node[1][1].in[0], compute_graph.flexml_layers[52].compute_node[1][1].in[0], compute_graph.flexml_layers[53].compute_node[1][1].in[0], compute_graph.flexml_layers[54].compute_node[1][1].in[0], compute_graph.flexml_layers[55].compute_node[1][1].in[0], compute_graph.flexml_layers[55].compute_node[1][1].in[2], compute_graph.flexml_layers[56].compute_node[1][1].in[0], compute_graph.flexml_layers[57].compute_node[1][1].in[0], compute_graph.flexml_layers[58].compute_node[1][1].in[0], compute_graph.flexml_layers[59].compute_node[1][1].in[0], compute_graph.flexml_layers[60].compute_node[1][1].in[0], compute_graph.flexml_layers[61].compute_node[1][1].in[0], compute_graph.flexml_layers[61].compute_node[1][1].in[2], compute_graph.flexml_layers[62].compute_node[1][1].in[0], compute_graph.flexml_layers[201].compute_node[1][1].in[0], compute_graph.flexml_layers[202].compute_node[1][1].in[0], compute_graph.flexml_layers[64].compute_node[1][1].in[0], compute_graph.flexml_layers[65].compute_node[1][1].in[0], compute_graph.flexml_layers[66].compute_node[1][1].in[0], compute_graph.flexml_layers[66].compute_node[1][1].in[2], compute_graph.flexml_layers[67].compute_node[1][1].in[0], compute_graph.flexml_layers[67].compute_node[1][1].in[2], compute_graph.flexml_layers[68].compute_node[1][1].in[0], compute_graph.flexml_layers[69].compute_node[1][1].in[0], compute_graph.flexml_layers[70].compute_node[1][1].in[0], compute_graph.flexml_layers[71].compute_node[1][1].in[0], compute_graph.flexml_layers[72].compute_node[1][1].in[0], compute_graph.flexml_layers[72].compute_node[1][1].in[2], compute_graph.flexml_layers[73].compute_node[1][1].in[0], compute_graph.flexml_layers[74].compute_node[1][1].in[0], compute_graph.flexml_layers[75].compute_node[1][1].in[0], compute_graph.flexml_layers[76].compute_node[1][1].in[0], compute_graph.flexml_layers[77].compute_node[1][1].in[0], compute_graph.flexml_layers[77].compute_node[1][1].in[2], compute_graph.flexml_layers[78].compute_node[1][1].in[0], compute_graph.flexml_layers[78].compute_node[1][1].in[2], compute_graph.flexml_layers[79].compute_node[1][1].in[0], compute_graph.flexml_layers[80].compute_node[1][1].in[0], compute_graph.flexml_layers[81].compute_node[1][1].in[0], compute_graph.flexml_layers[82].compute_node[1][1].in[0], compute_graph.flexml_layers[83].compute_node[1][1].in[0], compute_graph.flexml_layers[83].compute_node[1][1].in[2], compute_graph.flexml_layers[84].compute_node[1][1].in[0], compute_graph.flexml_layers[85].compute_node[1][1].in[0], compute_graph.flexml_layers[86].compute_node[1][1].in[0], compute_graph.flexml_layers[87].compute_node[1][1].in[0], compute_graph.flexml_layers[88].compute_node[1][1].in[0], compute_graph.flexml_layers[88].compute_node[1][1].in[2], compute_graph.flexml_layers[89].compute_node[1][1].in[0], compute_graph.flexml_layers[89].compute_node[1][1].in[2], compute_graph.flexml_layers[90].compute_node[1][1].in[0], compute_graph.flexml_layers[91].compute_node[1][1].in[0], compute_graph.flexml_layers[92].compute_node[1][1].in[0], compute_graph.flexml_layers[93].compute_node[1][1].in[0], compute_graph.flexml_layers[94].compute_node[1][1].in[0], compute_graph.flexml_layers[94].compute_node[1][1].in[2], compute_graph.flexml_layers[95].compute_node[1][1].in[0], compute_graph.flexml_layers[96].compute_node[1][1].in[0], compute_graph.flexml_layers[97].compute_node[1][1].in[0], compute_graph.flexml_layers[98].compute_node[1][1].in[0], compute_graph.flexml_layers[99].compute_node[1][1].in[0], compute_graph.flexml_layers[99].compute_node[1][1].in[2], compute_graph.flexml_layers[100].compute_node[1][1].in[0], compute_graph.flexml_layers[101].compute_node[1][1].in[0], compute_graph.flexml_layers[102].compute_node[1][1].in[0], compute_graph.flexml_layers[103].compute_node[1][1].in[0], compute_graph.flexml_layers[104].compute_node[1][1].in[0], compute_graph.flexml_layers[105].compute_node[1][1].in[0], compute_graph.flexml_layers[106].compute_node[1][1].in[0], compute_graph.flexml_layers[106].compute_node[1][1].in[2], compute_graph.flexml_layers[107].compute_node[1][1].in[0], compute_graph.flexml_layers[108].compute_node[1][1].in[0], compute_graph.flexml_layers[109].compute_node[1][1].in[0], compute_graph.flexml_layers[110].compute_node[1][1].in[0], compute_graph.flexml_layers[111].compute_node[1][1].in[0], compute_graph.flexml_layers[112].compute_node[1][1].in[0], compute_graph.flexml_layers[112].compute_node[1][1].in[2], compute_graph.flexml_layers[113].compute_node[1][1].in[0], compute_graph.flexml_layers[114].compute_node[1][1].in[0], compute_graph.flexml_layers[115].compute_node[1][1].in[0], compute_graph.flexml_layers[116].compute_node[1][1].in[0], compute_graph.flexml_layers[117].compute_node[1][1].in[0], compute_graph.flexml_layers[117].compute_node[1][1].in[2], compute_graph.flexml_layers[118].compute_node[1][1].in[0], compute_graph.flexml_layers[119].compute_node[1][1].in[0], compute_graph.flexml_layers[120].compute_node[1][1].in[0], compute_graph.flexml_layers[121].compute_node[1][1].in[0], compute_graph.flexml_layers[122].compute_node[1][1].in[0], compute_graph.flexml_layers[123].compute_node[1][1].in[0], compute_graph.flexml_layers[124].compute_node[1][1].in[0], compute_graph.flexml_layers[124].compute_node[1][1].in[2], compute_graph.flexml_layers[125].compute_node[1][1].in[0], compute_graph.flexml_layers[125].compute_node[1][1].in[2], compute_graph.flexml_layers[126].compute_node[1][1].in[0], compute_graph.flexml_layers[127].compute_node[1][1].in[0], compute_graph.flexml_layers[128].compute_node[1][1].in[0], compute_graph.flexml_layers[129].compute_node[1][1].in[0], compute_graph.flexml_layers[130].compute_node[1][1].in[0], compute_graph.flexml_layers[130].compute_node[1][1].in[2], compute_graph.flexml_layers[131].compute_node[1][1].in[0], compute_graph.flexml_layers[132].compute_node[1][1].in[0], compute_graph.flexml_layers[133].compute_node[1][1].in[0], compute_graph.flexml_layers[134].compute_node[1][1].in[0], compute_graph.flexml_layers[135].compute_node[1][1].in[0], compute_graph.flexml_layers[135].compute_node[1][1].in[2], compute_graph.flexml_layers[136].compute_node[1][1].in[0], compute_graph.flexml_layers[137].compute_node[1][1].in[0], compute_graph.flexml_layers[138].compute_node[1][1].in[0], compute_graph.flexml_layers[139].compute_node[1][1].in[0], compute_graph.flexml_layers[140].compute_node[1][1].in[0], compute_graph.flexml_layers[141].compute_node[1][1].in[0], compute_graph.flexml_layers[142].compute_node[1][1].in[0], compute_graph.flexml_layers[142].compute_node[1][1].in[2], compute_graph.flexml_layers[143].compute_node[1][1].in[0], compute_graph.flexml_layers[145].compute_node[1][1].in[0], compute_graph.flexml_layers[146].compute_node[1][1].in[0], compute_graph.flexml_layers[147].compute_node[1][1].in[0], compute_graph.flexml_layers[148].compute_node[1][1].in[0], compute_graph.flexml_layers[148].compute_node[1][1].in[2], compute_graph.flexml_layers[149].compute_node[1][1].in[0], compute_graph.flexml_layers[150].compute_node[1][1].in[0], compute_graph.flexml_layers[151].compute_node[1][1].in[0], compute_graph.flexml_layers[152].compute_node[1][1].in[0], compute_graph.flexml_layers[153].compute_node[1][1].in[0], compute_graph.flexml_layers[153].compute_node[1][1].in[2], compute_graph.flexml_layers[154].compute_node[1][1].in[0], compute_graph.flexml_layers[155].compute_node[1][1].in[0], compute_graph.flexml_layers[156].compute_node[1][1].in[0], compute_graph.flexml_layers[157].compute_node[1][1].in[0], compute_graph.flexml_layers[158].compute_node[1][1].in[0], compute_graph.flexml_layers[159].compute_node[1][1].in[0], compute_graph.flexml_layers[160].compute_node[1][1].in[0], compute_graph.flexml_layers[160].compute_node[1][1].in[2], compute_graph.flexml_layers[161].compute_node[1][1].in[0], compute_graph.flexml_layers[161].compute_node[1][1].in[2], compute_graph.flexml_layers[162].compute_node[1][1].in[0], compute_graph.flexml_layers[163].compute_node[1][1].in[0], compute_graph.flexml_layers[164].compute_node[1][1].in[0], compute_graph.flexml_layers[165].compute_node[1][1].in[0], compute_graph.flexml_layers[166].compute_node[1][1].in[0], compute_graph.flexml_layers[166].compute_node[1][1].in[2], compute_graph.flexml_layers[167].compute_node[1][1].in[0], compute_graph.flexml_layers[168].compute_node[1][1].in[0], compute_graph.flexml_layers[169].compute_node[1][1].in[0], compute_graph.flexml_layers[170].compute_node[1][1].in[0], compute_graph.flexml_layers[171].compute_node[1][1].in[0], compute_graph.flexml_layers[171].compute_node[1][1].in[2], compute_graph.flexml_layers[172].compute_node[1][1].in[0], compute_graph.flexml_layers[173].compute_node[1][1].in[0], compute_graph.flexml_layers[174].compute_node[1][1].in[0], compute_graph.flexml_layers[175].compute_node[1][1].in[0], compute_graph.flexml_layers[176].compute_node[1][1].in[0], compute_graph.flexml_layers[177].compute_node[1][1].in[0], compute_graph.flexml_layers[178].compute_node[1][1].in[0], compute_graph.flexml_layers[178].compute_node[1][1].in[2], compute_graph.flexml_layers[179].compute_node[1][1].in[0], compute_graph.flexml_layers[179].compute_node[1][1].in[2], compute_graph.flexml_layers[180].compute_node[1][1].in[0], compute_graph.flexml_layers[181].compute_node[1][1].in[0], compute_graph.flexml_layers[182].compute_node[1][1].in[0], compute_graph.flexml_layers[183].compute_node[1][1].in[0], compute_graph.flexml_layers[184].compute_node[1][1].in[0], compute_graph.flexml_layers[184].compute_node[1][1].in[2], compute_graph.flexml_layers[185].compute_node[1][1].in[0], compute_graph.flexml_layers[186].compute_node[1][1].in[0], compute_graph.flexml_layers[187].compute_node[1][1].in[0], compute_graph.flexml_layers[188].compute_node[1][1].in[0], compute_graph.flexml_layers[188].compute_node[1][1].in[2], compute_graph.flexml_layers[189].compute_node[1][1].in[0], compute_graph.flexml_layers[190].compute_node[1][1].in[0], compute_graph.flexml_layers[191].compute_node[1][1].in[0], compute_graph.flexml_layers[191].compute_node[1][1].in[2], compute_graph.flexml_layers[192].compute_node[1][1].in[0], compute_graph.flexml_layers[192].compute_node[1][1].in[2], compute_graph.flexml_layers[193].compute_node[1][1].in[0], compute_graph.flexml_layers[193].compute_node[1][1].in[2], compute_graph.flexml_layers[194].compute_node[1][1].in[0], compute_graph.flexml_layers[195].compute_node[1][1].in[0], compute_graph.flexml_layers[196].compute_node[1][1].in[0], compute_graph.flexml_layers[196].compute_node[1][1].in[2], compute_graph.flexml_layers[197].compute_node[1][1].in[0], compute_graph.flexml_layers[197].compute_node[1][1].in[2], compute_graph.flexml_layers[198].compute_node[1][1].in[0], compute_graph.flexml_layers[199].compute_node[1][1].in[0], compute_graph.flexml_layers[200].compute_node[1][1].in[0], compute_graph.flexml_layers[203].compute_node[1][1].in[0], compute_graph.flexml_layers[204].compute_node[1][1].in[0], compute_graph.flexml_layers[204].compute_node[1][1].in[2], compute_graph.flexml_layers[205].compute_node[1][1].in[0], compute_graph.flexml_layers[205].compute_node[1][1].in[2], compute_graph.flexml_layers[206].compute_node[1][1].in[0], compute_graph.flexml_layers[206].compute_node[1][1].in[2], compute_graph.flexml_layers[207].compute_node[1][1].in[0], compute_graph.flexml_layers[208].compute_node[1][1].in[0], compute_graph.flexml_layers[209].compute_node[1][1].in[0], compute_graph.flexml_layers[209].compute_node[1][1].in[2], compute_graph.flexml_layers[210].compute_node[1][1].in[0], compute_graph.flexml_layers[210].compute_node[1][1].in[2], compute_graph.flexml_layers[211].compute_node[1][1].in[0], compute_graph.flexml_layers[212].compute_node[1][1].in[0], compute_graph.flexml_layers[213].compute_node[1][1].in[0], compute_graph.flexml_layers[214].compute_node[1][1].in[0], compute_graph.flexml_layers[214].compute_node[1][1].in[2], compute_graph.flexml_layers[215].compute_node[1][1].in[0], compute_graph.flexml_layers[215].compute_node[1][1].in[2], compute_graph.flexml_layers[216].compute_node[1][1].in[0], compute_graph.flexml_layers[216].compute_node[1][1].in[2], compute_graph.flexml_layers[217].compute_node[1][1].in[0], compute_graph.flexml_layers[218].compute_node[1][1].in[0], compute_graph.flexml_layers[219].compute_node[1][1].in[0], compute_graph.flexml_layers[219].compute_node[1][1].in[2], compute_graph.flexml_layers[220].compute_node[1][1].in[0], compute_graph.flexml_layers[220].compute_node[1][1].in[2], compute_graph.flexml_layers[221].compute_node[1][1].in[0], compute_graph.flexml_layers[222].compute_node[1][1].in[0], compute_graph.flexml_layers[223].compute_node[1][1].in[0], compute_graph.flexml_layers[224].compute_node[1][1].in[0], compute_graph.flexml_layers[224].compute_node[1][1].in[2], compute_graph.flexml_layers[225].compute_node[1][1].in[0], compute_graph.flexml_layers[225].compute_node[1][1].in[2], compute_graph.flexml_layers[226].compute_node[1][1].in[0], compute_graph.flexml_layers[226].compute_node[1][1].in[2], compute_graph.flexml_layers[227].compute_node[1][1].in[0], compute_graph.flexml_layers[228].compute_node[1][1].in[0], compute_graph.flexml_layers[229].compute_node[1][1].in[0], compute_graph.flexml_layers[229].compute_node[1][1].in[2], compute_graph.flexml_layers[230].compute_node[1][1].in[0], compute_graph.flexml_layers[230].compute_node[1][1].in[2], compute_graph.flexml_layers[231].compute_node[1][1].in[0], compute_graph.flexml_layers[232].compute_node[1][1].in[0], compute_graph.flexml_layers[233].compute_node[1][1].in[0], compute_graph.flexml_layers[234].compute_node[1][1].in[0], compute_graph.flexml_layers[235].compute_node[1][1].in[0], compute_graph.flexml_layers[235].compute_node[1][1].in[2], compute_graph.flexml_layers[236].compute_node[1][1].in[0]) + +Column 1Row 1Channel 0 + +i9_po0(compute_graph.templated_graph_4.trans_comp_nd[1][1].out[0], compute_graph.flexml_layers[36].compute_node[1][1].out[0], compute_graph.templated_graph_231.mk[1][1].out[0], compute_graph.templated_graph_254.mk[1][1].out[0], compute_graph.templated_graph_258.compute_node[1][1].out[0], compute_graph.templated_graph_259.compute_node[1][1].out[0], compute_graph.templated_graph_260.mk[1][1].out[0], compute_graph.templated_graph_263.compute_node[1][1].out[0], compute_graph.templated_graph_266.compute_node[1][1].out[0], compute_graph.templated_graph_268.mk[1][1].out[0], compute_graph.templated_graph_273.mk[1][1].out[0], compute_graph.templated_graph_274.mk[1][1].out[0], compute_graph.flexml_layers[63].compute_node[1][1].out[0], compute_graph.templated_graph_293.mk[1][1].out[0], compute_graph.templated_graph_298.compute_node[1][1].out[0], compute_graph.templated_graph_300.compute_node[1][1].out[0], compute_graph.flexml_layers[0].compute_node[1][1].out[0], compute_graph.flexml_layers[1].compute_node[1][1].out[0], compute_graph.flexml_layers[2].compute_node[1][1].out[0], compute_graph.flexml_layers[3].compute_node[1][1].out[0], compute_graph.flexml_layers[4].compute_node[1][1].out[0], compute_graph.flexml_layers[5].compute_node[1][1].out[0], compute_graph.flexml_layers[6].compute_node[1][1].out[0], compute_graph.flexml_layers[7].compute_node[1][1].out[0], compute_graph.flexml_layers[8].compute_node[1][1].out[0], compute_graph.flexml_layers[9].compute_node[1][1].out[0], compute_graph.flexml_layers[10].compute_node[1][1].out[0], compute_graph.flexml_layers[11].compute_node[1][1].out[0], compute_graph.flexml_layers[12].compute_node[1][1].out[0], compute_graph.flexml_layers[13].compute_node[1][1].out[0], compute_graph.flexml_layers[14].compute_node[1][1].out[0], compute_graph.flexml_layers[15].compute_node[1][1].out[0], compute_graph.flexml_layers[16].compute_node[1][1].out[0], compute_graph.flexml_layers[17].compute_node[1][1].out[0], compute_graph.flexml_layers[18].compute_node[1][1].out[0], compute_graph.flexml_layers[19].compute_node[1][1].out[0], compute_graph.flexml_layers[20].compute_node[1][1].out[0], compute_graph.flexml_layers[21].compute_node[1][1].out[0], compute_graph.flexml_layers[22].compute_node[1][1].out[0], compute_graph.flexml_layers[23].compute_node[1][1].out[0], compute_graph.flexml_layers[24].compute_node[1][1].out[0], compute_graph.flexml_layers[25].compute_node[1][1].out[0], compute_graph.flexml_layers[26].compute_node[1][1].out[0], compute_graph.flexml_layers[27].compute_node[1][1].out[0], compute_graph.flexml_layers[28].compute_node[1][1].out[0], compute_graph.flexml_layers[29].compute_node[1][1].out[0], compute_graph.flexml_layers[30].compute_node[1][1].out[0], compute_graph.flexml_layers[31].compute_node[1][1].out[0], compute_graph.flexml_layers[32].compute_node[1][1].out[0], compute_graph.flexml_layers[33].compute_node[1][1].out[0], compute_graph.flexml_layers[34].compute_node[1][1].out[0], compute_graph.flexml_layers[35].compute_node[1][1].out[0], compute_graph.flexml_layers[144].compute_node[1][1].out[0], compute_graph.flexml_layers[37].compute_node[1][1].out[0], compute_graph.flexml_layers[38].compute_node[1][1].out[0], compute_graph.flexml_layers[39].compute_node[1][1].out[0], compute_graph.flexml_layers[40].compute_node[1][1].out[0], compute_graph.flexml_layers[41].compute_node[1][1].out[0], compute_graph.flexml_layers[42].compute_node[1][1].out[0], compute_graph.flexml_layers[43].compute_node[1][1].out[0], compute_graph.flexml_layers[44].compute_node[1][1].out[0], compute_graph.flexml_layers[45].compute_node[1][1].out[0], compute_graph.flexml_layers[46].compute_node[1][1].out[0], compute_graph.flexml_layers[47].compute_node[1][1].out[0], compute_graph.flexml_layers[48].compute_node[1][1].out[0], compute_graph.flexml_layers[49].compute_node[1][1].out[0], compute_graph.flexml_layers[50].compute_node[1][1].out[0], compute_graph.flexml_layers[51].compute_node[1][1].out[0], compute_graph.flexml_layers[52].compute_node[1][1].out[0], compute_graph.flexml_layers[53].compute_node[1][1].out[0], compute_graph.flexml_layers[54].compute_node[1][1].out[0], compute_graph.flexml_layers[55].compute_node[1][1].out[0], compute_graph.flexml_layers[56].compute_node[1][1].out[0], compute_graph.flexml_layers[57].compute_node[1][1].out[0], compute_graph.flexml_layers[58].compute_node[1][1].out[0], compute_graph.flexml_layers[59].compute_node[1][1].out[0], compute_graph.flexml_layers[60].compute_node[1][1].out[0], compute_graph.flexml_layers[61].compute_node[1][1].out[0], compute_graph.flexml_layers[62].compute_node[1][1].out[0], compute_graph.flexml_layers[201].compute_node[1][1].out[0], compute_graph.flexml_layers[202].compute_node[1][1].out[0], compute_graph.flexml_layers[64].compute_node[1][1].out[0], compute_graph.flexml_layers[65].compute_node[1][1].out[0], compute_graph.flexml_layers[66].compute_node[1][1].out[0], compute_graph.flexml_layers[67].compute_node[1][1].out[0], compute_graph.flexml_layers[68].compute_node[1][1].out[0], compute_graph.flexml_layers[69].compute_node[1][1].out[0], compute_graph.flexml_layers[70].compute_node[1][1].out[0], compute_graph.flexml_layers[71].compute_node[1][1].out[0], compute_graph.flexml_layers[72].compute_node[1][1].out[0], compute_graph.flexml_layers[73].compute_node[1][1].out[0], compute_graph.flexml_layers[74].compute_node[1][1].out[0], compute_graph.flexml_layers[75].compute_node[1][1].out[0], compute_graph.flexml_layers[76].compute_node[1][1].out[0], compute_graph.flexml_layers[77].compute_node[1][1].out[0], compute_graph.flexml_layers[78].compute_node[1][1].out[0], compute_graph.flexml_layers[79].compute_node[1][1].out[0], compute_graph.flexml_layers[80].compute_node[1][1].out[0], compute_graph.flexml_layers[81].compute_node[1][1].out[0], compute_graph.flexml_layers[82].compute_node[1][1].out[0], compute_graph.flexml_layers[83].compute_node[1][1].out[0], compute_graph.flexml_layers[84].compute_node[1][1].out[0], compute_graph.flexml_layers[85].compute_node[1][1].out[0], compute_graph.flexml_layers[86].compute_node[1][1].out[0], compute_graph.flexml_layers[87].compute_node[1][1].out[0], compute_graph.flexml_layers[88].compute_node[1][1].out[0], compute_graph.flexml_layers[89].compute_node[1][1].out[0], compute_graph.flexml_layers[90].compute_node[1][1].out[0], compute_graph.flexml_layers[91].compute_node[1][1].out[0], compute_graph.flexml_layers[92].compute_node[1][1].out[0], compute_graph.flexml_layers[93].compute_node[1][1].out[0], compute_graph.flexml_layers[94].compute_node[1][1].out[0], compute_graph.flexml_layers[95].compute_node[1][1].out[0], compute_graph.flexml_layers[96].compute_node[1][1].out[0], compute_graph.flexml_layers[97].compute_node[1][1].out[0], compute_graph.flexml_layers[98].compute_node[1][1].out[0], compute_graph.flexml_layers[99].compute_node[1][1].out[0], compute_graph.flexml_layers[100].compute_node[1][1].out[0], compute_graph.flexml_layers[101].compute_node[1][1].out[0], compute_graph.flexml_layers[102].compute_node[1][1].out[0], compute_graph.flexml_layers[103].compute_node[1][1].out[0], compute_graph.flexml_layers[104].compute_node[1][1].out[0], compute_graph.flexml_layers[105].compute_node[1][1].out[0], compute_graph.flexml_layers[106].compute_node[1][1].out[0], compute_graph.flexml_layers[107].compute_node[1][1].out[0], compute_graph.flexml_layers[108].compute_node[1][1].out[0], compute_graph.flexml_layers[109].compute_node[1][1].out[0], compute_graph.flexml_layers[110].compute_node[1][1].out[0], compute_graph.flexml_layers[111].compute_node[1][1].out[0], compute_graph.flexml_layers[112].compute_node[1][1].out[0], compute_graph.flexml_layers[113].compute_node[1][1].out[0], compute_graph.flexml_layers[114].compute_node[1][1].out[0], compute_graph.flexml_layers[115].compute_node[1][1].out[0], compute_graph.flexml_layers[116].compute_node[1][1].out[0], compute_graph.flexml_layers[117].compute_node[1][1].out[0], compute_graph.flexml_layers[118].compute_node[1][1].out[0], compute_graph.flexml_layers[119].compute_node[1][1].out[0], compute_graph.flexml_layers[120].compute_node[1][1].out[0], compute_graph.flexml_layers[121].compute_node[1][1].out[0], compute_graph.flexml_layers[122].compute_node[1][1].out[0], compute_graph.flexml_layers[123].compute_node[1][1].out[0], compute_graph.flexml_layers[124].compute_node[1][1].out[0], compute_graph.flexml_layers[125].compute_node[1][1].out[0], compute_graph.flexml_layers[126].compute_node[1][1].out[0], compute_graph.flexml_layers[127].compute_node[1][1].out[0], compute_graph.flexml_layers[128].compute_node[1][1].out[0], compute_graph.flexml_layers[129].compute_node[1][1].out[0], compute_graph.flexml_layers[130].compute_node[1][1].out[0], compute_graph.flexml_layers[131].compute_node[1][1].out[0], compute_graph.flexml_layers[132].compute_node[1][1].out[0], compute_graph.flexml_layers[133].compute_node[1][1].out[0], compute_graph.flexml_layers[134].compute_node[1][1].out[0], compute_graph.flexml_layers[135].compute_node[1][1].out[0], compute_graph.flexml_layers[136].compute_node[1][1].out[0], compute_graph.flexml_layers[137].compute_node[1][1].out[0], compute_graph.flexml_layers[138].compute_node[1][1].out[0], compute_graph.flexml_layers[139].compute_node[1][1].out[0], compute_graph.flexml_layers[140].compute_node[1][1].out[0], compute_graph.flexml_layers[141].compute_node[1][1].out[0], compute_graph.flexml_layers[142].compute_node[1][1].out[0], compute_graph.flexml_layers[143].compute_node[1][1].out[0], compute_graph.flexml_layers[145].compute_node[1][1].out[0], compute_graph.flexml_layers[146].compute_node[1][1].out[0], compute_graph.flexml_layers[147].compute_node[1][1].out[0], compute_graph.flexml_layers[148].compute_node[1][1].out[0], compute_graph.flexml_layers[149].compute_node[1][1].out[0], compute_graph.flexml_layers[150].compute_node[1][1].out[0], compute_graph.flexml_layers[151].compute_node[1][1].out[0], compute_graph.flexml_layers[152].compute_node[1][1].out[0], compute_graph.flexml_layers[153].compute_node[1][1].out[0], compute_graph.flexml_layers[154].compute_node[1][1].out[0], compute_graph.flexml_layers[155].compute_node[1][1].out[0], compute_graph.flexml_layers[156].compute_node[1][1].out[0], compute_graph.flexml_layers[157].compute_node[1][1].out[0], compute_graph.flexml_layers[158].compute_node[1][1].out[0], compute_graph.flexml_layers[159].compute_node[1][1].out[0], compute_graph.flexml_layers[160].compute_node[1][1].out[0], compute_graph.flexml_layers[161].compute_node[1][1].out[0], compute_graph.flexml_layers[162].compute_node[1][1].out[0], compute_graph.flexml_layers[163].compute_node[1][1].out[0], compute_graph.flexml_layers[164].compute_node[1][1].out[0], compute_graph.flexml_layers[165].compute_node[1][1].out[0], compute_graph.flexml_layers[166].compute_node[1][1].out[0], compute_graph.flexml_layers[167].compute_node[1][1].out[0], compute_graph.flexml_layers[168].compute_node[1][1].out[0], compute_graph.flexml_layers[169].compute_node[1][1].out[0], compute_graph.flexml_layers[170].compute_node[1][1].out[0], compute_graph.flexml_layers[171].compute_node[1][1].out[0], compute_graph.flexml_layers[172].compute_node[1][1].out[0], compute_graph.flexml_layers[173].compute_node[1][1].out[0], compute_graph.flexml_layers[174].compute_node[1][1].out[0], compute_graph.flexml_layers[175].compute_node[1][1].out[0], compute_graph.flexml_layers[176].compute_node[1][1].out[0], compute_graph.flexml_layers[177].compute_node[1][1].out[0], compute_graph.flexml_layers[178].compute_node[1][1].out[0], compute_graph.flexml_layers[179].compute_node[1][1].out[0], compute_graph.flexml_layers[180].compute_node[1][1].out[0], compute_graph.flexml_layers[181].compute_node[1][1].out[0], compute_graph.flexml_layers[182].compute_node[1][1].out[0], compute_graph.flexml_layers[183].compute_node[1][1].out[0], compute_graph.flexml_layers[184].compute_node[1][1].out[0], compute_graph.flexml_layers[185].compute_node[1][1].out[0], compute_graph.flexml_layers[186].compute_node[1][1].out[0], compute_graph.flexml_layers[187].compute_node[1][1].out[0], compute_graph.flexml_layers[188].compute_node[1][1].out[0], compute_graph.flexml_layers[189].compute_node[1][1].out[0], compute_graph.flexml_layers[190].compute_node[1][1].out[0], compute_graph.flexml_layers[191].compute_node[1][1].out[0], compute_graph.flexml_layers[192].compute_node[1][1].out[0], compute_graph.flexml_layers[193].compute_node[1][1].out[0], compute_graph.flexml_layers[194].compute_node[1][1].out[0], compute_graph.flexml_layers[195].compute_node[1][1].out[0], compute_graph.flexml_layers[196].compute_node[1][1].out[0], compute_graph.flexml_layers[197].compute_node[1][1].out[0], compute_graph.flexml_layers[198].compute_node[1][1].out[0], compute_graph.flexml_layers[199].compute_node[1][1].out[0], compute_graph.flexml_layers[200].compute_node[1][1].out[0], compute_graph.flexml_layers[203].compute_node[1][1].out[0], compute_graph.flexml_layers[204].compute_node[1][1].out[0], compute_graph.flexml_layers[205].compute_node[1][1].out[0], compute_graph.flexml_layers[206].compute_node[1][1].out[0], compute_graph.flexml_layers[207].compute_node[1][1].out[0], compute_graph.flexml_layers[208].compute_node[1][1].out[0], compute_graph.flexml_layers[209].compute_node[1][1].out[0], compute_graph.flexml_layers[210].compute_node[1][1].out[0], compute_graph.flexml_layers[211].compute_node[1][1].out[0], compute_graph.flexml_layers[212].compute_node[1][1].out[0], compute_graph.flexml_layers[213].compute_node[1][1].out[0], compute_graph.flexml_layers[214].compute_node[1][1].out[0], compute_graph.flexml_layers[215].compute_node[1][1].out[0], compute_graph.flexml_layers[216].compute_node[1][1].out[0], compute_graph.flexml_layers[217].compute_node[1][1].out[0], compute_graph.flexml_layers[218].compute_node[1][1].out[0], compute_graph.flexml_layers[219].compute_node[1][1].out[0], compute_graph.flexml_layers[220].compute_node[1][1].out[0], compute_graph.flexml_layers[221].compute_node[1][1].out[0], compute_graph.flexml_layers[222].compute_node[1][1].out[0], compute_graph.flexml_layers[223].compute_node[1][1].out[0], compute_graph.flexml_layers[224].compute_node[1][1].out[0], compute_graph.flexml_layers[225].compute_node[1][1].out[0], compute_graph.flexml_layers[226].compute_node[1][1].out[0], compute_graph.flexml_layers[227].compute_node[1][1].out[0], compute_graph.flexml_layers[228].compute_node[1][1].out[0], compute_graph.flexml_layers[229].compute_node[1][1].out[0], compute_graph.flexml_layers[230].compute_node[1][1].out[0], compute_graph.flexml_layers[231].compute_node[1][1].out[0], compute_graph.flexml_layers[232].compute_node[1][1].out[0], compute_graph.flexml_layers[233].compute_node[1][1].out[0], compute_graph.flexml_layers[234].compute_node[1][1].out[0], compute_graph.flexml_layers[235].compute_node[1][1].out[0], compute_graph.flexml_layers[236].compute_node[1][1].out[0]) + +Column 1Row 1Channel 1 + +i9_pi1(compute_graph.flexml_layers[36].compute_node[1][1].in[1], compute_graph.templated_graph_260.mk[1][1].in[1], compute_graph.templated_graph_268.mk[1][1].in[1], compute_graph.templated_graph_273.mk[1][1].in[1], compute_graph.flexml_layers[3].compute_node[1][1].in[1], compute_graph.flexml_layers[8].compute_node[1][1].in[1], compute_graph.flexml_layers[9].compute_node[1][1].in[1], compute_graph.flexml_layers[10].compute_node[1][1].in[1], compute_graph.flexml_layers[11].compute_node[1][1].in[1], compute_graph.flexml_layers[12].compute_node[1][1].in[1], compute_graph.flexml_layers[13].compute_node[1][1].in[1], compute_graph.flexml_layers[14].compute_node[1][1].in[1], compute_graph.flexml_layers[15].compute_node[1][1].in[1], compute_graph.flexml_layers[16].compute_node[1][1].in[1], compute_graph.flexml_layers[17].compute_node[1][1].in[1], compute_graph.flexml_layers[19].compute_node[1][1].in[1], compute_graph.flexml_layers[20].compute_node[1][1].in[1], compute_graph.flexml_layers[25].compute_node[1][1].in[1], compute_graph.flexml_layers[26].compute_node[1][1].in[1], compute_graph.flexml_layers[27].compute_node[1][1].in[1], compute_graph.flexml_layers[29].compute_node[1][1].in[1], compute_graph.flexml_layers[30].compute_node[1][1].in[1], compute_graph.flexml_layers[35].compute_node[1][1].in[1], compute_graph.flexml_layers[144].compute_node[1][1].in[1], compute_graph.flexml_layers[37].compute_node[1][1].in[1], compute_graph.flexml_layers[39].compute_node[1][1].in[1], compute_graph.flexml_layers[40].compute_node[1][1].in[1], compute_graph.flexml_layers[45].compute_node[1][1].in[1], compute_graph.flexml_layers[46].compute_node[1][1].in[1], compute_graph.flexml_layers[51].compute_node[1][1].in[1], compute_graph.flexml_layers[56].compute_node[1][1].in[1], compute_graph.flexml_layers[57].compute_node[1][1].in[1], compute_graph.flexml_layers[62].compute_node[1][1].in[1], compute_graph.flexml_layers[201].compute_node[1][1].in[1], compute_graph.flexml_layers[202].compute_node[1][1].in[1], compute_graph.flexml_layers[67].compute_node[1][1].in[1], compute_graph.flexml_layers[68].compute_node[1][1].in[1], compute_graph.flexml_layers[73].compute_node[1][1].in[1], compute_graph.flexml_layers[78].compute_node[1][1].in[1], compute_graph.flexml_layers[79].compute_node[1][1].in[1], compute_graph.flexml_layers[84].compute_node[1][1].in[1], compute_graph.flexml_layers[89].compute_node[1][1].in[1], compute_graph.flexml_layers[90].compute_node[1][1].in[1], compute_graph.flexml_layers[95].compute_node[1][1].in[1], compute_graph.flexml_layers[101].compute_node[1][1].in[1], compute_graph.flexml_layers[102].compute_node[1][1].in[1], compute_graph.flexml_layers[107].compute_node[1][1].in[1], compute_graph.flexml_layers[108].compute_node[1][1].in[1], compute_graph.flexml_layers[113].compute_node[1][1].in[1], compute_graph.flexml_layers[119].compute_node[1][1].in[1], compute_graph.flexml_layers[120].compute_node[1][1].in[1], compute_graph.flexml_layers[125].compute_node[1][1].in[1], compute_graph.flexml_layers[126].compute_node[1][1].in[1], compute_graph.flexml_layers[131].compute_node[1][1].in[1], compute_graph.flexml_layers[137].compute_node[1][1].in[1], compute_graph.flexml_layers[138].compute_node[1][1].in[1], compute_graph.flexml_layers[143].compute_node[1][1].in[1], compute_graph.flexml_layers[149].compute_node[1][1].in[1], compute_graph.flexml_layers[155].compute_node[1][1].in[1], compute_graph.flexml_layers[156].compute_node[1][1].in[1], compute_graph.flexml_layers[161].compute_node[1][1].in[1], compute_graph.flexml_layers[162].compute_node[1][1].in[1], compute_graph.flexml_layers[167].compute_node[1][1].in[1], compute_graph.flexml_layers[173].compute_node[1][1].in[1], compute_graph.flexml_layers[174].compute_node[1][1].in[1], compute_graph.flexml_layers[179].compute_node[1][1].in[1], compute_graph.flexml_layers[180].compute_node[1][1].in[1], compute_graph.flexml_layers[186].compute_node[1][1].in[1], compute_graph.flexml_layers[188].compute_node[1][1].in[1], compute_graph.flexml_layers[189].compute_node[1][1].in[1], compute_graph.flexml_layers[194].compute_node[1][1].in[1], compute_graph.flexml_layers[207].compute_node[1][1].in[1], compute_graph.flexml_layers[211].compute_node[1][1].in[1], compute_graph.flexml_layers[212].compute_node[1][1].in[1], compute_graph.flexml_layers[217].compute_node[1][1].in[1], compute_graph.flexml_layers[221].compute_node[1][1].in[1], compute_graph.flexml_layers[222].compute_node[1][1].in[1], compute_graph.flexml_layers[227].compute_node[1][1].in[1], compute_graph.flexml_layers[231].compute_node[1][1].in[1], compute_graph.flexml_layers[232].compute_node[1][1].in[1], compute_graph.flexml_layers[233].compute_node[1][1].in[1]) + +Column 1Row 2Channel 0 + +i10_pi0(compute_graph.templated_graph_4.trans_comp_nd[1][2].in[0], compute_graph.flexml_layers[36].compute_node[1][2].in[0], compute_graph.templated_graph_231.mk[1][2].in[0], compute_graph.templated_graph_254.mk[1][2].in[0], compute_graph.templated_graph_258.compute_node[1][2].in[0], compute_graph.templated_graph_259.compute_node[1][2].in[0], compute_graph.templated_graph_260.mk[1][2].in[0], compute_graph.templated_graph_263.compute_node[1][2].in[0], compute_graph.templated_graph_266.compute_node[1][2].in[0], compute_graph.templated_graph_268.mk[1][2].in[0], compute_graph.templated_graph_273.mk[1][2].in[0], compute_graph.templated_graph_274.mk[1][2].in[0], compute_graph.flexml_layers[63].compute_node[1][2].in[0], compute_graph.templated_graph_293.mk[1][2].in[0], compute_graph.templated_graph_298.compute_node[1][2].in[0], compute_graph.templated_graph_300.compute_node[1][2].in[0], compute_graph.flexml_layers[0].compute_node[1][2].in[0], compute_graph.flexml_layers[1].compute_node[1][2].in[0], compute_graph.flexml_layers[1].compute_node[1][2].in[2], compute_graph.flexml_layers[2].compute_node[1][2].in[0], compute_graph.flexml_layers[2].compute_node[1][2].in[2], compute_graph.flexml_layers[3].compute_node[1][2].in[0], compute_graph.flexml_layers[4].compute_node[1][2].in[0], compute_graph.flexml_layers[5].compute_node[1][2].in[0], compute_graph.flexml_layers[6].compute_node[1][2].in[0], compute_graph.flexml_layers[7].compute_node[1][2].in[0], compute_graph.flexml_layers[7].compute_node[1][2].in[2], compute_graph.flexml_layers[8].compute_node[1][2].in[0], compute_graph.flexml_layers[9].compute_node[1][2].in[0], compute_graph.flexml_layers[9].compute_node[1][2].in[2], compute_graph.flexml_layers[10].compute_node[1][2].in[0], compute_graph.flexml_layers[11].compute_node[1][2].in[0], compute_graph.flexml_layers[12].compute_node[1][2].in[0], compute_graph.flexml_layers[13].compute_node[1][2].in[0], compute_graph.flexml_layers[14].compute_node[1][2].in[0], compute_graph.flexml_layers[15].compute_node[1][2].in[0], compute_graph.flexml_layers[15].compute_node[1][2].in[2], compute_graph.flexml_layers[16].compute_node[1][2].in[0], compute_graph.flexml_layers[17].compute_node[1][2].in[0], compute_graph.flexml_layers[18].compute_node[1][2].in[0], compute_graph.flexml_layers[19].compute_node[1][2].in[0], compute_graph.flexml_layers[20].compute_node[1][2].in[0], compute_graph.flexml_layers[21].compute_node[1][2].in[0], compute_graph.flexml_layers[22].compute_node[1][2].in[0], compute_graph.flexml_layers[23].compute_node[1][2].in[0], compute_graph.flexml_layers[24].compute_node[1][2].in[0], compute_graph.flexml_layers[24].compute_node[1][2].in[2], compute_graph.flexml_layers[25].compute_node[1][2].in[0], compute_graph.flexml_layers[26].compute_node[1][2].in[0], compute_graph.flexml_layers[27].compute_node[1][2].in[0], compute_graph.flexml_layers[28].compute_node[1][2].in[0], compute_graph.flexml_layers[29].compute_node[1][2].in[0], compute_graph.flexml_layers[30].compute_node[1][2].in[0], compute_graph.flexml_layers[31].compute_node[1][2].in[0], compute_graph.flexml_layers[32].compute_node[1][2].in[0], compute_graph.flexml_layers[33].compute_node[1][2].in[0], compute_graph.flexml_layers[34].compute_node[1][2].in[0], compute_graph.flexml_layers[34].compute_node[1][2].in[2], compute_graph.flexml_layers[35].compute_node[1][2].in[0], compute_graph.flexml_layers[35].compute_node[1][2].in[2], compute_graph.flexml_layers[144].compute_node[1][2].in[0], compute_graph.flexml_layers[37].compute_node[1][2].in[0], compute_graph.flexml_layers[38].compute_node[1][2].in[0], compute_graph.flexml_layers[39].compute_node[1][2].in[0], compute_graph.flexml_layers[40].compute_node[1][2].in[0], compute_graph.flexml_layers[41].compute_node[1][2].in[0], compute_graph.flexml_layers[42].compute_node[1][2].in[0], compute_graph.flexml_layers[43].compute_node[1][2].in[0], compute_graph.flexml_layers[44].compute_node[1][2].in[0], compute_graph.flexml_layers[44].compute_node[1][2].in[2], compute_graph.flexml_layers[45].compute_node[1][2].in[0], compute_graph.flexml_layers[45].compute_node[1][2].in[2], compute_graph.flexml_layers[46].compute_node[1][2].in[0], compute_graph.flexml_layers[47].compute_node[1][2].in[0], compute_graph.flexml_layers[48].compute_node[1][2].in[0], compute_graph.flexml_layers[49].compute_node[1][2].in[0], compute_graph.flexml_layers[50].compute_node[1][2].in[0], compute_graph.flexml_layers[50].compute_node[1][2].in[2], compute_graph.flexml_layers[51].compute_node[1][2].in[0], compute_graph.flexml_layers[52].compute_node[1][2].in[0], compute_graph.flexml_layers[53].compute_node[1][2].in[0], compute_graph.flexml_layers[54].compute_node[1][2].in[0], compute_graph.flexml_layers[55].compute_node[1][2].in[0], compute_graph.flexml_layers[55].compute_node[1][2].in[2], compute_graph.flexml_layers[56].compute_node[1][2].in[0], compute_graph.flexml_layers[57].compute_node[1][2].in[0], compute_graph.flexml_layers[58].compute_node[1][2].in[0], compute_graph.flexml_layers[59].compute_node[1][2].in[0], compute_graph.flexml_layers[60].compute_node[1][2].in[0], compute_graph.flexml_layers[61].compute_node[1][2].in[0], compute_graph.flexml_layers[61].compute_node[1][2].in[2], compute_graph.flexml_layers[62].compute_node[1][2].in[0], compute_graph.flexml_layers[201].compute_node[1][2].in[0], compute_graph.flexml_layers[202].compute_node[1][2].in[0], compute_graph.flexml_layers[64].compute_node[1][2].in[0], compute_graph.flexml_layers[65].compute_node[1][2].in[0], compute_graph.flexml_layers[66].compute_node[1][2].in[0], compute_graph.flexml_layers[66].compute_node[1][2].in[2], compute_graph.flexml_layers[67].compute_node[1][2].in[0], compute_graph.flexml_layers[67].compute_node[1][2].in[2], compute_graph.flexml_layers[68].compute_node[1][2].in[0], compute_graph.flexml_layers[69].compute_node[1][2].in[0], compute_graph.flexml_layers[70].compute_node[1][2].in[0], compute_graph.flexml_layers[71].compute_node[1][2].in[0], compute_graph.flexml_layers[72].compute_node[1][2].in[0], compute_graph.flexml_layers[72].compute_node[1][2].in[2], compute_graph.flexml_layers[73].compute_node[1][2].in[0], compute_graph.flexml_layers[74].compute_node[1][2].in[0], compute_graph.flexml_layers[75].compute_node[1][2].in[0], compute_graph.flexml_layers[76].compute_node[1][2].in[0], compute_graph.flexml_layers[77].compute_node[1][2].in[0], compute_graph.flexml_layers[77].compute_node[1][2].in[2], compute_graph.flexml_layers[78].compute_node[1][2].in[0], compute_graph.flexml_layers[78].compute_node[1][2].in[2], compute_graph.flexml_layers[79].compute_node[1][2].in[0], compute_graph.flexml_layers[80].compute_node[1][2].in[0], compute_graph.flexml_layers[81].compute_node[1][2].in[0], compute_graph.flexml_layers[82].compute_node[1][2].in[0], compute_graph.flexml_layers[83].compute_node[1][2].in[0], compute_graph.flexml_layers[83].compute_node[1][2].in[2], compute_graph.flexml_layers[84].compute_node[1][2].in[0], compute_graph.flexml_layers[85].compute_node[1][2].in[0], compute_graph.flexml_layers[86].compute_node[1][2].in[0], compute_graph.flexml_layers[87].compute_node[1][2].in[0], compute_graph.flexml_layers[88].compute_node[1][2].in[0], compute_graph.flexml_layers[88].compute_node[1][2].in[2], compute_graph.flexml_layers[89].compute_node[1][2].in[0], compute_graph.flexml_layers[89].compute_node[1][2].in[2], compute_graph.flexml_layers[90].compute_node[1][2].in[0], compute_graph.flexml_layers[91].compute_node[1][2].in[0], compute_graph.flexml_layers[92].compute_node[1][2].in[0], compute_graph.flexml_layers[93].compute_node[1][2].in[0], compute_graph.flexml_layers[94].compute_node[1][2].in[0], compute_graph.flexml_layers[94].compute_node[1][2].in[2], compute_graph.flexml_layers[95].compute_node[1][2].in[0], compute_graph.flexml_layers[96].compute_node[1][2].in[0], compute_graph.flexml_layers[97].compute_node[1][2].in[0], compute_graph.flexml_layers[98].compute_node[1][2].in[0], compute_graph.flexml_layers[99].compute_node[1][2].in[0], compute_graph.flexml_layers[99].compute_node[1][2].in[2], compute_graph.flexml_layers[100].compute_node[1][2].in[0], compute_graph.flexml_layers[101].compute_node[1][2].in[0], compute_graph.flexml_layers[102].compute_node[1][2].in[0], compute_graph.flexml_layers[103].compute_node[1][2].in[0], compute_graph.flexml_layers[104].compute_node[1][2].in[0], compute_graph.flexml_layers[105].compute_node[1][2].in[0], compute_graph.flexml_layers[106].compute_node[1][2].in[0], compute_graph.flexml_layers[106].compute_node[1][2].in[2], compute_graph.flexml_layers[107].compute_node[1][2].in[0], compute_graph.flexml_layers[108].compute_node[1][2].in[0], compute_graph.flexml_layers[109].compute_node[1][2].in[0], compute_graph.flexml_layers[110].compute_node[1][2].in[0], compute_graph.flexml_layers[111].compute_node[1][2].in[0], compute_graph.flexml_layers[112].compute_node[1][2].in[0], compute_graph.flexml_layers[112].compute_node[1][2].in[2], compute_graph.flexml_layers[113].compute_node[1][2].in[0], compute_graph.flexml_layers[114].compute_node[1][2].in[0], compute_graph.flexml_layers[115].compute_node[1][2].in[0], compute_graph.flexml_layers[116].compute_node[1][2].in[0], compute_graph.flexml_layers[117].compute_node[1][2].in[0], compute_graph.flexml_layers[117].compute_node[1][2].in[2], compute_graph.flexml_layers[118].compute_node[1][2].in[0], compute_graph.flexml_layers[119].compute_node[1][2].in[0], compute_graph.flexml_layers[120].compute_node[1][2].in[0], compute_graph.flexml_layers[121].compute_node[1][2].in[0], compute_graph.flexml_layers[122].compute_node[1][2].in[0], compute_graph.flexml_layers[123].compute_node[1][2].in[0], compute_graph.flexml_layers[124].compute_node[1][2].in[0], compute_graph.flexml_layers[124].compute_node[1][2].in[2], compute_graph.flexml_layers[125].compute_node[1][2].in[0], compute_graph.flexml_layers[125].compute_node[1][2].in[2], compute_graph.flexml_layers[126].compute_node[1][2].in[0], compute_graph.flexml_layers[127].compute_node[1][2].in[0], compute_graph.flexml_layers[128].compute_node[1][2].in[0], compute_graph.flexml_layers[129].compute_node[1][2].in[0], compute_graph.flexml_layers[130].compute_node[1][2].in[0], compute_graph.flexml_layers[130].compute_node[1][2].in[2], compute_graph.flexml_layers[131].compute_node[1][2].in[0], compute_graph.flexml_layers[132].compute_node[1][2].in[0], compute_graph.flexml_layers[133].compute_node[1][2].in[0], compute_graph.flexml_layers[134].compute_node[1][2].in[0], compute_graph.flexml_layers[135].compute_node[1][2].in[0], compute_graph.flexml_layers[135].compute_node[1][2].in[2], compute_graph.flexml_layers[136].compute_node[1][2].in[0], compute_graph.flexml_layers[137].compute_node[1][2].in[0], compute_graph.flexml_layers[138].compute_node[1][2].in[0], compute_graph.flexml_layers[139].compute_node[1][2].in[0], compute_graph.flexml_layers[140].compute_node[1][2].in[0], compute_graph.flexml_layers[141].compute_node[1][2].in[0], compute_graph.flexml_layers[142].compute_node[1][2].in[0], compute_graph.flexml_layers[142].compute_node[1][2].in[2], compute_graph.flexml_layers[143].compute_node[1][2].in[0], compute_graph.flexml_layers[145].compute_node[1][2].in[0], compute_graph.flexml_layers[146].compute_node[1][2].in[0], compute_graph.flexml_layers[147].compute_node[1][2].in[0], compute_graph.flexml_layers[148].compute_node[1][2].in[0], compute_graph.flexml_layers[148].compute_node[1][2].in[2], compute_graph.flexml_layers[149].compute_node[1][2].in[0], compute_graph.flexml_layers[150].compute_node[1][2].in[0], compute_graph.flexml_layers[151].compute_node[1][2].in[0], compute_graph.flexml_layers[152].compute_node[1][2].in[0], compute_graph.flexml_layers[153].compute_node[1][2].in[0], compute_graph.flexml_layers[153].compute_node[1][2].in[2], compute_graph.flexml_layers[154].compute_node[1][2].in[0], compute_graph.flexml_layers[155].compute_node[1][2].in[0], compute_graph.flexml_layers[156].compute_node[1][2].in[0], compute_graph.flexml_layers[157].compute_node[1][2].in[0], compute_graph.flexml_layers[158].compute_node[1][2].in[0], compute_graph.flexml_layers[159].compute_node[1][2].in[0], compute_graph.flexml_layers[160].compute_node[1][2].in[0], compute_graph.flexml_layers[160].compute_node[1][2].in[2], compute_graph.flexml_layers[161].compute_node[1][2].in[0], compute_graph.flexml_layers[161].compute_node[1][2].in[2], compute_graph.flexml_layers[162].compute_node[1][2].in[0], compute_graph.flexml_layers[163].compute_node[1][2].in[0], compute_graph.flexml_layers[164].compute_node[1][2].in[0], compute_graph.flexml_layers[165].compute_node[1][2].in[0], compute_graph.flexml_layers[166].compute_node[1][2].in[0], compute_graph.flexml_layers[166].compute_node[1][2].in[2], compute_graph.flexml_layers[167].compute_node[1][2].in[0], compute_graph.flexml_layers[168].compute_node[1][2].in[0], compute_graph.flexml_layers[169].compute_node[1][2].in[0], compute_graph.flexml_layers[170].compute_node[1][2].in[0], compute_graph.flexml_layers[171].compute_node[1][2].in[0], compute_graph.flexml_layers[171].compute_node[1][2].in[2], compute_graph.flexml_layers[172].compute_node[1][2].in[0], compute_graph.flexml_layers[173].compute_node[1][2].in[0], compute_graph.flexml_layers[174].compute_node[1][2].in[0], compute_graph.flexml_layers[175].compute_node[1][2].in[0], compute_graph.flexml_layers[176].compute_node[1][2].in[0], compute_graph.flexml_layers[177].compute_node[1][2].in[0], compute_graph.flexml_layers[178].compute_node[1][2].in[0], compute_graph.flexml_layers[178].compute_node[1][2].in[2], compute_graph.flexml_layers[179].compute_node[1][2].in[0], compute_graph.flexml_layers[179].compute_node[1][2].in[2], compute_graph.flexml_layers[180].compute_node[1][2].in[0], compute_graph.flexml_layers[181].compute_node[1][2].in[0], compute_graph.flexml_layers[182].compute_node[1][2].in[0], compute_graph.flexml_layers[183].compute_node[1][2].in[0], compute_graph.flexml_layers[184].compute_node[1][2].in[0], compute_graph.flexml_layers[184].compute_node[1][2].in[2], compute_graph.flexml_layers[185].compute_node[1][2].in[0], compute_graph.flexml_layers[186].compute_node[1][2].in[0], compute_graph.flexml_layers[187].compute_node[1][2].in[0], compute_graph.flexml_layers[188].compute_node[1][2].in[0], compute_graph.flexml_layers[188].compute_node[1][2].in[2], compute_graph.flexml_layers[189].compute_node[1][2].in[0], compute_graph.flexml_layers[190].compute_node[1][2].in[0], compute_graph.flexml_layers[191].compute_node[1][2].in[0], compute_graph.flexml_layers[191].compute_node[1][2].in[2], compute_graph.flexml_layers[192].compute_node[1][2].in[0], compute_graph.flexml_layers[192].compute_node[1][2].in[2], compute_graph.flexml_layers[193].compute_node[1][2].in[0], compute_graph.flexml_layers[193].compute_node[1][2].in[2], compute_graph.flexml_layers[194].compute_node[1][2].in[0], compute_graph.flexml_layers[195].compute_node[1][2].in[0], compute_graph.flexml_layers[196].compute_node[1][2].in[0], compute_graph.flexml_layers[196].compute_node[1][2].in[2], compute_graph.flexml_layers[197].compute_node[1][2].in[0], compute_graph.flexml_layers[197].compute_node[1][2].in[2], compute_graph.flexml_layers[198].compute_node[1][2].in[0], compute_graph.flexml_layers[199].compute_node[1][2].in[0], compute_graph.flexml_layers[200].compute_node[1][2].in[0], compute_graph.flexml_layers[203].compute_node[1][2].in[0], compute_graph.flexml_layers[204].compute_node[1][2].in[0], compute_graph.flexml_layers[204].compute_node[1][2].in[2], compute_graph.flexml_layers[205].compute_node[1][2].in[0], compute_graph.flexml_layers[205].compute_node[1][2].in[2], compute_graph.flexml_layers[206].compute_node[1][2].in[0], compute_graph.flexml_layers[206].compute_node[1][2].in[2], compute_graph.flexml_layers[207].compute_node[1][2].in[0], compute_graph.flexml_layers[208].compute_node[1][2].in[0], compute_graph.flexml_layers[209].compute_node[1][2].in[0], compute_graph.flexml_layers[209].compute_node[1][2].in[2], compute_graph.flexml_layers[210].compute_node[1][2].in[0], compute_graph.flexml_layers[210].compute_node[1][2].in[2], compute_graph.flexml_layers[211].compute_node[1][2].in[0], compute_graph.flexml_layers[212].compute_node[1][2].in[0], compute_graph.flexml_layers[213].compute_node[1][2].in[0], compute_graph.flexml_layers[214].compute_node[1][2].in[0], compute_graph.flexml_layers[214].compute_node[1][2].in[2], compute_graph.flexml_layers[215].compute_node[1][2].in[0], compute_graph.flexml_layers[215].compute_node[1][2].in[2], compute_graph.flexml_layers[216].compute_node[1][2].in[0], compute_graph.flexml_layers[216].compute_node[1][2].in[2], compute_graph.flexml_layers[217].compute_node[1][2].in[0], compute_graph.flexml_layers[218].compute_node[1][2].in[0], compute_graph.flexml_layers[219].compute_node[1][2].in[0], compute_graph.flexml_layers[219].compute_node[1][2].in[2], compute_graph.flexml_layers[220].compute_node[1][2].in[0], compute_graph.flexml_layers[220].compute_node[1][2].in[2], compute_graph.flexml_layers[221].compute_node[1][2].in[0], compute_graph.flexml_layers[222].compute_node[1][2].in[0], compute_graph.flexml_layers[223].compute_node[1][2].in[0], compute_graph.flexml_layers[224].compute_node[1][2].in[0], compute_graph.flexml_layers[224].compute_node[1][2].in[2], compute_graph.flexml_layers[225].compute_node[1][2].in[0], compute_graph.flexml_layers[225].compute_node[1][2].in[2], compute_graph.flexml_layers[226].compute_node[1][2].in[0], compute_graph.flexml_layers[226].compute_node[1][2].in[2], compute_graph.flexml_layers[227].compute_node[1][2].in[0], compute_graph.flexml_layers[228].compute_node[1][2].in[0], compute_graph.flexml_layers[229].compute_node[1][2].in[0], compute_graph.flexml_layers[229].compute_node[1][2].in[2], compute_graph.flexml_layers[230].compute_node[1][2].in[0], compute_graph.flexml_layers[230].compute_node[1][2].in[2], compute_graph.flexml_layers[231].compute_node[1][2].in[0], compute_graph.flexml_layers[232].compute_node[1][2].in[0], compute_graph.flexml_layers[233].compute_node[1][2].in[0], compute_graph.flexml_layers[234].compute_node[1][2].in[0], compute_graph.flexml_layers[235].compute_node[1][2].in[0], compute_graph.flexml_layers[235].compute_node[1][2].in[2], compute_graph.flexml_layers[236].compute_node[1][2].in[0]) + +Column 1Row 2Channel 0 + +i10_po0(compute_graph.templated_graph_4.trans_comp_nd[1][2].out[0], compute_graph.flexml_layers[36].compute_node[1][2].out[0], compute_graph.templated_graph_231.mk[1][2].out[0], compute_graph.templated_graph_254.mk[1][2].out[0], compute_graph.templated_graph_258.compute_node[1][2].out[0], compute_graph.templated_graph_259.compute_node[1][2].out[0], compute_graph.templated_graph_260.mk[1][2].out[0], compute_graph.templated_graph_263.compute_node[1][2].out[0], compute_graph.templated_graph_266.compute_node[1][2].out[0], compute_graph.templated_graph_268.mk[1][2].out[0], compute_graph.templated_graph_273.mk[1][2].out[0], compute_graph.templated_graph_274.mk[1][2].out[0], compute_graph.flexml_layers[63].compute_node[1][2].out[0], compute_graph.templated_graph_293.mk[1][2].out[0], compute_graph.templated_graph_298.compute_node[1][2].out[0], compute_graph.templated_graph_300.compute_node[1][2].out[0], compute_graph.flexml_layers[0].compute_node[1][2].out[0], compute_graph.flexml_layers[1].compute_node[1][2].out[0], compute_graph.flexml_layers[2].compute_node[1][2].out[0], compute_graph.flexml_layers[3].compute_node[1][2].out[0], compute_graph.flexml_layers[4].compute_node[1][2].out[0], compute_graph.flexml_layers[5].compute_node[1][2].out[0], compute_graph.flexml_layers[6].compute_node[1][2].out[0], compute_graph.flexml_layers[7].compute_node[1][2].out[0], compute_graph.flexml_layers[8].compute_node[1][2].out[0], compute_graph.flexml_layers[9].compute_node[1][2].out[0], compute_graph.flexml_layers[10].compute_node[1][2].out[0], compute_graph.flexml_layers[11].compute_node[1][2].out[0], compute_graph.flexml_layers[12].compute_node[1][2].out[0], compute_graph.flexml_layers[13].compute_node[1][2].out[0], compute_graph.flexml_layers[14].compute_node[1][2].out[0], compute_graph.flexml_layers[15].compute_node[1][2].out[0], compute_graph.flexml_layers[16].compute_node[1][2].out[0], compute_graph.flexml_layers[17].compute_node[1][2].out[0], compute_graph.flexml_layers[18].compute_node[1][2].out[0], compute_graph.flexml_layers[19].compute_node[1][2].out[0], compute_graph.flexml_layers[20].compute_node[1][2].out[0], compute_graph.flexml_layers[21].compute_node[1][2].out[0], compute_graph.flexml_layers[22].compute_node[1][2].out[0], compute_graph.flexml_layers[23].compute_node[1][2].out[0], compute_graph.flexml_layers[24].compute_node[1][2].out[0], compute_graph.flexml_layers[25].compute_node[1][2].out[0], compute_graph.flexml_layers[26].compute_node[1][2].out[0], compute_graph.flexml_layers[27].compute_node[1][2].out[0], compute_graph.flexml_layers[28].compute_node[1][2].out[0], compute_graph.flexml_layers[29].compute_node[1][2].out[0], compute_graph.flexml_layers[30].compute_node[1][2].out[0], compute_graph.flexml_layers[31].compute_node[1][2].out[0], compute_graph.flexml_layers[32].compute_node[1][2].out[0], compute_graph.flexml_layers[33].compute_node[1][2].out[0], compute_graph.flexml_layers[34].compute_node[1][2].out[0], compute_graph.flexml_layers[35].compute_node[1][2].out[0], compute_graph.flexml_layers[144].compute_node[1][2].out[0], compute_graph.flexml_layers[37].compute_node[1][2].out[0], compute_graph.flexml_layers[38].compute_node[1][2].out[0], compute_graph.flexml_layers[39].compute_node[1][2].out[0], compute_graph.flexml_layers[40].compute_node[1][2].out[0], compute_graph.flexml_layers[41].compute_node[1][2].out[0], compute_graph.flexml_layers[42].compute_node[1][2].out[0], compute_graph.flexml_layers[43].compute_node[1][2].out[0], compute_graph.flexml_layers[44].compute_node[1][2].out[0], compute_graph.flexml_layers[45].compute_node[1][2].out[0], compute_graph.flexml_layers[46].compute_node[1][2].out[0], compute_graph.flexml_layers[47].compute_node[1][2].out[0], compute_graph.flexml_layers[48].compute_node[1][2].out[0], compute_graph.flexml_layers[49].compute_node[1][2].out[0], compute_graph.flexml_layers[50].compute_node[1][2].out[0], compute_graph.flexml_layers[51].compute_node[1][2].out[0], compute_graph.flexml_layers[52].compute_node[1][2].out[0], compute_graph.flexml_layers[53].compute_node[1][2].out[0], compute_graph.flexml_layers[54].compute_node[1][2].out[0], compute_graph.flexml_layers[55].compute_node[1][2].out[0], compute_graph.flexml_layers[56].compute_node[1][2].out[0], compute_graph.flexml_layers[57].compute_node[1][2].out[0], compute_graph.flexml_layers[58].compute_node[1][2].out[0], compute_graph.flexml_layers[59].compute_node[1][2].out[0], compute_graph.flexml_layers[60].compute_node[1][2].out[0], compute_graph.flexml_layers[61].compute_node[1][2].out[0], compute_graph.flexml_layers[62].compute_node[1][2].out[0], compute_graph.flexml_layers[201].compute_node[1][2].out[0], compute_graph.flexml_layers[202].compute_node[1][2].out[0], compute_graph.flexml_layers[64].compute_node[1][2].out[0], compute_graph.flexml_layers[65].compute_node[1][2].out[0], compute_graph.flexml_layers[66].compute_node[1][2].out[0], compute_graph.flexml_layers[67].compute_node[1][2].out[0], compute_graph.flexml_layers[68].compute_node[1][2].out[0], compute_graph.flexml_layers[69].compute_node[1][2].out[0], compute_graph.flexml_layers[70].compute_node[1][2].out[0], compute_graph.flexml_layers[71].compute_node[1][2].out[0], compute_graph.flexml_layers[72].compute_node[1][2].out[0], compute_graph.flexml_layers[73].compute_node[1][2].out[0], compute_graph.flexml_layers[74].compute_node[1][2].out[0], compute_graph.flexml_layers[75].compute_node[1][2].out[0], compute_graph.flexml_layers[76].compute_node[1][2].out[0], compute_graph.flexml_layers[77].compute_node[1][2].out[0], compute_graph.flexml_layers[78].compute_node[1][2].out[0], compute_graph.flexml_layers[79].compute_node[1][2].out[0], compute_graph.flexml_layers[80].compute_node[1][2].out[0], compute_graph.flexml_layers[81].compute_node[1][2].out[0], compute_graph.flexml_layers[82].compute_node[1][2].out[0], compute_graph.flexml_layers[83].compute_node[1][2].out[0], compute_graph.flexml_layers[84].compute_node[1][2].out[0], compute_graph.flexml_layers[85].compute_node[1][2].out[0], compute_graph.flexml_layers[86].compute_node[1][2].out[0], compute_graph.flexml_layers[87].compute_node[1][2].out[0], compute_graph.flexml_layers[88].compute_node[1][2].out[0], compute_graph.flexml_layers[89].compute_node[1][2].out[0], compute_graph.flexml_layers[90].compute_node[1][2].out[0], compute_graph.flexml_layers[91].compute_node[1][2].out[0], compute_graph.flexml_layers[92].compute_node[1][2].out[0], compute_graph.flexml_layers[93].compute_node[1][2].out[0], compute_graph.flexml_layers[94].compute_node[1][2].out[0], compute_graph.flexml_layers[95].compute_node[1][2].out[0], compute_graph.flexml_layers[96].compute_node[1][2].out[0], compute_graph.flexml_layers[97].compute_node[1][2].out[0], compute_graph.flexml_layers[98].compute_node[1][2].out[0], compute_graph.flexml_layers[99].compute_node[1][2].out[0], compute_graph.flexml_layers[100].compute_node[1][2].out[0], compute_graph.flexml_layers[101].compute_node[1][2].out[0], compute_graph.flexml_layers[102].compute_node[1][2].out[0], compute_graph.flexml_layers[103].compute_node[1][2].out[0], compute_graph.flexml_layers[104].compute_node[1][2].out[0], compute_graph.flexml_layers[105].compute_node[1][2].out[0], compute_graph.flexml_layers[106].compute_node[1][2].out[0], compute_graph.flexml_layers[107].compute_node[1][2].out[0], compute_graph.flexml_layers[108].compute_node[1][2].out[0], compute_graph.flexml_layers[109].compute_node[1][2].out[0], compute_graph.flexml_layers[110].compute_node[1][2].out[0], compute_graph.flexml_layers[111].compute_node[1][2].out[0], compute_graph.flexml_layers[112].compute_node[1][2].out[0], compute_graph.flexml_layers[113].compute_node[1][2].out[0], compute_graph.flexml_layers[114].compute_node[1][2].out[0], compute_graph.flexml_layers[115].compute_node[1][2].out[0], compute_graph.flexml_layers[116].compute_node[1][2].out[0], compute_graph.flexml_layers[117].compute_node[1][2].out[0], compute_graph.flexml_layers[118].compute_node[1][2].out[0], compute_graph.flexml_layers[119].compute_node[1][2].out[0], compute_graph.flexml_layers[120].compute_node[1][2].out[0], compute_graph.flexml_layers[121].compute_node[1][2].out[0], compute_graph.flexml_layers[122].compute_node[1][2].out[0], compute_graph.flexml_layers[123].compute_node[1][2].out[0], compute_graph.flexml_layers[124].compute_node[1][2].out[0], compute_graph.flexml_layers[125].compute_node[1][2].out[0], compute_graph.flexml_layers[126].compute_node[1][2].out[0], compute_graph.flexml_layers[127].compute_node[1][2].out[0], compute_graph.flexml_layers[128].compute_node[1][2].out[0], compute_graph.flexml_layers[129].compute_node[1][2].out[0], compute_graph.flexml_layers[130].compute_node[1][2].out[0], compute_graph.flexml_layers[131].compute_node[1][2].out[0], compute_graph.flexml_layers[132].compute_node[1][2].out[0], compute_graph.flexml_layers[133].compute_node[1][2].out[0], compute_graph.flexml_layers[134].compute_node[1][2].out[0], compute_graph.flexml_layers[135].compute_node[1][2].out[0], compute_graph.flexml_layers[136].compute_node[1][2].out[0], compute_graph.flexml_layers[137].compute_node[1][2].out[0], compute_graph.flexml_layers[138].compute_node[1][2].out[0], compute_graph.flexml_layers[139].compute_node[1][2].out[0], compute_graph.flexml_layers[140].compute_node[1][2].out[0], compute_graph.flexml_layers[141].compute_node[1][2].out[0], compute_graph.flexml_layers[142].compute_node[1][2].out[0], compute_graph.flexml_layers[143].compute_node[1][2].out[0], compute_graph.flexml_layers[145].compute_node[1][2].out[0], compute_graph.flexml_layers[146].compute_node[1][2].out[0], compute_graph.flexml_layers[147].compute_node[1][2].out[0], compute_graph.flexml_layers[148].compute_node[1][2].out[0], compute_graph.flexml_layers[149].compute_node[1][2].out[0], compute_graph.flexml_layers[150].compute_node[1][2].out[0], compute_graph.flexml_layers[151].compute_node[1][2].out[0], compute_graph.flexml_layers[152].compute_node[1][2].out[0], compute_graph.flexml_layers[153].compute_node[1][2].out[0], compute_graph.flexml_layers[154].compute_node[1][2].out[0], compute_graph.flexml_layers[155].compute_node[1][2].out[0], compute_graph.flexml_layers[156].compute_node[1][2].out[0], compute_graph.flexml_layers[157].compute_node[1][2].out[0], compute_graph.flexml_layers[158].compute_node[1][2].out[0], compute_graph.flexml_layers[159].compute_node[1][2].out[0], compute_graph.flexml_layers[160].compute_node[1][2].out[0], compute_graph.flexml_layers[161].compute_node[1][2].out[0], compute_graph.flexml_layers[162].compute_node[1][2].out[0], compute_graph.flexml_layers[163].compute_node[1][2].out[0], compute_graph.flexml_layers[164].compute_node[1][2].out[0], compute_graph.flexml_layers[165].compute_node[1][2].out[0], compute_graph.flexml_layers[166].compute_node[1][2].out[0], compute_graph.flexml_layers[167].compute_node[1][2].out[0], compute_graph.flexml_layers[168].compute_node[1][2].out[0], compute_graph.flexml_layers[169].compute_node[1][2].out[0], compute_graph.flexml_layers[170].compute_node[1][2].out[0], compute_graph.flexml_layers[171].compute_node[1][2].out[0], compute_graph.flexml_layers[172].compute_node[1][2].out[0], compute_graph.flexml_layers[173].compute_node[1][2].out[0], compute_graph.flexml_layers[174].compute_node[1][2].out[0], compute_graph.flexml_layers[175].compute_node[1][2].out[0], compute_graph.flexml_layers[176].compute_node[1][2].out[0], compute_graph.flexml_layers[177].compute_node[1][2].out[0], compute_graph.flexml_layers[178].compute_node[1][2].out[0], compute_graph.flexml_layers[179].compute_node[1][2].out[0], compute_graph.flexml_layers[180].compute_node[1][2].out[0], compute_graph.flexml_layers[181].compute_node[1][2].out[0], compute_graph.flexml_layers[182].compute_node[1][2].out[0], compute_graph.flexml_layers[183].compute_node[1][2].out[0], compute_graph.flexml_layers[184].compute_node[1][2].out[0], compute_graph.flexml_layers[185].compute_node[1][2].out[0], compute_graph.flexml_layers[186].compute_node[1][2].out[0], compute_graph.flexml_layers[187].compute_node[1][2].out[0], compute_graph.flexml_layers[188].compute_node[1][2].out[0], compute_graph.flexml_layers[189].compute_node[1][2].out[0], compute_graph.flexml_layers[190].compute_node[1][2].out[0], compute_graph.flexml_layers[191].compute_node[1][2].out[0], compute_graph.flexml_layers[192].compute_node[1][2].out[0], compute_graph.flexml_layers[193].compute_node[1][2].out[0], compute_graph.flexml_layers[194].compute_node[1][2].out[0], compute_graph.flexml_layers[195].compute_node[1][2].out[0], compute_graph.flexml_layers[196].compute_node[1][2].out[0], compute_graph.flexml_layers[197].compute_node[1][2].out[0], compute_graph.flexml_layers[198].compute_node[1][2].out[0], compute_graph.flexml_layers[199].compute_node[1][2].out[0], compute_graph.flexml_layers[200].compute_node[1][2].out[0], compute_graph.flexml_layers[203].compute_node[1][2].out[0], compute_graph.flexml_layers[204].compute_node[1][2].out[0], compute_graph.flexml_layers[205].compute_node[1][2].out[0], compute_graph.flexml_layers[206].compute_node[1][2].out[0], compute_graph.flexml_layers[207].compute_node[1][2].out[0], compute_graph.flexml_layers[208].compute_node[1][2].out[0], compute_graph.flexml_layers[209].compute_node[1][2].out[0], compute_graph.flexml_layers[210].compute_node[1][2].out[0], compute_graph.flexml_layers[211].compute_node[1][2].out[0], compute_graph.flexml_layers[212].compute_node[1][2].out[0], compute_graph.flexml_layers[213].compute_node[1][2].out[0], compute_graph.flexml_layers[214].compute_node[1][2].out[0], compute_graph.flexml_layers[215].compute_node[1][2].out[0], compute_graph.flexml_layers[216].compute_node[1][2].out[0], compute_graph.flexml_layers[217].compute_node[1][2].out[0], compute_graph.flexml_layers[218].compute_node[1][2].out[0], compute_graph.flexml_layers[219].compute_node[1][2].out[0], compute_graph.flexml_layers[220].compute_node[1][2].out[0], compute_graph.flexml_layers[221].compute_node[1][2].out[0], compute_graph.flexml_layers[222].compute_node[1][2].out[0], compute_graph.flexml_layers[223].compute_node[1][2].out[0], compute_graph.flexml_layers[224].compute_node[1][2].out[0], compute_graph.flexml_layers[225].compute_node[1][2].out[0], compute_graph.flexml_layers[226].compute_node[1][2].out[0], compute_graph.flexml_layers[227].compute_node[1][2].out[0], compute_graph.flexml_layers[228].compute_node[1][2].out[0], compute_graph.flexml_layers[229].compute_node[1][2].out[0], compute_graph.flexml_layers[230].compute_node[1][2].out[0], compute_graph.flexml_layers[231].compute_node[1][2].out[0], compute_graph.flexml_layers[232].compute_node[1][2].out[0], compute_graph.flexml_layers[233].compute_node[1][2].out[0], compute_graph.flexml_layers[234].compute_node[1][2].out[0], compute_graph.flexml_layers[235].compute_node[1][2].out[0], compute_graph.flexml_layers[236].compute_node[1][2].out[0]) + +Column 1Row 2Channel 1 + +i10_pi1(compute_graph.flexml_layers[36].compute_node[1][2].in[1], compute_graph.templated_graph_260.mk[1][2].in[1], compute_graph.templated_graph_268.mk[1][2].in[1], compute_graph.templated_graph_273.mk[1][2].in[1], compute_graph.flexml_layers[3].compute_node[1][2].in[1], compute_graph.flexml_layers[8].compute_node[1][2].in[1], compute_graph.flexml_layers[9].compute_node[1][2].in[1], compute_graph.flexml_layers[10].compute_node[1][2].in[1], compute_graph.flexml_layers[11].compute_node[1][2].in[1], compute_graph.flexml_layers[12].compute_node[1][2].in[1], compute_graph.flexml_layers[13].compute_node[1][2].in[1], compute_graph.flexml_layers[14].compute_node[1][2].in[1], compute_graph.flexml_layers[15].compute_node[1][2].in[1], compute_graph.flexml_layers[16].compute_node[1][2].in[1], compute_graph.flexml_layers[17].compute_node[1][2].in[1], compute_graph.flexml_layers[19].compute_node[1][2].in[1], compute_graph.flexml_layers[20].compute_node[1][2].in[1], compute_graph.flexml_layers[25].compute_node[1][2].in[1], compute_graph.flexml_layers[26].compute_node[1][2].in[1], compute_graph.flexml_layers[27].compute_node[1][2].in[1], compute_graph.flexml_layers[29].compute_node[1][2].in[1], compute_graph.flexml_layers[30].compute_node[1][2].in[1], compute_graph.flexml_layers[35].compute_node[1][2].in[1], compute_graph.flexml_layers[144].compute_node[1][2].in[1], compute_graph.flexml_layers[37].compute_node[1][2].in[1], compute_graph.flexml_layers[39].compute_node[1][2].in[1], compute_graph.flexml_layers[40].compute_node[1][2].in[1], compute_graph.flexml_layers[45].compute_node[1][2].in[1], compute_graph.flexml_layers[46].compute_node[1][2].in[1], compute_graph.flexml_layers[51].compute_node[1][2].in[1], compute_graph.flexml_layers[56].compute_node[1][2].in[1], compute_graph.flexml_layers[57].compute_node[1][2].in[1], compute_graph.flexml_layers[62].compute_node[1][2].in[1], compute_graph.flexml_layers[201].compute_node[1][2].in[1], compute_graph.flexml_layers[202].compute_node[1][2].in[1], compute_graph.flexml_layers[67].compute_node[1][2].in[1], compute_graph.flexml_layers[68].compute_node[1][2].in[1], compute_graph.flexml_layers[73].compute_node[1][2].in[1], compute_graph.flexml_layers[78].compute_node[1][2].in[1], compute_graph.flexml_layers[79].compute_node[1][2].in[1], compute_graph.flexml_layers[84].compute_node[1][2].in[1], compute_graph.flexml_layers[89].compute_node[1][2].in[1], compute_graph.flexml_layers[90].compute_node[1][2].in[1], compute_graph.flexml_layers[95].compute_node[1][2].in[1], compute_graph.flexml_layers[101].compute_node[1][2].in[1], compute_graph.flexml_layers[102].compute_node[1][2].in[1], compute_graph.flexml_layers[107].compute_node[1][2].in[1], compute_graph.flexml_layers[108].compute_node[1][2].in[1], compute_graph.flexml_layers[113].compute_node[1][2].in[1], compute_graph.flexml_layers[119].compute_node[1][2].in[1], compute_graph.flexml_layers[120].compute_node[1][2].in[1], compute_graph.flexml_layers[125].compute_node[1][2].in[1], compute_graph.flexml_layers[126].compute_node[1][2].in[1], compute_graph.flexml_layers[131].compute_node[1][2].in[1], compute_graph.flexml_layers[137].compute_node[1][2].in[1], compute_graph.flexml_layers[138].compute_node[1][2].in[1], compute_graph.flexml_layers[143].compute_node[1][2].in[1], compute_graph.flexml_layers[149].compute_node[1][2].in[1], compute_graph.flexml_layers[155].compute_node[1][2].in[1], compute_graph.flexml_layers[156].compute_node[1][2].in[1], compute_graph.flexml_layers[161].compute_node[1][2].in[1], compute_graph.flexml_layers[162].compute_node[1][2].in[1], compute_graph.flexml_layers[167].compute_node[1][2].in[1], compute_graph.flexml_layers[173].compute_node[1][2].in[1], compute_graph.flexml_layers[174].compute_node[1][2].in[1], compute_graph.flexml_layers[179].compute_node[1][2].in[1], compute_graph.flexml_layers[180].compute_node[1][2].in[1], compute_graph.flexml_layers[186].compute_node[1][2].in[1], compute_graph.flexml_layers[188].compute_node[1][2].in[1], compute_graph.flexml_layers[189].compute_node[1][2].in[1], compute_graph.flexml_layers[194].compute_node[1][2].in[1], compute_graph.flexml_layers[207].compute_node[1][2].in[1], compute_graph.flexml_layers[211].compute_node[1][2].in[1], compute_graph.flexml_layers[212].compute_node[1][2].in[1], compute_graph.flexml_layers[217].compute_node[1][2].in[1], compute_graph.flexml_layers[221].compute_node[1][2].in[1], compute_graph.flexml_layers[222].compute_node[1][2].in[1], compute_graph.flexml_layers[227].compute_node[1][2].in[1], compute_graph.flexml_layers[231].compute_node[1][2].in[1], compute_graph.flexml_layers[232].compute_node[1][2].in[1], compute_graph.flexml_layers[233].compute_node[1][2].in[1]) + +Column 1Row 3Channel 0 + +i11_pi0(compute_graph.templated_graph_4.trans_comp_nd[1][3].in[0], compute_graph.flexml_layers[36].compute_node[1][3].in[0], compute_graph.templated_graph_231.mk[1][3].in[0], compute_graph.templated_graph_254.mk[1][3].in[0], compute_graph.templated_graph_258.compute_node[1][3].in[0], compute_graph.templated_graph_259.compute_node[1][3].in[0], compute_graph.templated_graph_260.mk[1][3].in[0], compute_graph.templated_graph_263.compute_node[1][3].in[0], compute_graph.templated_graph_266.compute_node[1][3].in[0], compute_graph.templated_graph_268.mk[1][3].in[0], compute_graph.templated_graph_273.mk[1][3].in[0], compute_graph.templated_graph_274.mk[1][3].in[0], compute_graph.flexml_layers[63].compute_node[1][3].in[0], compute_graph.templated_graph_293.mk[1][3].in[0], compute_graph.templated_graph_298.compute_node[1][3].in[0], compute_graph.templated_graph_300.compute_node[1][3].in[0], compute_graph.flexml_layers[0].compute_node[1][3].in[0], compute_graph.flexml_layers[1].compute_node[1][3].in[0], compute_graph.flexml_layers[1].compute_node[1][3].in[2], compute_graph.flexml_layers[2].compute_node[1][3].in[0], compute_graph.flexml_layers[2].compute_node[1][3].in[2], compute_graph.flexml_layers[3].compute_node[1][3].in[0], compute_graph.flexml_layers[4].compute_node[1][3].in[0], compute_graph.flexml_layers[5].compute_node[1][3].in[0], compute_graph.flexml_layers[6].compute_node[1][3].in[0], compute_graph.flexml_layers[7].compute_node[1][3].in[0], compute_graph.flexml_layers[7].compute_node[1][3].in[2], compute_graph.flexml_layers[8].compute_node[1][3].in[0], compute_graph.flexml_layers[9].compute_node[1][3].in[0], compute_graph.flexml_layers[9].compute_node[1][3].in[2], compute_graph.flexml_layers[10].compute_node[1][3].in[0], compute_graph.flexml_layers[11].compute_node[1][3].in[0], compute_graph.flexml_layers[12].compute_node[1][3].in[0], compute_graph.flexml_layers[13].compute_node[1][3].in[0], compute_graph.flexml_layers[14].compute_node[1][3].in[0], compute_graph.flexml_layers[15].compute_node[1][3].in[0], compute_graph.flexml_layers[15].compute_node[1][3].in[2], compute_graph.flexml_layers[16].compute_node[1][3].in[0], compute_graph.flexml_layers[17].compute_node[1][3].in[0], compute_graph.flexml_layers[18].compute_node[1][3].in[0], compute_graph.flexml_layers[19].compute_node[1][3].in[0], compute_graph.flexml_layers[20].compute_node[1][3].in[0], compute_graph.flexml_layers[21].compute_node[1][3].in[0], compute_graph.flexml_layers[22].compute_node[1][3].in[0], compute_graph.flexml_layers[23].compute_node[1][3].in[0], compute_graph.flexml_layers[24].compute_node[1][3].in[0], compute_graph.flexml_layers[24].compute_node[1][3].in[2], compute_graph.flexml_layers[25].compute_node[1][3].in[0], compute_graph.flexml_layers[26].compute_node[1][3].in[0], compute_graph.flexml_layers[27].compute_node[1][3].in[0], compute_graph.flexml_layers[28].compute_node[1][3].in[0], compute_graph.flexml_layers[29].compute_node[1][3].in[0], compute_graph.flexml_layers[30].compute_node[1][3].in[0], compute_graph.flexml_layers[31].compute_node[1][3].in[0], compute_graph.flexml_layers[32].compute_node[1][3].in[0], compute_graph.flexml_layers[33].compute_node[1][3].in[0], compute_graph.flexml_layers[34].compute_node[1][3].in[0], compute_graph.flexml_layers[34].compute_node[1][3].in[2], compute_graph.flexml_layers[35].compute_node[1][3].in[0], compute_graph.flexml_layers[35].compute_node[1][3].in[2], compute_graph.flexml_layers[144].compute_node[1][3].in[0], compute_graph.flexml_layers[37].compute_node[1][3].in[0], compute_graph.flexml_layers[38].compute_node[1][3].in[0], compute_graph.flexml_layers[39].compute_node[1][3].in[0], compute_graph.flexml_layers[40].compute_node[1][3].in[0], compute_graph.flexml_layers[41].compute_node[1][3].in[0], compute_graph.flexml_layers[42].compute_node[1][3].in[0], compute_graph.flexml_layers[43].compute_node[1][3].in[0], compute_graph.flexml_layers[44].compute_node[1][3].in[0], compute_graph.flexml_layers[44].compute_node[1][3].in[2], compute_graph.flexml_layers[45].compute_node[1][3].in[0], compute_graph.flexml_layers[45].compute_node[1][3].in[2], compute_graph.flexml_layers[46].compute_node[1][3].in[0], compute_graph.flexml_layers[47].compute_node[1][3].in[0], compute_graph.flexml_layers[48].compute_node[1][3].in[0], compute_graph.flexml_layers[49].compute_node[1][3].in[0], compute_graph.flexml_layers[50].compute_node[1][3].in[0], compute_graph.flexml_layers[50].compute_node[1][3].in[2], compute_graph.flexml_layers[51].compute_node[1][3].in[0], compute_graph.flexml_layers[52].compute_node[1][3].in[0], compute_graph.flexml_layers[53].compute_node[1][3].in[0], compute_graph.flexml_layers[54].compute_node[1][3].in[0], compute_graph.flexml_layers[55].compute_node[1][3].in[0], compute_graph.flexml_layers[55].compute_node[1][3].in[2], compute_graph.flexml_layers[56].compute_node[1][3].in[0], compute_graph.flexml_layers[57].compute_node[1][3].in[0], compute_graph.flexml_layers[58].compute_node[1][3].in[0], compute_graph.flexml_layers[59].compute_node[1][3].in[0], compute_graph.flexml_layers[60].compute_node[1][3].in[0], compute_graph.flexml_layers[61].compute_node[1][3].in[0], compute_graph.flexml_layers[61].compute_node[1][3].in[2], compute_graph.flexml_layers[62].compute_node[1][3].in[0], compute_graph.flexml_layers[201].compute_node[1][3].in[0], compute_graph.flexml_layers[202].compute_node[1][3].in[0], compute_graph.flexml_layers[64].compute_node[1][3].in[0], compute_graph.flexml_layers[65].compute_node[1][3].in[0], compute_graph.flexml_layers[66].compute_node[1][3].in[0], compute_graph.flexml_layers[66].compute_node[1][3].in[2], compute_graph.flexml_layers[67].compute_node[1][3].in[0], compute_graph.flexml_layers[67].compute_node[1][3].in[2], compute_graph.flexml_layers[68].compute_node[1][3].in[0], compute_graph.flexml_layers[69].compute_node[1][3].in[0], compute_graph.flexml_layers[70].compute_node[1][3].in[0], compute_graph.flexml_layers[71].compute_node[1][3].in[0], compute_graph.flexml_layers[72].compute_node[1][3].in[0], compute_graph.flexml_layers[72].compute_node[1][3].in[2], compute_graph.flexml_layers[73].compute_node[1][3].in[0], compute_graph.flexml_layers[74].compute_node[1][3].in[0], compute_graph.flexml_layers[75].compute_node[1][3].in[0], compute_graph.flexml_layers[76].compute_node[1][3].in[0], compute_graph.flexml_layers[77].compute_node[1][3].in[0], compute_graph.flexml_layers[77].compute_node[1][3].in[2], compute_graph.flexml_layers[78].compute_node[1][3].in[0], compute_graph.flexml_layers[78].compute_node[1][3].in[2], compute_graph.flexml_layers[79].compute_node[1][3].in[0], compute_graph.flexml_layers[80].compute_node[1][3].in[0], compute_graph.flexml_layers[81].compute_node[1][3].in[0], compute_graph.flexml_layers[82].compute_node[1][3].in[0], compute_graph.flexml_layers[83].compute_node[1][3].in[0], compute_graph.flexml_layers[83].compute_node[1][3].in[2], compute_graph.flexml_layers[84].compute_node[1][3].in[0], compute_graph.flexml_layers[85].compute_node[1][3].in[0], compute_graph.flexml_layers[86].compute_node[1][3].in[0], compute_graph.flexml_layers[87].compute_node[1][3].in[0], compute_graph.flexml_layers[88].compute_node[1][3].in[0], compute_graph.flexml_layers[88].compute_node[1][3].in[2], compute_graph.flexml_layers[89].compute_node[1][3].in[0], compute_graph.flexml_layers[89].compute_node[1][3].in[2], compute_graph.flexml_layers[90].compute_node[1][3].in[0], compute_graph.flexml_layers[91].compute_node[1][3].in[0], compute_graph.flexml_layers[92].compute_node[1][3].in[0], compute_graph.flexml_layers[93].compute_node[1][3].in[0], compute_graph.flexml_layers[94].compute_node[1][3].in[0], compute_graph.flexml_layers[94].compute_node[1][3].in[2], compute_graph.flexml_layers[95].compute_node[1][3].in[0], compute_graph.flexml_layers[96].compute_node[1][3].in[0], compute_graph.flexml_layers[97].compute_node[1][3].in[0], compute_graph.flexml_layers[98].compute_node[1][3].in[0], compute_graph.flexml_layers[99].compute_node[1][3].in[0], compute_graph.flexml_layers[99].compute_node[1][3].in[2], compute_graph.flexml_layers[100].compute_node[1][3].in[0], compute_graph.flexml_layers[101].compute_node[1][3].in[0], compute_graph.flexml_layers[102].compute_node[1][3].in[0], compute_graph.flexml_layers[103].compute_node[1][3].in[0], compute_graph.flexml_layers[104].compute_node[1][3].in[0], compute_graph.flexml_layers[105].compute_node[1][3].in[0], compute_graph.flexml_layers[106].compute_node[1][3].in[0], compute_graph.flexml_layers[106].compute_node[1][3].in[2], compute_graph.flexml_layers[107].compute_node[1][3].in[0], compute_graph.flexml_layers[108].compute_node[1][3].in[0], compute_graph.flexml_layers[109].compute_node[1][3].in[0], compute_graph.flexml_layers[110].compute_node[1][3].in[0], compute_graph.flexml_layers[111].compute_node[1][3].in[0], compute_graph.flexml_layers[112].compute_node[1][3].in[0], compute_graph.flexml_layers[112].compute_node[1][3].in[2], compute_graph.flexml_layers[113].compute_node[1][3].in[0], compute_graph.flexml_layers[114].compute_node[1][3].in[0], compute_graph.flexml_layers[115].compute_node[1][3].in[0], compute_graph.flexml_layers[116].compute_node[1][3].in[0], compute_graph.flexml_layers[117].compute_node[1][3].in[0], compute_graph.flexml_layers[117].compute_node[1][3].in[2], compute_graph.flexml_layers[118].compute_node[1][3].in[0], compute_graph.flexml_layers[119].compute_node[1][3].in[0], compute_graph.flexml_layers[120].compute_node[1][3].in[0], compute_graph.flexml_layers[121].compute_node[1][3].in[0], compute_graph.flexml_layers[122].compute_node[1][3].in[0], compute_graph.flexml_layers[123].compute_node[1][3].in[0], compute_graph.flexml_layers[124].compute_node[1][3].in[0], compute_graph.flexml_layers[124].compute_node[1][3].in[2], compute_graph.flexml_layers[125].compute_node[1][3].in[0], compute_graph.flexml_layers[125].compute_node[1][3].in[2], compute_graph.flexml_layers[126].compute_node[1][3].in[0], compute_graph.flexml_layers[127].compute_node[1][3].in[0], compute_graph.flexml_layers[128].compute_node[1][3].in[0], compute_graph.flexml_layers[129].compute_node[1][3].in[0], compute_graph.flexml_layers[130].compute_node[1][3].in[0], compute_graph.flexml_layers[130].compute_node[1][3].in[2], compute_graph.flexml_layers[131].compute_node[1][3].in[0], compute_graph.flexml_layers[132].compute_node[1][3].in[0], compute_graph.flexml_layers[133].compute_node[1][3].in[0], compute_graph.flexml_layers[134].compute_node[1][3].in[0], compute_graph.flexml_layers[135].compute_node[1][3].in[0], compute_graph.flexml_layers[135].compute_node[1][3].in[2], compute_graph.flexml_layers[136].compute_node[1][3].in[0], compute_graph.flexml_layers[137].compute_node[1][3].in[0], compute_graph.flexml_layers[138].compute_node[1][3].in[0], compute_graph.flexml_layers[139].compute_node[1][3].in[0], compute_graph.flexml_layers[140].compute_node[1][3].in[0], compute_graph.flexml_layers[141].compute_node[1][3].in[0], compute_graph.flexml_layers[142].compute_node[1][3].in[0], compute_graph.flexml_layers[142].compute_node[1][3].in[2], compute_graph.flexml_layers[143].compute_node[1][3].in[0], compute_graph.flexml_layers[145].compute_node[1][3].in[0], compute_graph.flexml_layers[146].compute_node[1][3].in[0], compute_graph.flexml_layers[147].compute_node[1][3].in[0], compute_graph.flexml_layers[148].compute_node[1][3].in[0], compute_graph.flexml_layers[148].compute_node[1][3].in[2], compute_graph.flexml_layers[149].compute_node[1][3].in[0], compute_graph.flexml_layers[150].compute_node[1][3].in[0], compute_graph.flexml_layers[151].compute_node[1][3].in[0], compute_graph.flexml_layers[152].compute_node[1][3].in[0], compute_graph.flexml_layers[153].compute_node[1][3].in[0], compute_graph.flexml_layers[153].compute_node[1][3].in[2], compute_graph.flexml_layers[154].compute_node[1][3].in[0], compute_graph.flexml_layers[155].compute_node[1][3].in[0], compute_graph.flexml_layers[156].compute_node[1][3].in[0], compute_graph.flexml_layers[157].compute_node[1][3].in[0], compute_graph.flexml_layers[158].compute_node[1][3].in[0], compute_graph.flexml_layers[159].compute_node[1][3].in[0], compute_graph.flexml_layers[160].compute_node[1][3].in[0], compute_graph.flexml_layers[160].compute_node[1][3].in[2], compute_graph.flexml_layers[161].compute_node[1][3].in[0], compute_graph.flexml_layers[161].compute_node[1][3].in[2], compute_graph.flexml_layers[162].compute_node[1][3].in[0], compute_graph.flexml_layers[163].compute_node[1][3].in[0], compute_graph.flexml_layers[164].compute_node[1][3].in[0], compute_graph.flexml_layers[165].compute_node[1][3].in[0], compute_graph.flexml_layers[166].compute_node[1][3].in[0], compute_graph.flexml_layers[166].compute_node[1][3].in[2], compute_graph.flexml_layers[167].compute_node[1][3].in[0], compute_graph.flexml_layers[168].compute_node[1][3].in[0], compute_graph.flexml_layers[169].compute_node[1][3].in[0], compute_graph.flexml_layers[170].compute_node[1][3].in[0], compute_graph.flexml_layers[171].compute_node[1][3].in[0], compute_graph.flexml_layers[171].compute_node[1][3].in[2], compute_graph.flexml_layers[172].compute_node[1][3].in[0], compute_graph.flexml_layers[173].compute_node[1][3].in[0], compute_graph.flexml_layers[174].compute_node[1][3].in[0], compute_graph.flexml_layers[175].compute_node[1][3].in[0], compute_graph.flexml_layers[176].compute_node[1][3].in[0], compute_graph.flexml_layers[177].compute_node[1][3].in[0], compute_graph.flexml_layers[178].compute_node[1][3].in[0], compute_graph.flexml_layers[178].compute_node[1][3].in[2], compute_graph.flexml_layers[179].compute_node[1][3].in[0], compute_graph.flexml_layers[179].compute_node[1][3].in[2], compute_graph.flexml_layers[180].compute_node[1][3].in[0], compute_graph.flexml_layers[181].compute_node[1][3].in[0], compute_graph.flexml_layers[182].compute_node[1][3].in[0], compute_graph.flexml_layers[183].compute_node[1][3].in[0], compute_graph.flexml_layers[184].compute_node[1][3].in[0], compute_graph.flexml_layers[184].compute_node[1][3].in[2], compute_graph.flexml_layers[185].compute_node[1][3].in[0], compute_graph.flexml_layers[186].compute_node[1][3].in[0], compute_graph.flexml_layers[187].compute_node[1][3].in[0], compute_graph.flexml_layers[188].compute_node[1][3].in[0], compute_graph.flexml_layers[188].compute_node[1][3].in[2], compute_graph.flexml_layers[189].compute_node[1][3].in[0], compute_graph.flexml_layers[190].compute_node[1][3].in[0], compute_graph.flexml_layers[191].compute_node[1][3].in[0], compute_graph.flexml_layers[191].compute_node[1][3].in[2], compute_graph.flexml_layers[192].compute_node[1][3].in[0], compute_graph.flexml_layers[192].compute_node[1][3].in[2], compute_graph.flexml_layers[193].compute_node[1][3].in[0], compute_graph.flexml_layers[193].compute_node[1][3].in[2], compute_graph.flexml_layers[194].compute_node[1][3].in[0], compute_graph.flexml_layers[195].compute_node[1][3].in[0], compute_graph.flexml_layers[196].compute_node[1][3].in[0], compute_graph.flexml_layers[196].compute_node[1][3].in[2], compute_graph.flexml_layers[197].compute_node[1][3].in[0], compute_graph.flexml_layers[197].compute_node[1][3].in[2], compute_graph.flexml_layers[198].compute_node[1][3].in[0], compute_graph.flexml_layers[199].compute_node[1][3].in[0], compute_graph.flexml_layers[200].compute_node[1][3].in[0], compute_graph.flexml_layers[203].compute_node[1][3].in[0], compute_graph.flexml_layers[204].compute_node[1][3].in[0], compute_graph.flexml_layers[204].compute_node[1][3].in[2], compute_graph.flexml_layers[205].compute_node[1][3].in[0], compute_graph.flexml_layers[205].compute_node[1][3].in[2], compute_graph.flexml_layers[206].compute_node[1][3].in[0], compute_graph.flexml_layers[206].compute_node[1][3].in[2], compute_graph.flexml_layers[207].compute_node[1][3].in[0], compute_graph.flexml_layers[208].compute_node[1][3].in[0], compute_graph.flexml_layers[209].compute_node[1][3].in[0], compute_graph.flexml_layers[209].compute_node[1][3].in[2], compute_graph.flexml_layers[210].compute_node[1][3].in[0], compute_graph.flexml_layers[210].compute_node[1][3].in[2], compute_graph.flexml_layers[211].compute_node[1][3].in[0], compute_graph.flexml_layers[212].compute_node[1][3].in[0], compute_graph.flexml_layers[213].compute_node[1][3].in[0], compute_graph.flexml_layers[214].compute_node[1][3].in[0], compute_graph.flexml_layers[214].compute_node[1][3].in[2], compute_graph.flexml_layers[215].compute_node[1][3].in[0], compute_graph.flexml_layers[215].compute_node[1][3].in[2], compute_graph.flexml_layers[216].compute_node[1][3].in[0], compute_graph.flexml_layers[216].compute_node[1][3].in[2], compute_graph.flexml_layers[217].compute_node[1][3].in[0], compute_graph.flexml_layers[218].compute_node[1][3].in[0], compute_graph.flexml_layers[219].compute_node[1][3].in[0], compute_graph.flexml_layers[219].compute_node[1][3].in[2], compute_graph.flexml_layers[220].compute_node[1][3].in[0], compute_graph.flexml_layers[220].compute_node[1][3].in[2], compute_graph.flexml_layers[221].compute_node[1][3].in[0], compute_graph.flexml_layers[222].compute_node[1][3].in[0], compute_graph.flexml_layers[223].compute_node[1][3].in[0], compute_graph.flexml_layers[224].compute_node[1][3].in[0], compute_graph.flexml_layers[224].compute_node[1][3].in[2], compute_graph.flexml_layers[225].compute_node[1][3].in[0], compute_graph.flexml_layers[225].compute_node[1][3].in[2], compute_graph.flexml_layers[226].compute_node[1][3].in[0], compute_graph.flexml_layers[226].compute_node[1][3].in[2], compute_graph.flexml_layers[227].compute_node[1][3].in[0], compute_graph.flexml_layers[228].compute_node[1][3].in[0], compute_graph.flexml_layers[229].compute_node[1][3].in[0], compute_graph.flexml_layers[229].compute_node[1][3].in[2], compute_graph.flexml_layers[230].compute_node[1][3].in[0], compute_graph.flexml_layers[230].compute_node[1][3].in[2], compute_graph.flexml_layers[231].compute_node[1][3].in[0], compute_graph.flexml_layers[232].compute_node[1][3].in[0], compute_graph.flexml_layers[233].compute_node[1][3].in[0], compute_graph.flexml_layers[234].compute_node[1][3].in[0], compute_graph.flexml_layers[235].compute_node[1][3].in[0], compute_graph.flexml_layers[235].compute_node[1][3].in[2], compute_graph.flexml_layers[236].compute_node[1][3].in[0]) + +Column 1Row 3Channel 0 + +i11_po0(compute_graph.templated_graph_4.trans_comp_nd[1][3].out[0], compute_graph.flexml_layers[36].compute_node[1][3].out[0], compute_graph.templated_graph_231.mk[1][3].out[0], compute_graph.templated_graph_254.mk[1][3].out[0], compute_graph.templated_graph_258.compute_node[1][3].out[0], compute_graph.templated_graph_259.compute_node[1][3].out[0], compute_graph.templated_graph_260.mk[1][3].out[0], compute_graph.templated_graph_263.compute_node[1][3].out[0], compute_graph.templated_graph_266.compute_node[1][3].out[0], compute_graph.templated_graph_268.mk[1][3].out[0], compute_graph.templated_graph_273.mk[1][3].out[0], compute_graph.templated_graph_274.mk[1][3].out[0], compute_graph.flexml_layers[63].compute_node[1][3].out[0], compute_graph.templated_graph_293.mk[1][3].out[0], compute_graph.templated_graph_298.compute_node[1][3].out[0], compute_graph.templated_graph_300.compute_node[1][3].out[0], compute_graph.flexml_layers[0].compute_node[1][3].out[0], compute_graph.flexml_layers[1].compute_node[1][3].out[0], compute_graph.flexml_layers[2].compute_node[1][3].out[0], compute_graph.flexml_layers[3].compute_node[1][3].out[0], compute_graph.flexml_layers[4].compute_node[1][3].out[0], compute_graph.flexml_layers[5].compute_node[1][3].out[0], compute_graph.flexml_layers[6].compute_node[1][3].out[0], compute_graph.flexml_layers[7].compute_node[1][3].out[0], compute_graph.flexml_layers[8].compute_node[1][3].out[0], compute_graph.flexml_layers[9].compute_node[1][3].out[0], compute_graph.flexml_layers[10].compute_node[1][3].out[0], compute_graph.flexml_layers[11].compute_node[1][3].out[0], compute_graph.flexml_layers[12].compute_node[1][3].out[0], compute_graph.flexml_layers[13].compute_node[1][3].out[0], compute_graph.flexml_layers[14].compute_node[1][3].out[0], compute_graph.flexml_layers[15].compute_node[1][3].out[0], compute_graph.flexml_layers[16].compute_node[1][3].out[0], compute_graph.flexml_layers[17].compute_node[1][3].out[0], compute_graph.flexml_layers[18].compute_node[1][3].out[0], compute_graph.flexml_layers[19].compute_node[1][3].out[0], compute_graph.flexml_layers[20].compute_node[1][3].out[0], compute_graph.flexml_layers[21].compute_node[1][3].out[0], compute_graph.flexml_layers[22].compute_node[1][3].out[0], compute_graph.flexml_layers[23].compute_node[1][3].out[0], compute_graph.flexml_layers[24].compute_node[1][3].out[0], compute_graph.flexml_layers[25].compute_node[1][3].out[0], compute_graph.flexml_layers[26].compute_node[1][3].out[0], compute_graph.flexml_layers[27].compute_node[1][3].out[0], compute_graph.flexml_layers[28].compute_node[1][3].out[0], compute_graph.flexml_layers[29].compute_node[1][3].out[0], compute_graph.flexml_layers[30].compute_node[1][3].out[0], compute_graph.flexml_layers[31].compute_node[1][3].out[0], compute_graph.flexml_layers[32].compute_node[1][3].out[0], compute_graph.flexml_layers[33].compute_node[1][3].out[0], compute_graph.flexml_layers[34].compute_node[1][3].out[0], compute_graph.flexml_layers[35].compute_node[1][3].out[0], compute_graph.flexml_layers[143].compute_node[1][3].out[0], compute_graph.flexml_layers[144].compute_node[1][3].out[0], compute_graph.flexml_layers[37].compute_node[1][3].out[0], compute_graph.flexml_layers[38].compute_node[1][3].out[0], compute_graph.flexml_layers[39].compute_node[1][3].out[0], compute_graph.flexml_layers[40].compute_node[1][3].out[0], compute_graph.flexml_layers[41].compute_node[1][3].out[0], compute_graph.flexml_layers[42].compute_node[1][3].out[0], compute_graph.flexml_layers[43].compute_node[1][3].out[0], compute_graph.flexml_layers[44].compute_node[1][3].out[0], compute_graph.flexml_layers[45].compute_node[1][3].out[0], compute_graph.flexml_layers[46].compute_node[1][3].out[0], compute_graph.flexml_layers[47].compute_node[1][3].out[0], compute_graph.flexml_layers[48].compute_node[1][3].out[0], compute_graph.flexml_layers[49].compute_node[1][3].out[0], compute_graph.flexml_layers[50].compute_node[1][3].out[0], compute_graph.flexml_layers[51].compute_node[1][3].out[0], compute_graph.flexml_layers[52].compute_node[1][3].out[0], compute_graph.flexml_layers[53].compute_node[1][3].out[0], compute_graph.flexml_layers[54].compute_node[1][3].out[0], compute_graph.flexml_layers[55].compute_node[1][3].out[0], compute_graph.flexml_layers[56].compute_node[1][3].out[0], compute_graph.flexml_layers[57].compute_node[1][3].out[0], compute_graph.flexml_layers[58].compute_node[1][3].out[0], compute_graph.flexml_layers[59].compute_node[1][3].out[0], compute_graph.flexml_layers[60].compute_node[1][3].out[0], compute_graph.flexml_layers[61].compute_node[1][3].out[0], compute_graph.flexml_layers[62].compute_node[1][3].out[0], compute_graph.flexml_layers[201].compute_node[1][3].out[0], compute_graph.flexml_layers[202].compute_node[1][3].out[0], compute_graph.flexml_layers[64].compute_node[1][3].out[0], compute_graph.flexml_layers[65].compute_node[1][3].out[0], compute_graph.flexml_layers[66].compute_node[1][3].out[0], compute_graph.flexml_layers[67].compute_node[1][3].out[0], compute_graph.flexml_layers[68].compute_node[1][3].out[0], compute_graph.flexml_layers[69].compute_node[1][3].out[0], compute_graph.flexml_layers[70].compute_node[1][3].out[0], compute_graph.flexml_layers[71].compute_node[1][3].out[0], compute_graph.flexml_layers[72].compute_node[1][3].out[0], compute_graph.flexml_layers[73].compute_node[1][3].out[0], compute_graph.flexml_layers[74].compute_node[1][3].out[0], compute_graph.flexml_layers[75].compute_node[1][3].out[0], compute_graph.flexml_layers[76].compute_node[1][3].out[0], compute_graph.flexml_layers[77].compute_node[1][3].out[0], compute_graph.flexml_layers[78].compute_node[1][3].out[0], compute_graph.flexml_layers[79].compute_node[1][3].out[0], compute_graph.flexml_layers[80].compute_node[1][3].out[0], compute_graph.flexml_layers[81].compute_node[1][3].out[0], compute_graph.flexml_layers[82].compute_node[1][3].out[0], compute_graph.flexml_layers[83].compute_node[1][3].out[0], compute_graph.flexml_layers[84].compute_node[1][3].out[0], compute_graph.flexml_layers[85].compute_node[1][3].out[0], compute_graph.flexml_layers[86].compute_node[1][3].out[0], compute_graph.flexml_layers[87].compute_node[1][3].out[0], compute_graph.flexml_layers[88].compute_node[1][3].out[0], compute_graph.flexml_layers[89].compute_node[1][3].out[0], compute_graph.flexml_layers[90].compute_node[1][3].out[0], compute_graph.flexml_layers[91].compute_node[1][3].out[0], compute_graph.flexml_layers[92].compute_node[1][3].out[0], compute_graph.flexml_layers[93].compute_node[1][3].out[0], compute_graph.flexml_layers[94].compute_node[1][3].out[0], compute_graph.flexml_layers[95].compute_node[1][3].out[0], compute_graph.flexml_layers[96].compute_node[1][3].out[0], compute_graph.flexml_layers[97].compute_node[1][3].out[0], compute_graph.flexml_layers[98].compute_node[1][3].out[0], compute_graph.flexml_layers[99].compute_node[1][3].out[0], compute_graph.flexml_layers[100].compute_node[1][3].out[0], compute_graph.flexml_layers[101].compute_node[1][3].out[0], compute_graph.flexml_layers[102].compute_node[1][3].out[0], compute_graph.flexml_layers[103].compute_node[1][3].out[0], compute_graph.flexml_layers[104].compute_node[1][3].out[0], compute_graph.flexml_layers[105].compute_node[1][3].out[0], compute_graph.flexml_layers[106].compute_node[1][3].out[0], compute_graph.flexml_layers[107].compute_node[1][3].out[0], compute_graph.flexml_layers[108].compute_node[1][3].out[0], compute_graph.flexml_layers[109].compute_node[1][3].out[0], compute_graph.flexml_layers[110].compute_node[1][3].out[0], compute_graph.flexml_layers[111].compute_node[1][3].out[0], compute_graph.flexml_layers[112].compute_node[1][3].out[0], compute_graph.flexml_layers[113].compute_node[1][3].out[0], compute_graph.flexml_layers[114].compute_node[1][3].out[0], compute_graph.flexml_layers[115].compute_node[1][3].out[0], compute_graph.flexml_layers[116].compute_node[1][3].out[0], compute_graph.flexml_layers[117].compute_node[1][3].out[0], compute_graph.flexml_layers[118].compute_node[1][3].out[0], compute_graph.flexml_layers[119].compute_node[1][3].out[0], compute_graph.flexml_layers[120].compute_node[1][3].out[0], compute_graph.flexml_layers[121].compute_node[1][3].out[0], compute_graph.flexml_layers[122].compute_node[1][3].out[0], compute_graph.flexml_layers[123].compute_node[1][3].out[0], compute_graph.flexml_layers[124].compute_node[1][3].out[0], compute_graph.flexml_layers[125].compute_node[1][3].out[0], compute_graph.flexml_layers[126].compute_node[1][3].out[0], compute_graph.flexml_layers[127].compute_node[1][3].out[0], compute_graph.flexml_layers[128].compute_node[1][3].out[0], compute_graph.flexml_layers[129].compute_node[1][3].out[0], compute_graph.flexml_layers[130].compute_node[1][3].out[0], compute_graph.flexml_layers[131].compute_node[1][3].out[0], compute_graph.flexml_layers[132].compute_node[1][3].out[0], compute_graph.flexml_layers[133].compute_node[1][3].out[0], compute_graph.flexml_layers[134].compute_node[1][3].out[0], compute_graph.flexml_layers[135].compute_node[1][3].out[0], compute_graph.flexml_layers[136].compute_node[1][3].out[0], compute_graph.flexml_layers[137].compute_node[1][3].out[0], compute_graph.flexml_layers[138].compute_node[1][3].out[0], compute_graph.flexml_layers[139].compute_node[1][3].out[0], compute_graph.flexml_layers[140].compute_node[1][3].out[0], compute_graph.flexml_layers[141].compute_node[1][3].out[0], compute_graph.flexml_layers[142].compute_node[1][3].out[0], compute_graph.flexml_layers[145].compute_node[1][3].out[0], compute_graph.flexml_layers[146].compute_node[1][3].out[0], compute_graph.flexml_layers[147].compute_node[1][3].out[0], compute_graph.flexml_layers[148].compute_node[1][3].out[0], compute_graph.flexml_layers[149].compute_node[1][3].out[0], compute_graph.flexml_layers[150].compute_node[1][3].out[0], compute_graph.flexml_layers[151].compute_node[1][3].out[0], compute_graph.flexml_layers[152].compute_node[1][3].out[0], compute_graph.flexml_layers[153].compute_node[1][3].out[0], compute_graph.flexml_layers[154].compute_node[1][3].out[0], compute_graph.flexml_layers[155].compute_node[1][3].out[0], compute_graph.flexml_layers[156].compute_node[1][3].out[0], compute_graph.flexml_layers[157].compute_node[1][3].out[0], compute_graph.flexml_layers[158].compute_node[1][3].out[0], compute_graph.flexml_layers[159].compute_node[1][3].out[0], compute_graph.flexml_layers[160].compute_node[1][3].out[0], compute_graph.flexml_layers[161].compute_node[1][3].out[0], compute_graph.flexml_layers[162].compute_node[1][3].out[0], compute_graph.flexml_layers[163].compute_node[1][3].out[0], compute_graph.flexml_layers[164].compute_node[1][3].out[0], compute_graph.flexml_layers[165].compute_node[1][3].out[0], compute_graph.flexml_layers[166].compute_node[1][3].out[0], compute_graph.flexml_layers[167].compute_node[1][3].out[0], compute_graph.flexml_layers[168].compute_node[1][3].out[0], compute_graph.flexml_layers[169].compute_node[1][3].out[0], compute_graph.flexml_layers[170].compute_node[1][3].out[0], compute_graph.flexml_layers[171].compute_node[1][3].out[0], compute_graph.flexml_layers[172].compute_node[1][3].out[0], compute_graph.flexml_layers[173].compute_node[1][3].out[0], compute_graph.flexml_layers[174].compute_node[1][3].out[0], compute_graph.flexml_layers[175].compute_node[1][3].out[0], compute_graph.flexml_layers[176].compute_node[1][3].out[0], compute_graph.flexml_layers[177].compute_node[1][3].out[0], compute_graph.flexml_layers[178].compute_node[1][3].out[0], compute_graph.flexml_layers[179].compute_node[1][3].out[0], compute_graph.flexml_layers[180].compute_node[1][3].out[0], compute_graph.flexml_layers[181].compute_node[1][3].out[0], compute_graph.flexml_layers[182].compute_node[1][3].out[0], compute_graph.flexml_layers[183].compute_node[1][3].out[0], compute_graph.flexml_layers[184].compute_node[1][3].out[0], compute_graph.flexml_layers[185].compute_node[1][3].out[0], compute_graph.flexml_layers[186].compute_node[1][3].out[0], compute_graph.flexml_layers[187].compute_node[1][3].out[0], compute_graph.flexml_layers[188].compute_node[1][3].out[0], compute_graph.flexml_layers[189].compute_node[1][3].out[0], compute_graph.flexml_layers[190].compute_node[1][3].out[0], compute_graph.flexml_layers[191].compute_node[1][3].out[0], compute_graph.flexml_layers[192].compute_node[1][3].out[0], compute_graph.flexml_layers[193].compute_node[1][3].out[0], compute_graph.flexml_layers[194].compute_node[1][3].out[0], compute_graph.flexml_layers[195].compute_node[1][3].out[0], compute_graph.flexml_layers[196].compute_node[1][3].out[0], compute_graph.flexml_layers[197].compute_node[1][3].out[0], compute_graph.flexml_layers[198].compute_node[1][3].out[0], compute_graph.flexml_layers[199].compute_node[1][3].out[0], compute_graph.flexml_layers[200].compute_node[1][3].out[0], compute_graph.flexml_layers[203].compute_node[1][3].out[0], compute_graph.flexml_layers[204].compute_node[1][3].out[0], compute_graph.flexml_layers[205].compute_node[1][3].out[0], compute_graph.flexml_layers[206].compute_node[1][3].out[0], compute_graph.flexml_layers[207].compute_node[1][3].out[0], compute_graph.flexml_layers[208].compute_node[1][3].out[0], compute_graph.flexml_layers[209].compute_node[1][3].out[0], compute_graph.flexml_layers[210].compute_node[1][3].out[0], compute_graph.flexml_layers[211].compute_node[1][3].out[0], compute_graph.flexml_layers[212].compute_node[1][3].out[0], compute_graph.flexml_layers[213].compute_node[1][3].out[0], compute_graph.flexml_layers[214].compute_node[1][3].out[0], compute_graph.flexml_layers[215].compute_node[1][3].out[0], compute_graph.flexml_layers[216].compute_node[1][3].out[0], compute_graph.flexml_layers[217].compute_node[1][3].out[0], compute_graph.flexml_layers[218].compute_node[1][3].out[0], compute_graph.flexml_layers[219].compute_node[1][3].out[0], compute_graph.flexml_layers[220].compute_node[1][3].out[0], compute_graph.flexml_layers[221].compute_node[1][3].out[0], compute_graph.flexml_layers[222].compute_node[1][3].out[0], compute_graph.flexml_layers[223].compute_node[1][3].out[0], compute_graph.flexml_layers[224].compute_node[1][3].out[0], compute_graph.flexml_layers[225].compute_node[1][3].out[0], compute_graph.flexml_layers[226].compute_node[1][3].out[0], compute_graph.flexml_layers[227].compute_node[1][3].out[0], compute_graph.flexml_layers[228].compute_node[1][3].out[0], compute_graph.flexml_layers[229].compute_node[1][3].out[0], compute_graph.flexml_layers[230].compute_node[1][3].out[0], compute_graph.flexml_layers[231].compute_node[1][3].out[0], compute_graph.flexml_layers[232].compute_node[1][3].out[0], compute_graph.flexml_layers[233].compute_node[1][3].out[0], compute_graph.flexml_layers[234].compute_node[1][3].out[0], compute_graph.flexml_layers[235].compute_node[1][3].out[0], compute_graph.flexml_layers[236].compute_node[1][3].out[0]) + +Column 1Row 3Channel 1 + +i11_pi1(compute_graph.flexml_layers[36].compute_node[1][3].in[1], compute_graph.templated_graph_260.mk[1][3].in[1], compute_graph.templated_graph_268.mk[1][3].in[1], compute_graph.templated_graph_273.mk[1][3].in[1], compute_graph.flexml_layers[3].compute_node[1][3].in[1], compute_graph.flexml_layers[8].compute_node[1][3].in[1], compute_graph.flexml_layers[9].compute_node[1][3].in[1], compute_graph.flexml_layers[10].compute_node[1][3].in[1], compute_graph.flexml_layers[11].compute_node[1][3].in[1], compute_graph.flexml_layers[12].compute_node[1][3].in[1], compute_graph.flexml_layers[13].compute_node[1][3].in[1], compute_graph.flexml_layers[14].compute_node[1][3].in[1], compute_graph.flexml_layers[15].compute_node[1][3].in[1], compute_graph.flexml_layers[16].compute_node[1][3].in[1], compute_graph.flexml_layers[17].compute_node[1][3].in[1], compute_graph.flexml_layers[19].compute_node[1][3].in[1], compute_graph.flexml_layers[20].compute_node[1][3].in[1], compute_graph.flexml_layers[25].compute_node[1][3].in[1], compute_graph.flexml_layers[26].compute_node[1][3].in[1], compute_graph.flexml_layers[27].compute_node[1][3].in[1], compute_graph.flexml_layers[29].compute_node[1][3].in[1], compute_graph.flexml_layers[30].compute_node[1][3].in[1], compute_graph.flexml_layers[35].compute_node[1][3].in[1], compute_graph.flexml_layers[143].compute_node[1][3].in[1], compute_graph.flexml_layers[144].compute_node[1][3].in[1], compute_graph.flexml_layers[37].compute_node[1][3].in[1], compute_graph.flexml_layers[39].compute_node[1][3].in[1], compute_graph.flexml_layers[40].compute_node[1][3].in[1], compute_graph.flexml_layers[45].compute_node[1][3].in[1], compute_graph.flexml_layers[46].compute_node[1][3].in[1], compute_graph.flexml_layers[51].compute_node[1][3].in[1], compute_graph.flexml_layers[56].compute_node[1][3].in[1], compute_graph.flexml_layers[57].compute_node[1][3].in[1], compute_graph.flexml_layers[62].compute_node[1][3].in[1], compute_graph.flexml_layers[201].compute_node[1][3].in[1], compute_graph.flexml_layers[202].compute_node[1][3].in[1], compute_graph.flexml_layers[67].compute_node[1][3].in[1], compute_graph.flexml_layers[68].compute_node[1][3].in[1], compute_graph.flexml_layers[73].compute_node[1][3].in[1], compute_graph.flexml_layers[78].compute_node[1][3].in[1], compute_graph.flexml_layers[79].compute_node[1][3].in[1], compute_graph.flexml_layers[84].compute_node[1][3].in[1], compute_graph.flexml_layers[89].compute_node[1][3].in[1], compute_graph.flexml_layers[90].compute_node[1][3].in[1], compute_graph.flexml_layers[95].compute_node[1][3].in[1], compute_graph.flexml_layers[101].compute_node[1][3].in[1], compute_graph.flexml_layers[102].compute_node[1][3].in[1], compute_graph.flexml_layers[107].compute_node[1][3].in[1], compute_graph.flexml_layers[108].compute_node[1][3].in[1], compute_graph.flexml_layers[113].compute_node[1][3].in[1], compute_graph.flexml_layers[119].compute_node[1][3].in[1], compute_graph.flexml_layers[120].compute_node[1][3].in[1], compute_graph.flexml_layers[125].compute_node[1][3].in[1], compute_graph.flexml_layers[126].compute_node[1][3].in[1], compute_graph.flexml_layers[131].compute_node[1][3].in[1], compute_graph.flexml_layers[137].compute_node[1][3].in[1], compute_graph.flexml_layers[138].compute_node[1][3].in[1], compute_graph.flexml_layers[149].compute_node[1][3].in[1], compute_graph.flexml_layers[155].compute_node[1][3].in[1], compute_graph.flexml_layers[156].compute_node[1][3].in[1], compute_graph.flexml_layers[161].compute_node[1][3].in[1], compute_graph.flexml_layers[162].compute_node[1][3].in[1], compute_graph.flexml_layers[167].compute_node[1][3].in[1], compute_graph.flexml_layers[173].compute_node[1][3].in[1], compute_graph.flexml_layers[174].compute_node[1][3].in[1], compute_graph.flexml_layers[179].compute_node[1][3].in[1], compute_graph.flexml_layers[180].compute_node[1][3].in[1], compute_graph.flexml_layers[186].compute_node[1][3].in[1], compute_graph.flexml_layers[188].compute_node[1][3].in[1], compute_graph.flexml_layers[189].compute_node[1][3].in[1], compute_graph.flexml_layers[194].compute_node[1][3].in[1], compute_graph.flexml_layers[207].compute_node[1][3].in[1], compute_graph.flexml_layers[211].compute_node[1][3].in[1], compute_graph.flexml_layers[212].compute_node[1][3].in[1], compute_graph.flexml_layers[217].compute_node[1][3].in[1], compute_graph.flexml_layers[221].compute_node[1][3].in[1], compute_graph.flexml_layers[222].compute_node[1][3].in[1], compute_graph.flexml_layers[227].compute_node[1][3].in[1], compute_graph.flexml_layers[231].compute_node[1][3].in[1], compute_graph.flexml_layers[232].compute_node[1][3].in[1], compute_graph.flexml_layers[233].compute_node[1][3].in[1]) + +Column 2Row 0Channel 0 + +i12_pi0(compute_graph.templated_graph_4.trans_comp_nd[2][0].in[0], compute_graph.flexml_layers[36].compute_node[2][0].in[0], compute_graph.templated_graph_231.mk[2][0].in[0], compute_graph.templated_graph_254.mk[2][0].in[0], compute_graph.templated_graph_258.compute_node[2][0].in[0], compute_graph.templated_graph_259.compute_node[2][0].in[0], compute_graph.templated_graph_260.mk[2][0].in[0], compute_graph.templated_graph_263.compute_node[2][0].in[0], compute_graph.templated_graph_266.compute_node[2][0].in[0], compute_graph.templated_graph_268.mk[2][0].in[0], compute_graph.templated_graph_273.mk[2][0].in[0], compute_graph.templated_graph_274.mk[2][0].in[0], compute_graph.flexml_layers[63].compute_node[2][0].in[0], compute_graph.templated_graph_293.mk[2][0].in[0], compute_graph.templated_graph_298.compute_node[2][0].in[0], compute_graph.templated_graph_300.compute_node[2][0].in[0], compute_graph.flexml_layers[0].compute_node[2][0].in[0], compute_graph.flexml_layers[1].compute_node[2][0].in[0], compute_graph.flexml_layers[1].compute_node[2][0].in[2], compute_graph.flexml_layers[2].compute_node[2][0].in[0], compute_graph.flexml_layers[2].compute_node[2][0].in[2], compute_graph.flexml_layers[3].compute_node[2][0].in[0], compute_graph.flexml_layers[4].compute_node[2][0].in[0], compute_graph.flexml_layers[5].compute_node[2][0].in[0], compute_graph.flexml_layers[6].compute_node[2][0].in[0], compute_graph.flexml_layers[7].compute_node[2][0].in[0], compute_graph.flexml_layers[7].compute_node[2][0].in[2], compute_graph.flexml_layers[8].compute_node[2][0].in[0], compute_graph.flexml_layers[9].compute_node[2][0].in[0], compute_graph.flexml_layers[9].compute_node[2][0].in[2], compute_graph.flexml_layers[10].compute_node[2][0].in[0], compute_graph.flexml_layers[11].compute_node[2][0].in[0], compute_graph.flexml_layers[12].compute_node[2][0].in[0], compute_graph.flexml_layers[13].compute_node[2][0].in[0], compute_graph.flexml_layers[14].compute_node[2][0].in[0], compute_graph.flexml_layers[15].compute_node[2][0].in[0], compute_graph.flexml_layers[15].compute_node[2][0].in[2], compute_graph.flexml_layers[16].compute_node[2][0].in[0], compute_graph.flexml_layers[17].compute_node[2][0].in[0], compute_graph.flexml_layers[18].compute_node[2][0].in[0], compute_graph.flexml_layers[19].compute_node[2][0].in[0], compute_graph.flexml_layers[20].compute_node[2][0].in[0], compute_graph.flexml_layers[21].compute_node[2][0].in[0], compute_graph.flexml_layers[22].compute_node[2][0].in[0], compute_graph.flexml_layers[23].compute_node[2][0].in[0], compute_graph.flexml_layers[24].compute_node[2][0].in[0], compute_graph.flexml_layers[24].compute_node[2][0].in[2], compute_graph.flexml_layers[25].compute_node[2][0].in[0], compute_graph.flexml_layers[26].compute_node[2][0].in[0], compute_graph.flexml_layers[27].compute_node[2][0].in[0], compute_graph.flexml_layers[28].compute_node[2][0].in[0], compute_graph.flexml_layers[29].compute_node[2][0].in[0], compute_graph.flexml_layers[30].compute_node[2][0].in[0], compute_graph.flexml_layers[31].compute_node[2][0].in[0], compute_graph.flexml_layers[32].compute_node[2][0].in[0], compute_graph.flexml_layers[33].compute_node[2][0].in[0], compute_graph.flexml_layers[34].compute_node[2][0].in[0], compute_graph.flexml_layers[34].compute_node[2][0].in[2], compute_graph.flexml_layers[35].compute_node[2][0].in[0], compute_graph.flexml_layers[35].compute_node[2][0].in[2], compute_graph.flexml_layers[143].compute_node[2][0].in[0], compute_graph.flexml_layers[144].compute_node[2][0].in[0], compute_graph.flexml_layers[37].compute_node[2][0].in[0], compute_graph.flexml_layers[38].compute_node[2][0].in[0], compute_graph.flexml_layers[39].compute_node[2][0].in[0], compute_graph.flexml_layers[40].compute_node[2][0].in[0], compute_graph.flexml_layers[41].compute_node[2][0].in[0], compute_graph.flexml_layers[42].compute_node[2][0].in[0], compute_graph.flexml_layers[43].compute_node[2][0].in[0], compute_graph.flexml_layers[44].compute_node[2][0].in[0], compute_graph.flexml_layers[44].compute_node[2][0].in[2], compute_graph.flexml_layers[45].compute_node[2][0].in[0], compute_graph.flexml_layers[45].compute_node[2][0].in[2], compute_graph.flexml_layers[46].compute_node[2][0].in[0], compute_graph.flexml_layers[47].compute_node[2][0].in[0], compute_graph.flexml_layers[48].compute_node[2][0].in[0], compute_graph.flexml_layers[49].compute_node[2][0].in[0], compute_graph.flexml_layers[50].compute_node[2][0].in[0], compute_graph.flexml_layers[50].compute_node[2][0].in[2], compute_graph.flexml_layers[51].compute_node[2][0].in[0], compute_graph.flexml_layers[52].compute_node[2][0].in[0], compute_graph.flexml_layers[53].compute_node[2][0].in[0], compute_graph.flexml_layers[54].compute_node[2][0].in[0], compute_graph.flexml_layers[55].compute_node[2][0].in[0], compute_graph.flexml_layers[55].compute_node[2][0].in[2], compute_graph.flexml_layers[56].compute_node[2][0].in[0], compute_graph.flexml_layers[57].compute_node[2][0].in[0], compute_graph.flexml_layers[58].compute_node[2][0].in[0], compute_graph.flexml_layers[59].compute_node[2][0].in[0], compute_graph.flexml_layers[60].compute_node[2][0].in[0], compute_graph.flexml_layers[61].compute_node[2][0].in[0], compute_graph.flexml_layers[61].compute_node[2][0].in[2], compute_graph.flexml_layers[62].compute_node[2][0].in[0], compute_graph.flexml_layers[200].compute_node[2][0].in[0], compute_graph.flexml_layers[201].compute_node[2][0].in[0], compute_graph.flexml_layers[202].compute_node[2][0].in[0], compute_graph.flexml_layers[64].compute_node[2][0].in[0], compute_graph.flexml_layers[65].compute_node[2][0].in[0], compute_graph.flexml_layers[66].compute_node[2][0].in[0], compute_graph.flexml_layers[66].compute_node[2][0].in[2], compute_graph.flexml_layers[67].compute_node[2][0].in[0], compute_graph.flexml_layers[67].compute_node[2][0].in[2], compute_graph.flexml_layers[68].compute_node[2][0].in[0], compute_graph.flexml_layers[69].compute_node[2][0].in[0], compute_graph.flexml_layers[70].compute_node[2][0].in[0], compute_graph.flexml_layers[71].compute_node[2][0].in[0], compute_graph.flexml_layers[72].compute_node[2][0].in[0], compute_graph.flexml_layers[72].compute_node[2][0].in[2], compute_graph.flexml_layers[73].compute_node[2][0].in[0], compute_graph.flexml_layers[74].compute_node[2][0].in[0], compute_graph.flexml_layers[75].compute_node[2][0].in[0], compute_graph.flexml_layers[76].compute_node[2][0].in[0], compute_graph.flexml_layers[77].compute_node[2][0].in[0], compute_graph.flexml_layers[77].compute_node[2][0].in[2], compute_graph.flexml_layers[78].compute_node[2][0].in[0], compute_graph.flexml_layers[78].compute_node[2][0].in[2], compute_graph.flexml_layers[79].compute_node[2][0].in[0], compute_graph.flexml_layers[80].compute_node[2][0].in[0], compute_graph.flexml_layers[81].compute_node[2][0].in[0], compute_graph.flexml_layers[82].compute_node[2][0].in[0], compute_graph.flexml_layers[83].compute_node[2][0].in[0], compute_graph.flexml_layers[83].compute_node[2][0].in[2], compute_graph.flexml_layers[84].compute_node[2][0].in[0], compute_graph.flexml_layers[85].compute_node[2][0].in[0], compute_graph.flexml_layers[86].compute_node[2][0].in[0], compute_graph.flexml_layers[87].compute_node[2][0].in[0], compute_graph.flexml_layers[88].compute_node[2][0].in[0], compute_graph.flexml_layers[88].compute_node[2][0].in[2], compute_graph.flexml_layers[89].compute_node[2][0].in[0], compute_graph.flexml_layers[89].compute_node[2][0].in[2], compute_graph.flexml_layers[90].compute_node[2][0].in[0], compute_graph.flexml_layers[91].compute_node[2][0].in[0], compute_graph.flexml_layers[92].compute_node[2][0].in[0], compute_graph.flexml_layers[93].compute_node[2][0].in[0], compute_graph.flexml_layers[94].compute_node[2][0].in[0], compute_graph.flexml_layers[94].compute_node[2][0].in[2], compute_graph.flexml_layers[95].compute_node[2][0].in[0], compute_graph.flexml_layers[96].compute_node[2][0].in[0], compute_graph.flexml_layers[97].compute_node[2][0].in[0], compute_graph.flexml_layers[98].compute_node[2][0].in[0], compute_graph.flexml_layers[99].compute_node[2][0].in[0], compute_graph.flexml_layers[99].compute_node[2][0].in[2], compute_graph.flexml_layers[100].compute_node[2][0].in[0], compute_graph.flexml_layers[101].compute_node[2][0].in[0], compute_graph.flexml_layers[102].compute_node[2][0].in[0], compute_graph.flexml_layers[103].compute_node[2][0].in[0], compute_graph.flexml_layers[104].compute_node[2][0].in[0], compute_graph.flexml_layers[105].compute_node[2][0].in[0], compute_graph.flexml_layers[106].compute_node[2][0].in[0], compute_graph.flexml_layers[106].compute_node[2][0].in[2], compute_graph.flexml_layers[107].compute_node[2][0].in[0], compute_graph.flexml_layers[108].compute_node[2][0].in[0], compute_graph.flexml_layers[109].compute_node[2][0].in[0], compute_graph.flexml_layers[110].compute_node[2][0].in[0], compute_graph.flexml_layers[111].compute_node[2][0].in[0], compute_graph.flexml_layers[112].compute_node[2][0].in[0], compute_graph.flexml_layers[112].compute_node[2][0].in[2], compute_graph.flexml_layers[113].compute_node[2][0].in[0], compute_graph.flexml_layers[114].compute_node[2][0].in[0], compute_graph.flexml_layers[115].compute_node[2][0].in[0], compute_graph.flexml_layers[116].compute_node[2][0].in[0], compute_graph.flexml_layers[117].compute_node[2][0].in[0], compute_graph.flexml_layers[117].compute_node[2][0].in[2], compute_graph.flexml_layers[118].compute_node[2][0].in[0], compute_graph.flexml_layers[119].compute_node[2][0].in[0], compute_graph.flexml_layers[120].compute_node[2][0].in[0], compute_graph.flexml_layers[121].compute_node[2][0].in[0], compute_graph.flexml_layers[122].compute_node[2][0].in[0], compute_graph.flexml_layers[123].compute_node[2][0].in[0], compute_graph.flexml_layers[124].compute_node[2][0].in[0], compute_graph.flexml_layers[124].compute_node[2][0].in[2], compute_graph.flexml_layers[125].compute_node[2][0].in[0], compute_graph.flexml_layers[125].compute_node[2][0].in[2], compute_graph.flexml_layers[126].compute_node[2][0].in[0], compute_graph.flexml_layers[127].compute_node[2][0].in[0], compute_graph.flexml_layers[128].compute_node[2][0].in[0], compute_graph.flexml_layers[129].compute_node[2][0].in[0], compute_graph.flexml_layers[130].compute_node[2][0].in[0], compute_graph.flexml_layers[130].compute_node[2][0].in[2], compute_graph.flexml_layers[131].compute_node[2][0].in[0], compute_graph.flexml_layers[132].compute_node[2][0].in[0], compute_graph.flexml_layers[133].compute_node[2][0].in[0], compute_graph.flexml_layers[134].compute_node[2][0].in[0], compute_graph.flexml_layers[135].compute_node[2][0].in[0], compute_graph.flexml_layers[135].compute_node[2][0].in[2], compute_graph.flexml_layers[136].compute_node[2][0].in[0], compute_graph.flexml_layers[137].compute_node[2][0].in[0], compute_graph.flexml_layers[138].compute_node[2][0].in[0], compute_graph.flexml_layers[139].compute_node[2][0].in[0], compute_graph.flexml_layers[140].compute_node[2][0].in[0], compute_graph.flexml_layers[141].compute_node[2][0].in[0], compute_graph.flexml_layers[142].compute_node[2][0].in[0], compute_graph.flexml_layers[142].compute_node[2][0].in[2], compute_graph.flexml_layers[145].compute_node[2][0].in[0], compute_graph.flexml_layers[146].compute_node[2][0].in[0], compute_graph.flexml_layers[147].compute_node[2][0].in[0], compute_graph.flexml_layers[148].compute_node[2][0].in[0], compute_graph.flexml_layers[148].compute_node[2][0].in[2], compute_graph.flexml_layers[149].compute_node[2][0].in[0], compute_graph.flexml_layers[150].compute_node[2][0].in[0], compute_graph.flexml_layers[151].compute_node[2][0].in[0], compute_graph.flexml_layers[152].compute_node[2][0].in[0], compute_graph.flexml_layers[153].compute_node[2][0].in[0], compute_graph.flexml_layers[153].compute_node[2][0].in[2], compute_graph.flexml_layers[154].compute_node[2][0].in[0], compute_graph.flexml_layers[155].compute_node[2][0].in[0], compute_graph.flexml_layers[156].compute_node[2][0].in[0], compute_graph.flexml_layers[157].compute_node[2][0].in[0], compute_graph.flexml_layers[158].compute_node[2][0].in[0], compute_graph.flexml_layers[159].compute_node[2][0].in[0], compute_graph.flexml_layers[160].compute_node[2][0].in[0], compute_graph.flexml_layers[160].compute_node[2][0].in[2], compute_graph.flexml_layers[161].compute_node[2][0].in[0], compute_graph.flexml_layers[161].compute_node[2][0].in[2], compute_graph.flexml_layers[162].compute_node[2][0].in[0], compute_graph.flexml_layers[163].compute_node[2][0].in[0], compute_graph.flexml_layers[164].compute_node[2][0].in[0], compute_graph.flexml_layers[165].compute_node[2][0].in[0], compute_graph.flexml_layers[166].compute_node[2][0].in[0], compute_graph.flexml_layers[166].compute_node[2][0].in[2], compute_graph.flexml_layers[167].compute_node[2][0].in[0], compute_graph.flexml_layers[168].compute_node[2][0].in[0], compute_graph.flexml_layers[169].compute_node[2][0].in[0], compute_graph.flexml_layers[170].compute_node[2][0].in[0], compute_graph.flexml_layers[171].compute_node[2][0].in[0], compute_graph.flexml_layers[171].compute_node[2][0].in[2], compute_graph.flexml_layers[172].compute_node[2][0].in[0], compute_graph.flexml_layers[173].compute_node[2][0].in[0], compute_graph.flexml_layers[174].compute_node[2][0].in[0], compute_graph.flexml_layers[175].compute_node[2][0].in[0], compute_graph.flexml_layers[176].compute_node[2][0].in[0], compute_graph.flexml_layers[177].compute_node[2][0].in[0], compute_graph.flexml_layers[178].compute_node[2][0].in[0], compute_graph.flexml_layers[178].compute_node[2][0].in[2], compute_graph.flexml_layers[179].compute_node[2][0].in[0], compute_graph.flexml_layers[179].compute_node[2][0].in[2], compute_graph.flexml_layers[180].compute_node[2][0].in[0], compute_graph.flexml_layers[181].compute_node[2][0].in[0], compute_graph.flexml_layers[182].compute_node[2][0].in[0], compute_graph.flexml_layers[183].compute_node[2][0].in[0], compute_graph.flexml_layers[184].compute_node[2][0].in[0], compute_graph.flexml_layers[184].compute_node[2][0].in[2], compute_graph.flexml_layers[185].compute_node[2][0].in[0], compute_graph.flexml_layers[186].compute_node[2][0].in[0], compute_graph.flexml_layers[187].compute_node[2][0].in[0], compute_graph.flexml_layers[188].compute_node[2][0].in[0], compute_graph.flexml_layers[188].compute_node[2][0].in[2], compute_graph.flexml_layers[189].compute_node[2][0].in[0], compute_graph.flexml_layers[190].compute_node[2][0].in[0], compute_graph.flexml_layers[191].compute_node[2][0].in[0], compute_graph.flexml_layers[191].compute_node[2][0].in[2], compute_graph.flexml_layers[192].compute_node[2][0].in[0], compute_graph.flexml_layers[192].compute_node[2][0].in[2], compute_graph.flexml_layers[193].compute_node[2][0].in[0], compute_graph.flexml_layers[193].compute_node[2][0].in[2], compute_graph.flexml_layers[194].compute_node[2][0].in[0], compute_graph.flexml_layers[195].compute_node[2][0].in[0], compute_graph.flexml_layers[196].compute_node[2][0].in[0], compute_graph.flexml_layers[196].compute_node[2][0].in[2], compute_graph.flexml_layers[197].compute_node[2][0].in[0], compute_graph.flexml_layers[197].compute_node[2][0].in[2], compute_graph.flexml_layers[198].compute_node[2][0].in[0], compute_graph.flexml_layers[199].compute_node[2][0].in[0], compute_graph.flexml_layers[203].compute_node[2][0].in[0], compute_graph.flexml_layers[204].compute_node[2][0].in[0], compute_graph.flexml_layers[204].compute_node[2][0].in[2], compute_graph.flexml_layers[205].compute_node[2][0].in[0], compute_graph.flexml_layers[205].compute_node[2][0].in[2], compute_graph.flexml_layers[206].compute_node[2][0].in[0], compute_graph.flexml_layers[206].compute_node[2][0].in[2], compute_graph.flexml_layers[207].compute_node[2][0].in[0], compute_graph.flexml_layers[208].compute_node[2][0].in[0], compute_graph.flexml_layers[209].compute_node[2][0].in[0], compute_graph.flexml_layers[209].compute_node[2][0].in[2], compute_graph.flexml_layers[210].compute_node[2][0].in[0], compute_graph.flexml_layers[210].compute_node[2][0].in[2], compute_graph.flexml_layers[211].compute_node[2][0].in[0], compute_graph.flexml_layers[212].compute_node[2][0].in[0], compute_graph.flexml_layers[213].compute_node[2][0].in[0], compute_graph.flexml_layers[214].compute_node[2][0].in[0], compute_graph.flexml_layers[214].compute_node[2][0].in[2], compute_graph.flexml_layers[215].compute_node[2][0].in[0], compute_graph.flexml_layers[215].compute_node[2][0].in[2], compute_graph.flexml_layers[216].compute_node[2][0].in[0], compute_graph.flexml_layers[216].compute_node[2][0].in[2], compute_graph.flexml_layers[217].compute_node[2][0].in[0], compute_graph.flexml_layers[218].compute_node[2][0].in[0], compute_graph.flexml_layers[219].compute_node[2][0].in[0], compute_graph.flexml_layers[219].compute_node[2][0].in[2], compute_graph.flexml_layers[220].compute_node[2][0].in[0], compute_graph.flexml_layers[220].compute_node[2][0].in[2], compute_graph.flexml_layers[221].compute_node[2][0].in[0], compute_graph.flexml_layers[222].compute_node[2][0].in[0], compute_graph.flexml_layers[223].compute_node[2][0].in[0], compute_graph.flexml_layers[224].compute_node[2][0].in[0], compute_graph.flexml_layers[224].compute_node[2][0].in[2], compute_graph.flexml_layers[225].compute_node[2][0].in[0], compute_graph.flexml_layers[225].compute_node[2][0].in[2], compute_graph.flexml_layers[226].compute_node[2][0].in[0], compute_graph.flexml_layers[226].compute_node[2][0].in[2], compute_graph.flexml_layers[227].compute_node[2][0].in[0], compute_graph.flexml_layers[228].compute_node[2][0].in[0], compute_graph.flexml_layers[229].compute_node[2][0].in[0], compute_graph.flexml_layers[229].compute_node[2][0].in[2], compute_graph.flexml_layers[230].compute_node[2][0].in[0], compute_graph.flexml_layers[230].compute_node[2][0].in[2], compute_graph.flexml_layers[231].compute_node[2][0].in[0], compute_graph.flexml_layers[232].compute_node[2][0].in[0], compute_graph.flexml_layers[233].compute_node[2][0].in[0], compute_graph.flexml_layers[234].compute_node[2][0].in[0], compute_graph.flexml_layers[235].compute_node[2][0].in[0], compute_graph.flexml_layers[235].compute_node[2][0].in[2], compute_graph.flexml_layers[236].compute_node[2][0].in[0]) + +Column 2Row 0Channel 0 + +i12_po0(compute_graph.templated_graph_4.trans_comp_nd[2][0].out[0], compute_graph.flexml_layers[36].compute_node[2][0].out[0], compute_graph.templated_graph_231.mk[2][0].out[0], compute_graph.templated_graph_254.mk[2][0].out[0], compute_graph.templated_graph_258.compute_node[2][0].out[0], compute_graph.templated_graph_259.compute_node[2][0].out[0], compute_graph.templated_graph_260.mk[2][0].out[0], compute_graph.templated_graph_263.compute_node[2][0].out[0], compute_graph.templated_graph_266.compute_node[2][0].out[0], compute_graph.templated_graph_268.mk[2][0].out[0], compute_graph.templated_graph_273.mk[2][0].out[0], compute_graph.templated_graph_274.mk[2][0].out[0], compute_graph.flexml_layers[63].compute_node[2][0].out[0], compute_graph.templated_graph_293.mk[2][0].out[0], compute_graph.templated_graph_298.compute_node[2][0].out[0], compute_graph.templated_graph_300.compute_node[2][0].out[0], compute_graph.flexml_layers[0].compute_node[2][0].out[0], compute_graph.flexml_layers[1].compute_node[2][0].out[0], compute_graph.flexml_layers[2].compute_node[2][0].out[0], compute_graph.flexml_layers[3].compute_node[2][0].out[0], compute_graph.flexml_layers[4].compute_node[2][0].out[0], compute_graph.flexml_layers[5].compute_node[2][0].out[0], compute_graph.flexml_layers[6].compute_node[2][0].out[0], compute_graph.flexml_layers[7].compute_node[2][0].out[0], compute_graph.flexml_layers[8].compute_node[2][0].out[0], compute_graph.flexml_layers[9].compute_node[2][0].out[0], compute_graph.flexml_layers[10].compute_node[2][0].out[0], compute_graph.flexml_layers[11].compute_node[2][0].out[0], compute_graph.flexml_layers[12].compute_node[2][0].out[0], compute_graph.flexml_layers[13].compute_node[2][0].out[0], compute_graph.flexml_layers[14].compute_node[2][0].out[0], compute_graph.flexml_layers[15].compute_node[2][0].out[0], compute_graph.flexml_layers[16].compute_node[2][0].out[0], compute_graph.flexml_layers[17].compute_node[2][0].out[0], compute_graph.flexml_layers[18].compute_node[2][0].out[0], compute_graph.flexml_layers[19].compute_node[2][0].out[0], compute_graph.flexml_layers[20].compute_node[2][0].out[0], compute_graph.flexml_layers[21].compute_node[2][0].out[0], compute_graph.flexml_layers[22].compute_node[2][0].out[0], compute_graph.flexml_layers[23].compute_node[2][0].out[0], compute_graph.flexml_layers[24].compute_node[2][0].out[0], compute_graph.flexml_layers[25].compute_node[2][0].out[0], compute_graph.flexml_layers[26].compute_node[2][0].out[0], compute_graph.flexml_layers[27].compute_node[2][0].out[0], compute_graph.flexml_layers[28].compute_node[2][0].out[0], compute_graph.flexml_layers[29].compute_node[2][0].out[0], compute_graph.flexml_layers[30].compute_node[2][0].out[0], compute_graph.flexml_layers[31].compute_node[2][0].out[0], compute_graph.flexml_layers[32].compute_node[2][0].out[0], compute_graph.flexml_layers[33].compute_node[2][0].out[0], compute_graph.flexml_layers[34].compute_node[2][0].out[0], compute_graph.flexml_layers[35].compute_node[2][0].out[0], compute_graph.flexml_layers[143].compute_node[2][0].out[0], compute_graph.flexml_layers[144].compute_node[2][0].out[0], compute_graph.flexml_layers[37].compute_node[2][0].out[0], compute_graph.flexml_layers[38].compute_node[2][0].out[0], compute_graph.flexml_layers[39].compute_node[2][0].out[0], compute_graph.flexml_layers[40].compute_node[2][0].out[0], compute_graph.flexml_layers[41].compute_node[2][0].out[0], compute_graph.flexml_layers[42].compute_node[2][0].out[0], compute_graph.flexml_layers[43].compute_node[2][0].out[0], compute_graph.flexml_layers[44].compute_node[2][0].out[0], compute_graph.flexml_layers[45].compute_node[2][0].out[0], compute_graph.flexml_layers[46].compute_node[2][0].out[0], compute_graph.flexml_layers[47].compute_node[2][0].out[0], compute_graph.flexml_layers[48].compute_node[2][0].out[0], compute_graph.flexml_layers[49].compute_node[2][0].out[0], compute_graph.flexml_layers[50].compute_node[2][0].out[0], compute_graph.flexml_layers[51].compute_node[2][0].out[0], compute_graph.flexml_layers[52].compute_node[2][0].out[0], compute_graph.flexml_layers[53].compute_node[2][0].out[0], compute_graph.flexml_layers[54].compute_node[2][0].out[0], compute_graph.flexml_layers[55].compute_node[2][0].out[0], compute_graph.flexml_layers[56].compute_node[2][0].out[0], compute_graph.flexml_layers[57].compute_node[2][0].out[0], compute_graph.flexml_layers[58].compute_node[2][0].out[0], compute_graph.flexml_layers[59].compute_node[2][0].out[0], compute_graph.flexml_layers[60].compute_node[2][0].out[0], compute_graph.flexml_layers[61].compute_node[2][0].out[0], compute_graph.flexml_layers[62].compute_node[2][0].out[0], compute_graph.flexml_layers[200].compute_node[2][0].out[0], compute_graph.flexml_layers[201].compute_node[2][0].out[0], compute_graph.flexml_layers[202].compute_node[2][0].out[0], compute_graph.flexml_layers[64].compute_node[2][0].out[0], compute_graph.flexml_layers[65].compute_node[2][0].out[0], compute_graph.flexml_layers[66].compute_node[2][0].out[0], compute_graph.flexml_layers[67].compute_node[2][0].out[0], compute_graph.flexml_layers[68].compute_node[2][0].out[0], compute_graph.flexml_layers[69].compute_node[2][0].out[0], compute_graph.flexml_layers[70].compute_node[2][0].out[0], compute_graph.flexml_layers[71].compute_node[2][0].out[0], compute_graph.flexml_layers[72].compute_node[2][0].out[0], compute_graph.flexml_layers[73].compute_node[2][0].out[0], compute_graph.flexml_layers[74].compute_node[2][0].out[0], compute_graph.flexml_layers[75].compute_node[2][0].out[0], compute_graph.flexml_layers[76].compute_node[2][0].out[0], compute_graph.flexml_layers[77].compute_node[2][0].out[0], compute_graph.flexml_layers[78].compute_node[2][0].out[0], compute_graph.flexml_layers[79].compute_node[2][0].out[0], compute_graph.flexml_layers[80].compute_node[2][0].out[0], compute_graph.flexml_layers[81].compute_node[2][0].out[0], compute_graph.flexml_layers[82].compute_node[2][0].out[0], compute_graph.flexml_layers[83].compute_node[2][0].out[0], compute_graph.flexml_layers[84].compute_node[2][0].out[0], compute_graph.flexml_layers[85].compute_node[2][0].out[0], compute_graph.flexml_layers[86].compute_node[2][0].out[0], compute_graph.flexml_layers[87].compute_node[2][0].out[0], compute_graph.flexml_layers[88].compute_node[2][0].out[0], compute_graph.flexml_layers[89].compute_node[2][0].out[0], compute_graph.flexml_layers[90].compute_node[2][0].out[0], compute_graph.flexml_layers[91].compute_node[2][0].out[0], compute_graph.flexml_layers[92].compute_node[2][0].out[0], compute_graph.flexml_layers[93].compute_node[2][0].out[0], compute_graph.flexml_layers[94].compute_node[2][0].out[0], compute_graph.flexml_layers[95].compute_node[2][0].out[0], compute_graph.flexml_layers[96].compute_node[2][0].out[0], compute_graph.flexml_layers[97].compute_node[2][0].out[0], compute_graph.flexml_layers[98].compute_node[2][0].out[0], compute_graph.flexml_layers[99].compute_node[2][0].out[0], compute_graph.flexml_layers[100].compute_node[2][0].out[0], compute_graph.flexml_layers[101].compute_node[2][0].out[0], compute_graph.flexml_layers[102].compute_node[2][0].out[0], compute_graph.flexml_layers[103].compute_node[2][0].out[0], compute_graph.flexml_layers[104].compute_node[2][0].out[0], compute_graph.flexml_layers[105].compute_node[2][0].out[0], compute_graph.flexml_layers[106].compute_node[2][0].out[0], compute_graph.flexml_layers[107].compute_node[2][0].out[0], compute_graph.flexml_layers[108].compute_node[2][0].out[0], compute_graph.flexml_layers[109].compute_node[2][0].out[0], compute_graph.flexml_layers[110].compute_node[2][0].out[0], compute_graph.flexml_layers[111].compute_node[2][0].out[0], compute_graph.flexml_layers[112].compute_node[2][0].out[0], compute_graph.flexml_layers[113].compute_node[2][0].out[0], compute_graph.flexml_layers[114].compute_node[2][0].out[0], compute_graph.flexml_layers[115].compute_node[2][0].out[0], compute_graph.flexml_layers[116].compute_node[2][0].out[0], compute_graph.flexml_layers[117].compute_node[2][0].out[0], compute_graph.flexml_layers[118].compute_node[2][0].out[0], compute_graph.flexml_layers[119].compute_node[2][0].out[0], compute_graph.flexml_layers[120].compute_node[2][0].out[0], compute_graph.flexml_layers[121].compute_node[2][0].out[0], compute_graph.flexml_layers[122].compute_node[2][0].out[0], compute_graph.flexml_layers[123].compute_node[2][0].out[0], compute_graph.flexml_layers[124].compute_node[2][0].out[0], compute_graph.flexml_layers[125].compute_node[2][0].out[0], compute_graph.flexml_layers[126].compute_node[2][0].out[0], compute_graph.flexml_layers[127].compute_node[2][0].out[0], compute_graph.flexml_layers[128].compute_node[2][0].out[0], compute_graph.flexml_layers[129].compute_node[2][0].out[0], compute_graph.flexml_layers[130].compute_node[2][0].out[0], compute_graph.flexml_layers[131].compute_node[2][0].out[0], compute_graph.flexml_layers[132].compute_node[2][0].out[0], compute_graph.flexml_layers[133].compute_node[2][0].out[0], compute_graph.flexml_layers[134].compute_node[2][0].out[0], compute_graph.flexml_layers[135].compute_node[2][0].out[0], compute_graph.flexml_layers[136].compute_node[2][0].out[0], compute_graph.flexml_layers[137].compute_node[2][0].out[0], compute_graph.flexml_layers[138].compute_node[2][0].out[0], compute_graph.flexml_layers[139].compute_node[2][0].out[0], compute_graph.flexml_layers[140].compute_node[2][0].out[0], compute_graph.flexml_layers[141].compute_node[2][0].out[0], compute_graph.flexml_layers[142].compute_node[2][0].out[0], compute_graph.flexml_layers[145].compute_node[2][0].out[0], compute_graph.flexml_layers[146].compute_node[2][0].out[0], compute_graph.flexml_layers[147].compute_node[2][0].out[0], compute_graph.flexml_layers[148].compute_node[2][0].out[0], compute_graph.flexml_layers[149].compute_node[2][0].out[0], compute_graph.flexml_layers[150].compute_node[2][0].out[0], compute_graph.flexml_layers[151].compute_node[2][0].out[0], compute_graph.flexml_layers[152].compute_node[2][0].out[0], compute_graph.flexml_layers[153].compute_node[2][0].out[0], compute_graph.flexml_layers[154].compute_node[2][0].out[0], compute_graph.flexml_layers[155].compute_node[2][0].out[0], compute_graph.flexml_layers[156].compute_node[2][0].out[0], compute_graph.flexml_layers[157].compute_node[2][0].out[0], compute_graph.flexml_layers[158].compute_node[2][0].out[0], compute_graph.flexml_layers[159].compute_node[2][0].out[0], compute_graph.flexml_layers[160].compute_node[2][0].out[0], compute_graph.flexml_layers[161].compute_node[2][0].out[0], compute_graph.flexml_layers[162].compute_node[2][0].out[0], compute_graph.flexml_layers[163].compute_node[2][0].out[0], compute_graph.flexml_layers[164].compute_node[2][0].out[0], compute_graph.flexml_layers[165].compute_node[2][0].out[0], compute_graph.flexml_layers[166].compute_node[2][0].out[0], compute_graph.flexml_layers[167].compute_node[2][0].out[0], compute_graph.flexml_layers[168].compute_node[2][0].out[0], compute_graph.flexml_layers[169].compute_node[2][0].out[0], compute_graph.flexml_layers[170].compute_node[2][0].out[0], compute_graph.flexml_layers[171].compute_node[2][0].out[0], compute_graph.flexml_layers[172].compute_node[2][0].out[0], compute_graph.flexml_layers[173].compute_node[2][0].out[0], compute_graph.flexml_layers[174].compute_node[2][0].out[0], compute_graph.flexml_layers[175].compute_node[2][0].out[0], compute_graph.flexml_layers[176].compute_node[2][0].out[0], compute_graph.flexml_layers[177].compute_node[2][0].out[0], compute_graph.flexml_layers[178].compute_node[2][0].out[0], compute_graph.flexml_layers[179].compute_node[2][0].out[0], compute_graph.flexml_layers[180].compute_node[2][0].out[0], compute_graph.flexml_layers[181].compute_node[2][0].out[0], compute_graph.flexml_layers[182].compute_node[2][0].out[0], compute_graph.flexml_layers[183].compute_node[2][0].out[0], compute_graph.flexml_layers[184].compute_node[2][0].out[0], compute_graph.flexml_layers[185].compute_node[2][0].out[0], compute_graph.flexml_layers[186].compute_node[2][0].out[0], compute_graph.flexml_layers[187].compute_node[2][0].out[0], compute_graph.flexml_layers[188].compute_node[2][0].out[0], compute_graph.flexml_layers[189].compute_node[2][0].out[0], compute_graph.flexml_layers[190].compute_node[2][0].out[0], compute_graph.flexml_layers[191].compute_node[2][0].out[0], compute_graph.flexml_layers[192].compute_node[2][0].out[0], compute_graph.flexml_layers[193].compute_node[2][0].out[0], compute_graph.flexml_layers[194].compute_node[2][0].out[0], compute_graph.flexml_layers[195].compute_node[2][0].out[0], compute_graph.flexml_layers[196].compute_node[2][0].out[0], compute_graph.flexml_layers[197].compute_node[2][0].out[0], compute_graph.flexml_layers[198].compute_node[2][0].out[0], compute_graph.flexml_layers[199].compute_node[2][0].out[0], compute_graph.flexml_layers[203].compute_node[2][0].out[0], compute_graph.flexml_layers[204].compute_node[2][0].out[0], compute_graph.flexml_layers[205].compute_node[2][0].out[0], compute_graph.flexml_layers[206].compute_node[2][0].out[0], compute_graph.flexml_layers[207].compute_node[2][0].out[0], compute_graph.flexml_layers[208].compute_node[2][0].out[0], compute_graph.flexml_layers[209].compute_node[2][0].out[0], compute_graph.flexml_layers[210].compute_node[2][0].out[0], compute_graph.flexml_layers[211].compute_node[2][0].out[0], compute_graph.flexml_layers[212].compute_node[2][0].out[0], compute_graph.flexml_layers[213].compute_node[2][0].out[0], compute_graph.flexml_layers[214].compute_node[2][0].out[0], compute_graph.flexml_layers[215].compute_node[2][0].out[0], compute_graph.flexml_layers[216].compute_node[2][0].out[0], compute_graph.flexml_layers[217].compute_node[2][0].out[0], compute_graph.flexml_layers[218].compute_node[2][0].out[0], compute_graph.flexml_layers[219].compute_node[2][0].out[0], compute_graph.flexml_layers[220].compute_node[2][0].out[0], compute_graph.flexml_layers[221].compute_node[2][0].out[0], compute_graph.flexml_layers[222].compute_node[2][0].out[0], compute_graph.flexml_layers[223].compute_node[2][0].out[0], compute_graph.flexml_layers[224].compute_node[2][0].out[0], compute_graph.flexml_layers[225].compute_node[2][0].out[0], compute_graph.flexml_layers[226].compute_node[2][0].out[0], compute_graph.flexml_layers[227].compute_node[2][0].out[0], compute_graph.flexml_layers[228].compute_node[2][0].out[0], compute_graph.flexml_layers[229].compute_node[2][0].out[0], compute_graph.flexml_layers[230].compute_node[2][0].out[0], compute_graph.flexml_layers[231].compute_node[2][0].out[0], compute_graph.flexml_layers[232].compute_node[2][0].out[0], compute_graph.flexml_layers[233].compute_node[2][0].out[0], compute_graph.flexml_layers[234].compute_node[2][0].out[0], compute_graph.flexml_layers[235].compute_node[2][0].out[0], compute_graph.flexml_layers[236].compute_node[2][0].out[0]) + +Column 2Row 0Channel 1 + +i12_pi1(compute_graph.flexml_layers[36].compute_node[2][0].in[1], compute_graph.templated_graph_260.mk[2][0].in[1], compute_graph.templated_graph_268.mk[2][0].in[1], compute_graph.templated_graph_273.mk[2][0].in[1], compute_graph.flexml_layers[3].compute_node[2][0].in[1], compute_graph.flexml_layers[8].compute_node[2][0].in[1], compute_graph.flexml_layers[9].compute_node[2][0].in[1], compute_graph.flexml_layers[10].compute_node[2][0].in[1], compute_graph.flexml_layers[11].compute_node[2][0].in[1], compute_graph.flexml_layers[12].compute_node[2][0].in[1], compute_graph.flexml_layers[13].compute_node[2][0].in[1], compute_graph.flexml_layers[14].compute_node[2][0].in[1], compute_graph.flexml_layers[15].compute_node[2][0].in[1], compute_graph.flexml_layers[16].compute_node[2][0].in[1], compute_graph.flexml_layers[17].compute_node[2][0].in[1], compute_graph.flexml_layers[19].compute_node[2][0].in[1], compute_graph.flexml_layers[20].compute_node[2][0].in[1], compute_graph.flexml_layers[25].compute_node[2][0].in[1], compute_graph.flexml_layers[26].compute_node[2][0].in[1], compute_graph.flexml_layers[27].compute_node[2][0].in[1], compute_graph.flexml_layers[29].compute_node[2][0].in[1], compute_graph.flexml_layers[30].compute_node[2][0].in[1], compute_graph.flexml_layers[35].compute_node[2][0].in[1], compute_graph.flexml_layers[143].compute_node[2][0].in[1], compute_graph.flexml_layers[144].compute_node[2][0].in[1], compute_graph.flexml_layers[37].compute_node[2][0].in[1], compute_graph.flexml_layers[39].compute_node[2][0].in[1], compute_graph.flexml_layers[40].compute_node[2][0].in[1], compute_graph.flexml_layers[45].compute_node[2][0].in[1], compute_graph.flexml_layers[46].compute_node[2][0].in[1], compute_graph.flexml_layers[51].compute_node[2][0].in[1], compute_graph.flexml_layers[56].compute_node[2][0].in[1], compute_graph.flexml_layers[57].compute_node[2][0].in[1], compute_graph.flexml_layers[62].compute_node[2][0].in[1], compute_graph.flexml_layers[201].compute_node[2][0].in[1], compute_graph.flexml_layers[202].compute_node[2][0].in[1], compute_graph.flexml_layers[67].compute_node[2][0].in[1], compute_graph.flexml_layers[68].compute_node[2][0].in[1], compute_graph.flexml_layers[73].compute_node[2][0].in[1], compute_graph.flexml_layers[78].compute_node[2][0].in[1], compute_graph.flexml_layers[79].compute_node[2][0].in[1], compute_graph.flexml_layers[84].compute_node[2][0].in[1], compute_graph.flexml_layers[89].compute_node[2][0].in[1], compute_graph.flexml_layers[90].compute_node[2][0].in[1], compute_graph.flexml_layers[95].compute_node[2][0].in[1], compute_graph.flexml_layers[101].compute_node[2][0].in[1], compute_graph.flexml_layers[102].compute_node[2][0].in[1], compute_graph.flexml_layers[107].compute_node[2][0].in[1], compute_graph.flexml_layers[108].compute_node[2][0].in[1], compute_graph.flexml_layers[113].compute_node[2][0].in[1], compute_graph.flexml_layers[119].compute_node[2][0].in[1], compute_graph.flexml_layers[120].compute_node[2][0].in[1], compute_graph.flexml_layers[125].compute_node[2][0].in[1], compute_graph.flexml_layers[126].compute_node[2][0].in[1], compute_graph.flexml_layers[131].compute_node[2][0].in[1], compute_graph.flexml_layers[137].compute_node[2][0].in[1], compute_graph.flexml_layers[138].compute_node[2][0].in[1], compute_graph.flexml_layers[149].compute_node[2][0].in[1], compute_graph.flexml_layers[155].compute_node[2][0].in[1], compute_graph.flexml_layers[156].compute_node[2][0].in[1], compute_graph.flexml_layers[161].compute_node[2][0].in[1], compute_graph.flexml_layers[162].compute_node[2][0].in[1], compute_graph.flexml_layers[167].compute_node[2][0].in[1], compute_graph.flexml_layers[173].compute_node[2][0].in[1], compute_graph.flexml_layers[174].compute_node[2][0].in[1], compute_graph.flexml_layers[179].compute_node[2][0].in[1], compute_graph.flexml_layers[180].compute_node[2][0].in[1], compute_graph.flexml_layers[186].compute_node[2][0].in[1], compute_graph.flexml_layers[188].compute_node[2][0].in[1], compute_graph.flexml_layers[189].compute_node[2][0].in[1], compute_graph.flexml_layers[194].compute_node[2][0].in[1], compute_graph.flexml_layers[207].compute_node[2][0].in[1], compute_graph.flexml_layers[211].compute_node[2][0].in[1], compute_graph.flexml_layers[212].compute_node[2][0].in[1], compute_graph.flexml_layers[217].compute_node[2][0].in[1], compute_graph.flexml_layers[221].compute_node[2][0].in[1], compute_graph.flexml_layers[222].compute_node[2][0].in[1], compute_graph.flexml_layers[227].compute_node[2][0].in[1], compute_graph.flexml_layers[231].compute_node[2][0].in[1], compute_graph.flexml_layers[232].compute_node[2][0].in[1], compute_graph.flexml_layers[233].compute_node[2][0].in[1]) + +Column 2Row 1Channel 0 + +i13_pi0(compute_graph.templated_graph_4.trans_comp_nd[2][1].in[0], compute_graph.flexml_layers[36].compute_node[2][1].in[0], compute_graph.templated_graph_231.mk[2][1].in[0], compute_graph.templated_graph_254.mk[2][1].in[0], compute_graph.templated_graph_258.compute_node[2][1].in[0], compute_graph.templated_graph_259.compute_node[2][1].in[0], compute_graph.templated_graph_260.mk[2][1].in[0], compute_graph.templated_graph_263.compute_node[2][1].in[0], compute_graph.templated_graph_266.compute_node[2][1].in[0], compute_graph.templated_graph_268.mk[2][1].in[0], compute_graph.templated_graph_273.mk[2][1].in[0], compute_graph.templated_graph_274.mk[2][1].in[0], compute_graph.flexml_layers[63].compute_node[2][1].in[0], compute_graph.templated_graph_293.mk[2][1].in[0], compute_graph.templated_graph_298.compute_node[2][1].in[0], compute_graph.templated_graph_300.compute_node[2][1].in[0], compute_graph.flexml_layers[0].compute_node[2][1].in[0], compute_graph.flexml_layers[1].compute_node[2][1].in[0], compute_graph.flexml_layers[1].compute_node[2][1].in[2], compute_graph.flexml_layers[2].compute_node[2][1].in[0], compute_graph.flexml_layers[2].compute_node[2][1].in[2], compute_graph.flexml_layers[3].compute_node[2][1].in[0], compute_graph.flexml_layers[4].compute_node[2][1].in[0], compute_graph.flexml_layers[5].compute_node[2][1].in[0], compute_graph.flexml_layers[6].compute_node[2][1].in[0], compute_graph.flexml_layers[7].compute_node[2][1].in[0], compute_graph.flexml_layers[7].compute_node[2][1].in[2], compute_graph.flexml_layers[8].compute_node[2][1].in[0], compute_graph.flexml_layers[9].compute_node[2][1].in[0], compute_graph.flexml_layers[9].compute_node[2][1].in[2], compute_graph.flexml_layers[10].compute_node[2][1].in[0], compute_graph.flexml_layers[11].compute_node[2][1].in[0], compute_graph.flexml_layers[12].compute_node[2][1].in[0], compute_graph.flexml_layers[13].compute_node[2][1].in[0], compute_graph.flexml_layers[14].compute_node[2][1].in[0], compute_graph.flexml_layers[15].compute_node[2][1].in[0], compute_graph.flexml_layers[15].compute_node[2][1].in[2], compute_graph.flexml_layers[16].compute_node[2][1].in[0], compute_graph.flexml_layers[17].compute_node[2][1].in[0], compute_graph.flexml_layers[18].compute_node[2][1].in[0], compute_graph.flexml_layers[19].compute_node[2][1].in[0], compute_graph.flexml_layers[20].compute_node[2][1].in[0], compute_graph.flexml_layers[21].compute_node[2][1].in[0], compute_graph.flexml_layers[22].compute_node[2][1].in[0], compute_graph.flexml_layers[23].compute_node[2][1].in[0], compute_graph.flexml_layers[24].compute_node[2][1].in[0], compute_graph.flexml_layers[24].compute_node[2][1].in[2], compute_graph.flexml_layers[25].compute_node[2][1].in[0], compute_graph.flexml_layers[26].compute_node[2][1].in[0], compute_graph.flexml_layers[27].compute_node[2][1].in[0], compute_graph.flexml_layers[28].compute_node[2][1].in[0], compute_graph.flexml_layers[29].compute_node[2][1].in[0], compute_graph.flexml_layers[30].compute_node[2][1].in[0], compute_graph.flexml_layers[31].compute_node[2][1].in[0], compute_graph.flexml_layers[32].compute_node[2][1].in[0], compute_graph.flexml_layers[33].compute_node[2][1].in[0], compute_graph.flexml_layers[34].compute_node[2][1].in[0], compute_graph.flexml_layers[34].compute_node[2][1].in[2], compute_graph.flexml_layers[35].compute_node[2][1].in[0], compute_graph.flexml_layers[35].compute_node[2][1].in[2], compute_graph.flexml_layers[143].compute_node[2][1].in[0], compute_graph.flexml_layers[144].compute_node[2][1].in[0], compute_graph.flexml_layers[37].compute_node[2][1].in[0], compute_graph.flexml_layers[38].compute_node[2][1].in[0], compute_graph.flexml_layers[39].compute_node[2][1].in[0], compute_graph.flexml_layers[40].compute_node[2][1].in[0], compute_graph.flexml_layers[41].compute_node[2][1].in[0], compute_graph.flexml_layers[42].compute_node[2][1].in[0], compute_graph.flexml_layers[43].compute_node[2][1].in[0], compute_graph.flexml_layers[44].compute_node[2][1].in[0], compute_graph.flexml_layers[44].compute_node[2][1].in[2], compute_graph.flexml_layers[45].compute_node[2][1].in[0], compute_graph.flexml_layers[45].compute_node[2][1].in[2], compute_graph.flexml_layers[46].compute_node[2][1].in[0], compute_graph.flexml_layers[47].compute_node[2][1].in[0], compute_graph.flexml_layers[48].compute_node[2][1].in[0], compute_graph.flexml_layers[49].compute_node[2][1].in[0], compute_graph.flexml_layers[50].compute_node[2][1].in[0], compute_graph.flexml_layers[50].compute_node[2][1].in[2], compute_graph.flexml_layers[51].compute_node[2][1].in[0], compute_graph.flexml_layers[52].compute_node[2][1].in[0], compute_graph.flexml_layers[53].compute_node[2][1].in[0], compute_graph.flexml_layers[54].compute_node[2][1].in[0], compute_graph.flexml_layers[55].compute_node[2][1].in[0], compute_graph.flexml_layers[55].compute_node[2][1].in[2], compute_graph.flexml_layers[56].compute_node[2][1].in[0], compute_graph.flexml_layers[57].compute_node[2][1].in[0], compute_graph.flexml_layers[58].compute_node[2][1].in[0], compute_graph.flexml_layers[59].compute_node[2][1].in[0], compute_graph.flexml_layers[60].compute_node[2][1].in[0], compute_graph.flexml_layers[61].compute_node[2][1].in[0], compute_graph.flexml_layers[61].compute_node[2][1].in[2], compute_graph.flexml_layers[62].compute_node[2][1].in[0], compute_graph.flexml_layers[200].compute_node[2][1].in[0], compute_graph.flexml_layers[201].compute_node[2][1].in[0], compute_graph.flexml_layers[202].compute_node[2][1].in[0], compute_graph.flexml_layers[64].compute_node[2][1].in[0], compute_graph.flexml_layers[65].compute_node[2][1].in[0], compute_graph.flexml_layers[66].compute_node[2][1].in[0], compute_graph.flexml_layers[66].compute_node[2][1].in[2], compute_graph.flexml_layers[67].compute_node[2][1].in[0], compute_graph.flexml_layers[67].compute_node[2][1].in[2], compute_graph.flexml_layers[68].compute_node[2][1].in[0], compute_graph.flexml_layers[69].compute_node[2][1].in[0], compute_graph.flexml_layers[70].compute_node[2][1].in[0], compute_graph.flexml_layers[71].compute_node[2][1].in[0], compute_graph.flexml_layers[72].compute_node[2][1].in[0], compute_graph.flexml_layers[72].compute_node[2][1].in[2], compute_graph.flexml_layers[73].compute_node[2][1].in[0], compute_graph.flexml_layers[74].compute_node[2][1].in[0], compute_graph.flexml_layers[75].compute_node[2][1].in[0], compute_graph.flexml_layers[76].compute_node[2][1].in[0], compute_graph.flexml_layers[77].compute_node[2][1].in[0], compute_graph.flexml_layers[77].compute_node[2][1].in[2], compute_graph.flexml_layers[78].compute_node[2][1].in[0], compute_graph.flexml_layers[78].compute_node[2][1].in[2], compute_graph.flexml_layers[79].compute_node[2][1].in[0], compute_graph.flexml_layers[80].compute_node[2][1].in[0], compute_graph.flexml_layers[81].compute_node[2][1].in[0], compute_graph.flexml_layers[82].compute_node[2][1].in[0], compute_graph.flexml_layers[83].compute_node[2][1].in[0], compute_graph.flexml_layers[83].compute_node[2][1].in[2], compute_graph.flexml_layers[84].compute_node[2][1].in[0], compute_graph.flexml_layers[85].compute_node[2][1].in[0], compute_graph.flexml_layers[86].compute_node[2][1].in[0], compute_graph.flexml_layers[87].compute_node[2][1].in[0], compute_graph.flexml_layers[88].compute_node[2][1].in[0], compute_graph.flexml_layers[88].compute_node[2][1].in[2], compute_graph.flexml_layers[89].compute_node[2][1].in[0], compute_graph.flexml_layers[89].compute_node[2][1].in[2], compute_graph.flexml_layers[90].compute_node[2][1].in[0], compute_graph.flexml_layers[91].compute_node[2][1].in[0], compute_graph.flexml_layers[92].compute_node[2][1].in[0], compute_graph.flexml_layers[93].compute_node[2][1].in[0], compute_graph.flexml_layers[94].compute_node[2][1].in[0], compute_graph.flexml_layers[94].compute_node[2][1].in[2], compute_graph.flexml_layers[95].compute_node[2][1].in[0], compute_graph.flexml_layers[96].compute_node[2][1].in[0], compute_graph.flexml_layers[97].compute_node[2][1].in[0], compute_graph.flexml_layers[98].compute_node[2][1].in[0], compute_graph.flexml_layers[99].compute_node[2][1].in[0], compute_graph.flexml_layers[99].compute_node[2][1].in[2], compute_graph.flexml_layers[100].compute_node[2][1].in[0], compute_graph.flexml_layers[101].compute_node[2][1].in[0], compute_graph.flexml_layers[102].compute_node[2][1].in[0], compute_graph.flexml_layers[103].compute_node[2][1].in[0], compute_graph.flexml_layers[104].compute_node[2][1].in[0], compute_graph.flexml_layers[105].compute_node[2][1].in[0], compute_graph.flexml_layers[106].compute_node[2][1].in[0], compute_graph.flexml_layers[106].compute_node[2][1].in[2], compute_graph.flexml_layers[107].compute_node[2][1].in[0], compute_graph.flexml_layers[108].compute_node[2][1].in[0], compute_graph.flexml_layers[109].compute_node[2][1].in[0], compute_graph.flexml_layers[110].compute_node[2][1].in[0], compute_graph.flexml_layers[111].compute_node[2][1].in[0], compute_graph.flexml_layers[112].compute_node[2][1].in[0], compute_graph.flexml_layers[112].compute_node[2][1].in[2], compute_graph.flexml_layers[113].compute_node[2][1].in[0], compute_graph.flexml_layers[114].compute_node[2][1].in[0], compute_graph.flexml_layers[115].compute_node[2][1].in[0], compute_graph.flexml_layers[116].compute_node[2][1].in[0], compute_graph.flexml_layers[117].compute_node[2][1].in[0], compute_graph.flexml_layers[117].compute_node[2][1].in[2], compute_graph.flexml_layers[118].compute_node[2][1].in[0], compute_graph.flexml_layers[119].compute_node[2][1].in[0], compute_graph.flexml_layers[120].compute_node[2][1].in[0], compute_graph.flexml_layers[121].compute_node[2][1].in[0], compute_graph.flexml_layers[122].compute_node[2][1].in[0], compute_graph.flexml_layers[123].compute_node[2][1].in[0], compute_graph.flexml_layers[124].compute_node[2][1].in[0], compute_graph.flexml_layers[124].compute_node[2][1].in[2], compute_graph.flexml_layers[125].compute_node[2][1].in[0], compute_graph.flexml_layers[125].compute_node[2][1].in[2], compute_graph.flexml_layers[126].compute_node[2][1].in[0], compute_graph.flexml_layers[127].compute_node[2][1].in[0], compute_graph.flexml_layers[128].compute_node[2][1].in[0], compute_graph.flexml_layers[129].compute_node[2][1].in[0], compute_graph.flexml_layers[130].compute_node[2][1].in[0], compute_graph.flexml_layers[130].compute_node[2][1].in[2], compute_graph.flexml_layers[131].compute_node[2][1].in[0], compute_graph.flexml_layers[132].compute_node[2][1].in[0], compute_graph.flexml_layers[133].compute_node[2][1].in[0], compute_graph.flexml_layers[134].compute_node[2][1].in[0], compute_graph.flexml_layers[135].compute_node[2][1].in[0], compute_graph.flexml_layers[135].compute_node[2][1].in[2], compute_graph.flexml_layers[136].compute_node[2][1].in[0], compute_graph.flexml_layers[137].compute_node[2][1].in[0], compute_graph.flexml_layers[138].compute_node[2][1].in[0], compute_graph.flexml_layers[139].compute_node[2][1].in[0], compute_graph.flexml_layers[140].compute_node[2][1].in[0], compute_graph.flexml_layers[141].compute_node[2][1].in[0], compute_graph.flexml_layers[142].compute_node[2][1].in[0], compute_graph.flexml_layers[142].compute_node[2][1].in[2], compute_graph.flexml_layers[145].compute_node[2][1].in[0], compute_graph.flexml_layers[146].compute_node[2][1].in[0], compute_graph.flexml_layers[147].compute_node[2][1].in[0], compute_graph.flexml_layers[148].compute_node[2][1].in[0], compute_graph.flexml_layers[148].compute_node[2][1].in[2], compute_graph.flexml_layers[149].compute_node[2][1].in[0], compute_graph.flexml_layers[150].compute_node[2][1].in[0], compute_graph.flexml_layers[151].compute_node[2][1].in[0], compute_graph.flexml_layers[152].compute_node[2][1].in[0], compute_graph.flexml_layers[153].compute_node[2][1].in[0], compute_graph.flexml_layers[153].compute_node[2][1].in[2], compute_graph.flexml_layers[154].compute_node[2][1].in[0], compute_graph.flexml_layers[155].compute_node[2][1].in[0], compute_graph.flexml_layers[156].compute_node[2][1].in[0], compute_graph.flexml_layers[157].compute_node[2][1].in[0], compute_graph.flexml_layers[158].compute_node[2][1].in[0], compute_graph.flexml_layers[159].compute_node[2][1].in[0], compute_graph.flexml_layers[160].compute_node[2][1].in[0], compute_graph.flexml_layers[160].compute_node[2][1].in[2], compute_graph.flexml_layers[161].compute_node[2][1].in[0], compute_graph.flexml_layers[161].compute_node[2][1].in[2], compute_graph.flexml_layers[162].compute_node[2][1].in[0], compute_graph.flexml_layers[163].compute_node[2][1].in[0], compute_graph.flexml_layers[164].compute_node[2][1].in[0], compute_graph.flexml_layers[165].compute_node[2][1].in[0], compute_graph.flexml_layers[166].compute_node[2][1].in[0], compute_graph.flexml_layers[166].compute_node[2][1].in[2], compute_graph.flexml_layers[167].compute_node[2][1].in[0], compute_graph.flexml_layers[168].compute_node[2][1].in[0], compute_graph.flexml_layers[169].compute_node[2][1].in[0], compute_graph.flexml_layers[170].compute_node[2][1].in[0], compute_graph.flexml_layers[171].compute_node[2][1].in[0], compute_graph.flexml_layers[171].compute_node[2][1].in[2], compute_graph.flexml_layers[172].compute_node[2][1].in[0], compute_graph.flexml_layers[173].compute_node[2][1].in[0], compute_graph.flexml_layers[174].compute_node[2][1].in[0], compute_graph.flexml_layers[175].compute_node[2][1].in[0], compute_graph.flexml_layers[176].compute_node[2][1].in[0], compute_graph.flexml_layers[177].compute_node[2][1].in[0], compute_graph.flexml_layers[178].compute_node[2][1].in[0], compute_graph.flexml_layers[178].compute_node[2][1].in[2], compute_graph.flexml_layers[179].compute_node[2][1].in[0], compute_graph.flexml_layers[179].compute_node[2][1].in[2], compute_graph.flexml_layers[180].compute_node[2][1].in[0], compute_graph.flexml_layers[181].compute_node[2][1].in[0], compute_graph.flexml_layers[182].compute_node[2][1].in[0], compute_graph.flexml_layers[183].compute_node[2][1].in[0], compute_graph.flexml_layers[184].compute_node[2][1].in[0], compute_graph.flexml_layers[184].compute_node[2][1].in[2], compute_graph.flexml_layers[185].compute_node[2][1].in[0], compute_graph.flexml_layers[186].compute_node[2][1].in[0], compute_graph.flexml_layers[187].compute_node[2][1].in[0], compute_graph.flexml_layers[188].compute_node[2][1].in[0], compute_graph.flexml_layers[188].compute_node[2][1].in[2], compute_graph.flexml_layers[189].compute_node[2][1].in[0], compute_graph.flexml_layers[190].compute_node[2][1].in[0], compute_graph.flexml_layers[191].compute_node[2][1].in[0], compute_graph.flexml_layers[191].compute_node[2][1].in[2], compute_graph.flexml_layers[192].compute_node[2][1].in[0], compute_graph.flexml_layers[192].compute_node[2][1].in[2], compute_graph.flexml_layers[193].compute_node[2][1].in[0], compute_graph.flexml_layers[193].compute_node[2][1].in[2], compute_graph.flexml_layers[194].compute_node[2][1].in[0], compute_graph.flexml_layers[195].compute_node[2][1].in[0], compute_graph.flexml_layers[196].compute_node[2][1].in[0], compute_graph.flexml_layers[196].compute_node[2][1].in[2], compute_graph.flexml_layers[197].compute_node[2][1].in[0], compute_graph.flexml_layers[197].compute_node[2][1].in[2], compute_graph.flexml_layers[198].compute_node[2][1].in[0], compute_graph.flexml_layers[199].compute_node[2][1].in[0], compute_graph.flexml_layers[203].compute_node[2][1].in[0], compute_graph.flexml_layers[204].compute_node[2][1].in[0], compute_graph.flexml_layers[204].compute_node[2][1].in[2], compute_graph.flexml_layers[205].compute_node[2][1].in[0], compute_graph.flexml_layers[205].compute_node[2][1].in[2], compute_graph.flexml_layers[206].compute_node[2][1].in[0], compute_graph.flexml_layers[206].compute_node[2][1].in[2], compute_graph.flexml_layers[207].compute_node[2][1].in[0], compute_graph.flexml_layers[208].compute_node[2][1].in[0], compute_graph.flexml_layers[209].compute_node[2][1].in[0], compute_graph.flexml_layers[209].compute_node[2][1].in[2], compute_graph.flexml_layers[210].compute_node[2][1].in[0], compute_graph.flexml_layers[210].compute_node[2][1].in[2], compute_graph.flexml_layers[211].compute_node[2][1].in[0], compute_graph.flexml_layers[212].compute_node[2][1].in[0], compute_graph.flexml_layers[213].compute_node[2][1].in[0], compute_graph.flexml_layers[214].compute_node[2][1].in[0], compute_graph.flexml_layers[214].compute_node[2][1].in[2], compute_graph.flexml_layers[215].compute_node[2][1].in[0], compute_graph.flexml_layers[215].compute_node[2][1].in[2], compute_graph.flexml_layers[216].compute_node[2][1].in[0], compute_graph.flexml_layers[216].compute_node[2][1].in[2], compute_graph.flexml_layers[217].compute_node[2][1].in[0], compute_graph.flexml_layers[218].compute_node[2][1].in[0], compute_graph.flexml_layers[219].compute_node[2][1].in[0], compute_graph.flexml_layers[219].compute_node[2][1].in[2], compute_graph.flexml_layers[220].compute_node[2][1].in[0], compute_graph.flexml_layers[220].compute_node[2][1].in[2], compute_graph.flexml_layers[221].compute_node[2][1].in[0], compute_graph.flexml_layers[222].compute_node[2][1].in[0], compute_graph.flexml_layers[223].compute_node[2][1].in[0], compute_graph.flexml_layers[224].compute_node[2][1].in[0], compute_graph.flexml_layers[224].compute_node[2][1].in[2], compute_graph.flexml_layers[225].compute_node[2][1].in[0], compute_graph.flexml_layers[225].compute_node[2][1].in[2], compute_graph.flexml_layers[226].compute_node[2][1].in[0], compute_graph.flexml_layers[226].compute_node[2][1].in[2], compute_graph.flexml_layers[227].compute_node[2][1].in[0], compute_graph.flexml_layers[228].compute_node[2][1].in[0], compute_graph.flexml_layers[229].compute_node[2][1].in[0], compute_graph.flexml_layers[229].compute_node[2][1].in[2], compute_graph.flexml_layers[230].compute_node[2][1].in[0], compute_graph.flexml_layers[230].compute_node[2][1].in[2], compute_graph.flexml_layers[231].compute_node[2][1].in[0], compute_graph.flexml_layers[232].compute_node[2][1].in[0], compute_graph.flexml_layers[233].compute_node[2][1].in[0], compute_graph.flexml_layers[234].compute_node[2][1].in[0], compute_graph.flexml_layers[235].compute_node[2][1].in[0], compute_graph.flexml_layers[235].compute_node[2][1].in[2], compute_graph.flexml_layers[236].compute_node[2][1].in[0]) + +Column 2Row 1Channel 0 + +i13_po0(compute_graph.templated_graph_4.trans_comp_nd[2][1].out[0], compute_graph.flexml_layers[36].compute_node[2][1].out[0], compute_graph.templated_graph_231.mk[2][1].out[0], compute_graph.templated_graph_254.mk[2][1].out[0], compute_graph.templated_graph_258.compute_node[2][1].out[0], compute_graph.templated_graph_259.compute_node[2][1].out[0], compute_graph.templated_graph_260.mk[2][1].out[0], compute_graph.templated_graph_263.compute_node[2][1].out[0], compute_graph.templated_graph_266.compute_node[2][1].out[0], compute_graph.templated_graph_268.mk[2][1].out[0], compute_graph.templated_graph_273.mk[2][1].out[0], compute_graph.templated_graph_274.mk[2][1].out[0], compute_graph.flexml_layers[63].compute_node[2][1].out[0], compute_graph.templated_graph_293.mk[2][1].out[0], compute_graph.templated_graph_298.compute_node[2][1].out[0], compute_graph.templated_graph_300.compute_node[2][1].out[0], compute_graph.flexml_layers[0].compute_node[2][1].out[0], compute_graph.flexml_layers[1].compute_node[2][1].out[0], compute_graph.flexml_layers[2].compute_node[2][1].out[0], compute_graph.flexml_layers[3].compute_node[2][1].out[0], compute_graph.flexml_layers[4].compute_node[2][1].out[0], compute_graph.flexml_layers[5].compute_node[2][1].out[0], compute_graph.flexml_layers[6].compute_node[2][1].out[0], compute_graph.flexml_layers[7].compute_node[2][1].out[0], compute_graph.flexml_layers[8].compute_node[2][1].out[0], compute_graph.flexml_layers[9].compute_node[2][1].out[0], compute_graph.flexml_layers[10].compute_node[2][1].out[0], compute_graph.flexml_layers[11].compute_node[2][1].out[0], compute_graph.flexml_layers[12].compute_node[2][1].out[0], compute_graph.flexml_layers[13].compute_node[2][1].out[0], compute_graph.flexml_layers[14].compute_node[2][1].out[0], compute_graph.flexml_layers[15].compute_node[2][1].out[0], compute_graph.flexml_layers[16].compute_node[2][1].out[0], compute_graph.flexml_layers[17].compute_node[2][1].out[0], compute_graph.flexml_layers[18].compute_node[2][1].out[0], compute_graph.flexml_layers[19].compute_node[2][1].out[0], compute_graph.flexml_layers[20].compute_node[2][1].out[0], compute_graph.flexml_layers[21].compute_node[2][1].out[0], compute_graph.flexml_layers[22].compute_node[2][1].out[0], compute_graph.flexml_layers[23].compute_node[2][1].out[0], compute_graph.flexml_layers[24].compute_node[2][1].out[0], compute_graph.flexml_layers[25].compute_node[2][1].out[0], compute_graph.flexml_layers[26].compute_node[2][1].out[0], compute_graph.flexml_layers[27].compute_node[2][1].out[0], compute_graph.flexml_layers[28].compute_node[2][1].out[0], compute_graph.flexml_layers[29].compute_node[2][1].out[0], compute_graph.flexml_layers[30].compute_node[2][1].out[0], compute_graph.flexml_layers[31].compute_node[2][1].out[0], compute_graph.flexml_layers[32].compute_node[2][1].out[0], compute_graph.flexml_layers[33].compute_node[2][1].out[0], compute_graph.flexml_layers[34].compute_node[2][1].out[0], compute_graph.flexml_layers[35].compute_node[2][1].out[0], compute_graph.flexml_layers[143].compute_node[2][1].out[0], compute_graph.flexml_layers[144].compute_node[2][1].out[0], compute_graph.flexml_layers[37].compute_node[2][1].out[0], compute_graph.flexml_layers[38].compute_node[2][1].out[0], compute_graph.flexml_layers[39].compute_node[2][1].out[0], compute_graph.flexml_layers[40].compute_node[2][1].out[0], compute_graph.flexml_layers[41].compute_node[2][1].out[0], compute_graph.flexml_layers[42].compute_node[2][1].out[0], compute_graph.flexml_layers[43].compute_node[2][1].out[0], compute_graph.flexml_layers[44].compute_node[2][1].out[0], compute_graph.flexml_layers[45].compute_node[2][1].out[0], compute_graph.flexml_layers[46].compute_node[2][1].out[0], compute_graph.flexml_layers[47].compute_node[2][1].out[0], compute_graph.flexml_layers[48].compute_node[2][1].out[0], compute_graph.flexml_layers[49].compute_node[2][1].out[0], compute_graph.flexml_layers[50].compute_node[2][1].out[0], compute_graph.flexml_layers[51].compute_node[2][1].out[0], compute_graph.flexml_layers[52].compute_node[2][1].out[0], compute_graph.flexml_layers[53].compute_node[2][1].out[0], compute_graph.flexml_layers[54].compute_node[2][1].out[0], compute_graph.flexml_layers[55].compute_node[2][1].out[0], compute_graph.flexml_layers[56].compute_node[2][1].out[0], compute_graph.flexml_layers[57].compute_node[2][1].out[0], compute_graph.flexml_layers[58].compute_node[2][1].out[0], compute_graph.flexml_layers[59].compute_node[2][1].out[0], compute_graph.flexml_layers[60].compute_node[2][1].out[0], compute_graph.flexml_layers[61].compute_node[2][1].out[0], compute_graph.flexml_layers[62].compute_node[2][1].out[0], compute_graph.flexml_layers[200].compute_node[2][1].out[0], compute_graph.flexml_layers[201].compute_node[2][1].out[0], compute_graph.flexml_layers[202].compute_node[2][1].out[0], compute_graph.flexml_layers[64].compute_node[2][1].out[0], compute_graph.flexml_layers[65].compute_node[2][1].out[0], compute_graph.flexml_layers[66].compute_node[2][1].out[0], compute_graph.flexml_layers[67].compute_node[2][1].out[0], compute_graph.flexml_layers[68].compute_node[2][1].out[0], compute_graph.flexml_layers[69].compute_node[2][1].out[0], compute_graph.flexml_layers[70].compute_node[2][1].out[0], compute_graph.flexml_layers[71].compute_node[2][1].out[0], compute_graph.flexml_layers[72].compute_node[2][1].out[0], compute_graph.flexml_layers[73].compute_node[2][1].out[0], compute_graph.flexml_layers[74].compute_node[2][1].out[0], compute_graph.flexml_layers[75].compute_node[2][1].out[0], compute_graph.flexml_layers[76].compute_node[2][1].out[0], compute_graph.flexml_layers[77].compute_node[2][1].out[0], compute_graph.flexml_layers[78].compute_node[2][1].out[0], compute_graph.flexml_layers[79].compute_node[2][1].out[0], compute_graph.flexml_layers[80].compute_node[2][1].out[0], compute_graph.flexml_layers[81].compute_node[2][1].out[0], compute_graph.flexml_layers[82].compute_node[2][1].out[0], compute_graph.flexml_layers[83].compute_node[2][1].out[0], compute_graph.flexml_layers[84].compute_node[2][1].out[0], compute_graph.flexml_layers[85].compute_node[2][1].out[0], compute_graph.flexml_layers[86].compute_node[2][1].out[0], compute_graph.flexml_layers[87].compute_node[2][1].out[0], compute_graph.flexml_layers[88].compute_node[2][1].out[0], compute_graph.flexml_layers[89].compute_node[2][1].out[0], compute_graph.flexml_layers[90].compute_node[2][1].out[0], compute_graph.flexml_layers[91].compute_node[2][1].out[0], compute_graph.flexml_layers[92].compute_node[2][1].out[0], compute_graph.flexml_layers[93].compute_node[2][1].out[0], compute_graph.flexml_layers[94].compute_node[2][1].out[0], compute_graph.flexml_layers[95].compute_node[2][1].out[0], compute_graph.flexml_layers[96].compute_node[2][1].out[0], compute_graph.flexml_layers[97].compute_node[2][1].out[0], compute_graph.flexml_layers[98].compute_node[2][1].out[0], compute_graph.flexml_layers[99].compute_node[2][1].out[0], compute_graph.flexml_layers[100].compute_node[2][1].out[0], compute_graph.flexml_layers[101].compute_node[2][1].out[0], compute_graph.flexml_layers[102].compute_node[2][1].out[0], compute_graph.flexml_layers[103].compute_node[2][1].out[0], compute_graph.flexml_layers[104].compute_node[2][1].out[0], compute_graph.flexml_layers[105].compute_node[2][1].out[0], compute_graph.flexml_layers[106].compute_node[2][1].out[0], compute_graph.flexml_layers[107].compute_node[2][1].out[0], compute_graph.flexml_layers[108].compute_node[2][1].out[0], compute_graph.flexml_layers[109].compute_node[2][1].out[0], compute_graph.flexml_layers[110].compute_node[2][1].out[0], compute_graph.flexml_layers[111].compute_node[2][1].out[0], compute_graph.flexml_layers[112].compute_node[2][1].out[0], compute_graph.flexml_layers[113].compute_node[2][1].out[0], compute_graph.flexml_layers[114].compute_node[2][1].out[0], compute_graph.flexml_layers[115].compute_node[2][1].out[0], compute_graph.flexml_layers[116].compute_node[2][1].out[0], compute_graph.flexml_layers[117].compute_node[2][1].out[0], compute_graph.flexml_layers[118].compute_node[2][1].out[0], compute_graph.flexml_layers[119].compute_node[2][1].out[0], compute_graph.flexml_layers[120].compute_node[2][1].out[0], compute_graph.flexml_layers[121].compute_node[2][1].out[0], compute_graph.flexml_layers[122].compute_node[2][1].out[0], compute_graph.flexml_layers[123].compute_node[2][1].out[0], compute_graph.flexml_layers[124].compute_node[2][1].out[0], compute_graph.flexml_layers[125].compute_node[2][1].out[0], compute_graph.flexml_layers[126].compute_node[2][1].out[0], compute_graph.flexml_layers[127].compute_node[2][1].out[0], compute_graph.flexml_layers[128].compute_node[2][1].out[0], compute_graph.flexml_layers[129].compute_node[2][1].out[0], compute_graph.flexml_layers[130].compute_node[2][1].out[0], compute_graph.flexml_layers[131].compute_node[2][1].out[0], compute_graph.flexml_layers[132].compute_node[2][1].out[0], compute_graph.flexml_layers[133].compute_node[2][1].out[0], compute_graph.flexml_layers[134].compute_node[2][1].out[0], compute_graph.flexml_layers[135].compute_node[2][1].out[0], compute_graph.flexml_layers[136].compute_node[2][1].out[0], compute_graph.flexml_layers[137].compute_node[2][1].out[0], compute_graph.flexml_layers[138].compute_node[2][1].out[0], compute_graph.flexml_layers[139].compute_node[2][1].out[0], compute_graph.flexml_layers[140].compute_node[2][1].out[0], compute_graph.flexml_layers[141].compute_node[2][1].out[0], compute_graph.flexml_layers[142].compute_node[2][1].out[0], compute_graph.flexml_layers[145].compute_node[2][1].out[0], compute_graph.flexml_layers[146].compute_node[2][1].out[0], compute_graph.flexml_layers[147].compute_node[2][1].out[0], compute_graph.flexml_layers[148].compute_node[2][1].out[0], compute_graph.flexml_layers[149].compute_node[2][1].out[0], compute_graph.flexml_layers[150].compute_node[2][1].out[0], compute_graph.flexml_layers[151].compute_node[2][1].out[0], compute_graph.flexml_layers[152].compute_node[2][1].out[0], compute_graph.flexml_layers[153].compute_node[2][1].out[0], compute_graph.flexml_layers[154].compute_node[2][1].out[0], compute_graph.flexml_layers[155].compute_node[2][1].out[0], compute_graph.flexml_layers[156].compute_node[2][1].out[0], compute_graph.flexml_layers[157].compute_node[2][1].out[0], compute_graph.flexml_layers[158].compute_node[2][1].out[0], compute_graph.flexml_layers[159].compute_node[2][1].out[0], compute_graph.flexml_layers[160].compute_node[2][1].out[0], compute_graph.flexml_layers[161].compute_node[2][1].out[0], compute_graph.flexml_layers[162].compute_node[2][1].out[0], compute_graph.flexml_layers[163].compute_node[2][1].out[0], compute_graph.flexml_layers[164].compute_node[2][1].out[0], compute_graph.flexml_layers[165].compute_node[2][1].out[0], compute_graph.flexml_layers[166].compute_node[2][1].out[0], compute_graph.flexml_layers[167].compute_node[2][1].out[0], compute_graph.flexml_layers[168].compute_node[2][1].out[0], compute_graph.flexml_layers[169].compute_node[2][1].out[0], compute_graph.flexml_layers[170].compute_node[2][1].out[0], compute_graph.flexml_layers[171].compute_node[2][1].out[0], compute_graph.flexml_layers[172].compute_node[2][1].out[0], compute_graph.flexml_layers[173].compute_node[2][1].out[0], compute_graph.flexml_layers[174].compute_node[2][1].out[0], compute_graph.flexml_layers[175].compute_node[2][1].out[0], compute_graph.flexml_layers[176].compute_node[2][1].out[0], compute_graph.flexml_layers[177].compute_node[2][1].out[0], compute_graph.flexml_layers[178].compute_node[2][1].out[0], compute_graph.flexml_layers[179].compute_node[2][1].out[0], compute_graph.flexml_layers[180].compute_node[2][1].out[0], compute_graph.flexml_layers[181].compute_node[2][1].out[0], compute_graph.flexml_layers[182].compute_node[2][1].out[0], compute_graph.flexml_layers[183].compute_node[2][1].out[0], compute_graph.flexml_layers[184].compute_node[2][1].out[0], compute_graph.flexml_layers[185].compute_node[2][1].out[0], compute_graph.flexml_layers[186].compute_node[2][1].out[0], compute_graph.flexml_layers[187].compute_node[2][1].out[0], compute_graph.flexml_layers[188].compute_node[2][1].out[0], compute_graph.flexml_layers[189].compute_node[2][1].out[0], compute_graph.flexml_layers[190].compute_node[2][1].out[0], compute_graph.flexml_layers[191].compute_node[2][1].out[0], compute_graph.flexml_layers[192].compute_node[2][1].out[0], compute_graph.flexml_layers[193].compute_node[2][1].out[0], compute_graph.flexml_layers[194].compute_node[2][1].out[0], compute_graph.flexml_layers[195].compute_node[2][1].out[0], compute_graph.flexml_layers[196].compute_node[2][1].out[0], compute_graph.flexml_layers[197].compute_node[2][1].out[0], compute_graph.flexml_layers[198].compute_node[2][1].out[0], compute_graph.flexml_layers[199].compute_node[2][1].out[0], compute_graph.flexml_layers[203].compute_node[2][1].out[0], compute_graph.flexml_layers[204].compute_node[2][1].out[0], compute_graph.flexml_layers[205].compute_node[2][1].out[0], compute_graph.flexml_layers[206].compute_node[2][1].out[0], compute_graph.flexml_layers[207].compute_node[2][1].out[0], compute_graph.flexml_layers[208].compute_node[2][1].out[0], compute_graph.flexml_layers[209].compute_node[2][1].out[0], compute_graph.flexml_layers[210].compute_node[2][1].out[0], compute_graph.flexml_layers[211].compute_node[2][1].out[0], compute_graph.flexml_layers[212].compute_node[2][1].out[0], compute_graph.flexml_layers[213].compute_node[2][1].out[0], compute_graph.flexml_layers[214].compute_node[2][1].out[0], compute_graph.flexml_layers[215].compute_node[2][1].out[0], compute_graph.flexml_layers[216].compute_node[2][1].out[0], compute_graph.flexml_layers[217].compute_node[2][1].out[0], compute_graph.flexml_layers[218].compute_node[2][1].out[0], compute_graph.flexml_layers[219].compute_node[2][1].out[0], compute_graph.flexml_layers[220].compute_node[2][1].out[0], compute_graph.flexml_layers[221].compute_node[2][1].out[0], compute_graph.flexml_layers[222].compute_node[2][1].out[0], compute_graph.flexml_layers[223].compute_node[2][1].out[0], compute_graph.flexml_layers[224].compute_node[2][1].out[0], compute_graph.flexml_layers[225].compute_node[2][1].out[0], compute_graph.flexml_layers[226].compute_node[2][1].out[0], compute_graph.flexml_layers[227].compute_node[2][1].out[0], compute_graph.flexml_layers[228].compute_node[2][1].out[0], compute_graph.flexml_layers[229].compute_node[2][1].out[0], compute_graph.flexml_layers[230].compute_node[2][1].out[0], compute_graph.flexml_layers[231].compute_node[2][1].out[0], compute_graph.flexml_layers[232].compute_node[2][1].out[0], compute_graph.flexml_layers[233].compute_node[2][1].out[0], compute_graph.flexml_layers[234].compute_node[2][1].out[0], compute_graph.flexml_layers[235].compute_node[2][1].out[0], compute_graph.flexml_layers[236].compute_node[2][1].out[0]) + +Column 2Row 1Channel 1 + +i13_pi1(compute_graph.flexml_layers[36].compute_node[2][1].in[1], compute_graph.templated_graph_260.mk[2][1].in[1], compute_graph.templated_graph_268.mk[2][1].in[1], compute_graph.templated_graph_273.mk[2][1].in[1], compute_graph.flexml_layers[3].compute_node[2][1].in[1], compute_graph.flexml_layers[8].compute_node[2][1].in[1], compute_graph.flexml_layers[9].compute_node[2][1].in[1], compute_graph.flexml_layers[10].compute_node[2][1].in[1], compute_graph.flexml_layers[11].compute_node[2][1].in[1], compute_graph.flexml_layers[12].compute_node[2][1].in[1], compute_graph.flexml_layers[13].compute_node[2][1].in[1], compute_graph.flexml_layers[14].compute_node[2][1].in[1], compute_graph.flexml_layers[15].compute_node[2][1].in[1], compute_graph.flexml_layers[16].compute_node[2][1].in[1], compute_graph.flexml_layers[17].compute_node[2][1].in[1], compute_graph.flexml_layers[19].compute_node[2][1].in[1], compute_graph.flexml_layers[20].compute_node[2][1].in[1], compute_graph.flexml_layers[25].compute_node[2][1].in[1], compute_graph.flexml_layers[26].compute_node[2][1].in[1], compute_graph.flexml_layers[27].compute_node[2][1].in[1], compute_graph.flexml_layers[29].compute_node[2][1].in[1], compute_graph.flexml_layers[30].compute_node[2][1].in[1], compute_graph.flexml_layers[35].compute_node[2][1].in[1], compute_graph.flexml_layers[143].compute_node[2][1].in[1], compute_graph.flexml_layers[144].compute_node[2][1].in[1], compute_graph.flexml_layers[37].compute_node[2][1].in[1], compute_graph.flexml_layers[39].compute_node[2][1].in[1], compute_graph.flexml_layers[40].compute_node[2][1].in[1], compute_graph.flexml_layers[45].compute_node[2][1].in[1], compute_graph.flexml_layers[46].compute_node[2][1].in[1], compute_graph.flexml_layers[51].compute_node[2][1].in[1], compute_graph.flexml_layers[56].compute_node[2][1].in[1], compute_graph.flexml_layers[57].compute_node[2][1].in[1], compute_graph.flexml_layers[62].compute_node[2][1].in[1], compute_graph.flexml_layers[201].compute_node[2][1].in[1], compute_graph.flexml_layers[202].compute_node[2][1].in[1], compute_graph.flexml_layers[67].compute_node[2][1].in[1], compute_graph.flexml_layers[68].compute_node[2][1].in[1], compute_graph.flexml_layers[73].compute_node[2][1].in[1], compute_graph.flexml_layers[78].compute_node[2][1].in[1], compute_graph.flexml_layers[79].compute_node[2][1].in[1], compute_graph.flexml_layers[84].compute_node[2][1].in[1], compute_graph.flexml_layers[89].compute_node[2][1].in[1], compute_graph.flexml_layers[90].compute_node[2][1].in[1], compute_graph.flexml_layers[95].compute_node[2][1].in[1], compute_graph.flexml_layers[101].compute_node[2][1].in[1], compute_graph.flexml_layers[102].compute_node[2][1].in[1], compute_graph.flexml_layers[107].compute_node[2][1].in[1], compute_graph.flexml_layers[108].compute_node[2][1].in[1], compute_graph.flexml_layers[113].compute_node[2][1].in[1], compute_graph.flexml_layers[119].compute_node[2][1].in[1], compute_graph.flexml_layers[120].compute_node[2][1].in[1], compute_graph.flexml_layers[125].compute_node[2][1].in[1], compute_graph.flexml_layers[126].compute_node[2][1].in[1], compute_graph.flexml_layers[131].compute_node[2][1].in[1], compute_graph.flexml_layers[137].compute_node[2][1].in[1], compute_graph.flexml_layers[138].compute_node[2][1].in[1], compute_graph.flexml_layers[149].compute_node[2][1].in[1], compute_graph.flexml_layers[155].compute_node[2][1].in[1], compute_graph.flexml_layers[156].compute_node[2][1].in[1], compute_graph.flexml_layers[161].compute_node[2][1].in[1], compute_graph.flexml_layers[162].compute_node[2][1].in[1], compute_graph.flexml_layers[167].compute_node[2][1].in[1], compute_graph.flexml_layers[173].compute_node[2][1].in[1], compute_graph.flexml_layers[174].compute_node[2][1].in[1], compute_graph.flexml_layers[179].compute_node[2][1].in[1], compute_graph.flexml_layers[180].compute_node[2][1].in[1], compute_graph.flexml_layers[186].compute_node[2][1].in[1], compute_graph.flexml_layers[188].compute_node[2][1].in[1], compute_graph.flexml_layers[189].compute_node[2][1].in[1], compute_graph.flexml_layers[194].compute_node[2][1].in[1], compute_graph.flexml_layers[207].compute_node[2][1].in[1], compute_graph.flexml_layers[211].compute_node[2][1].in[1], compute_graph.flexml_layers[212].compute_node[2][1].in[1], compute_graph.flexml_layers[217].compute_node[2][1].in[1], compute_graph.flexml_layers[221].compute_node[2][1].in[1], compute_graph.flexml_layers[222].compute_node[2][1].in[1], compute_graph.flexml_layers[227].compute_node[2][1].in[1], compute_graph.flexml_layers[231].compute_node[2][1].in[1], compute_graph.flexml_layers[232].compute_node[2][1].in[1], compute_graph.flexml_layers[233].compute_node[2][1].in[1]) + +Column 2Row 2Channel 0 + +i14_pi0(compute_graph.templated_graph_4.trans_comp_nd[2][2].in[0], compute_graph.flexml_layers[36].compute_node[2][2].in[0], compute_graph.templated_graph_231.mk[2][2].in[0], compute_graph.templated_graph_254.mk[2][2].in[0], compute_graph.templated_graph_258.compute_node[2][2].in[0], compute_graph.templated_graph_259.compute_node[2][2].in[0], compute_graph.templated_graph_260.mk[2][2].in[0], compute_graph.templated_graph_263.compute_node[2][2].in[0], compute_graph.templated_graph_266.compute_node[2][2].in[0], compute_graph.templated_graph_268.mk[2][2].in[0], compute_graph.templated_graph_273.mk[2][2].in[0], compute_graph.templated_graph_274.mk[2][2].in[0], compute_graph.flexml_layers[63].compute_node[2][2].in[0], compute_graph.templated_graph_293.mk[2][2].in[0], compute_graph.templated_graph_298.compute_node[2][2].in[0], compute_graph.templated_graph_300.compute_node[2][2].in[0], compute_graph.flexml_layers[0].compute_node[2][2].in[0], compute_graph.flexml_layers[1].compute_node[2][2].in[0], compute_graph.flexml_layers[1].compute_node[2][2].in[2], compute_graph.flexml_layers[2].compute_node[2][2].in[0], compute_graph.flexml_layers[2].compute_node[2][2].in[2], compute_graph.flexml_layers[3].compute_node[2][2].in[0], compute_graph.flexml_layers[4].compute_node[2][2].in[0], compute_graph.flexml_layers[5].compute_node[2][2].in[0], compute_graph.flexml_layers[6].compute_node[2][2].in[0], compute_graph.flexml_layers[7].compute_node[2][2].in[0], compute_graph.flexml_layers[7].compute_node[2][2].in[2], compute_graph.flexml_layers[8].compute_node[2][2].in[0], compute_graph.flexml_layers[9].compute_node[2][2].in[0], compute_graph.flexml_layers[9].compute_node[2][2].in[2], compute_graph.flexml_layers[10].compute_node[2][2].in[0], compute_graph.flexml_layers[11].compute_node[2][2].in[0], compute_graph.flexml_layers[12].compute_node[2][2].in[0], compute_graph.flexml_layers[13].compute_node[2][2].in[0], compute_graph.flexml_layers[14].compute_node[2][2].in[0], compute_graph.flexml_layers[15].compute_node[2][2].in[0], compute_graph.flexml_layers[15].compute_node[2][2].in[2], compute_graph.flexml_layers[16].compute_node[2][2].in[0], compute_graph.flexml_layers[17].compute_node[2][2].in[0], compute_graph.flexml_layers[18].compute_node[2][2].in[0], compute_graph.flexml_layers[19].compute_node[2][2].in[0], compute_graph.flexml_layers[20].compute_node[2][2].in[0], compute_graph.flexml_layers[21].compute_node[2][2].in[0], compute_graph.flexml_layers[22].compute_node[2][2].in[0], compute_graph.flexml_layers[23].compute_node[2][2].in[0], compute_graph.flexml_layers[24].compute_node[2][2].in[0], compute_graph.flexml_layers[24].compute_node[2][2].in[2], compute_graph.flexml_layers[25].compute_node[2][2].in[0], compute_graph.flexml_layers[26].compute_node[2][2].in[0], compute_graph.flexml_layers[27].compute_node[2][2].in[0], compute_graph.flexml_layers[28].compute_node[2][2].in[0], compute_graph.flexml_layers[29].compute_node[2][2].in[0], compute_graph.flexml_layers[30].compute_node[2][2].in[0], compute_graph.flexml_layers[31].compute_node[2][2].in[0], compute_graph.flexml_layers[32].compute_node[2][2].in[0], compute_graph.flexml_layers[33].compute_node[2][2].in[0], compute_graph.flexml_layers[34].compute_node[2][2].in[0], compute_graph.flexml_layers[34].compute_node[2][2].in[2], compute_graph.flexml_layers[35].compute_node[2][2].in[0], compute_graph.flexml_layers[35].compute_node[2][2].in[2], compute_graph.flexml_layers[143].compute_node[2][2].in[0], compute_graph.flexml_layers[144].compute_node[2][2].in[0], compute_graph.flexml_layers[37].compute_node[2][2].in[0], compute_graph.flexml_layers[38].compute_node[2][2].in[0], compute_graph.flexml_layers[39].compute_node[2][2].in[0], compute_graph.flexml_layers[40].compute_node[2][2].in[0], compute_graph.flexml_layers[41].compute_node[2][2].in[0], compute_graph.flexml_layers[42].compute_node[2][2].in[0], compute_graph.flexml_layers[43].compute_node[2][2].in[0], compute_graph.flexml_layers[44].compute_node[2][2].in[0], compute_graph.flexml_layers[44].compute_node[2][2].in[2], compute_graph.flexml_layers[45].compute_node[2][2].in[0], compute_graph.flexml_layers[45].compute_node[2][2].in[2], compute_graph.flexml_layers[46].compute_node[2][2].in[0], compute_graph.flexml_layers[47].compute_node[2][2].in[0], compute_graph.flexml_layers[48].compute_node[2][2].in[0], compute_graph.flexml_layers[49].compute_node[2][2].in[0], compute_graph.flexml_layers[50].compute_node[2][2].in[0], compute_graph.flexml_layers[50].compute_node[2][2].in[2], compute_graph.flexml_layers[51].compute_node[2][2].in[0], compute_graph.flexml_layers[52].compute_node[2][2].in[0], compute_graph.flexml_layers[53].compute_node[2][2].in[0], compute_graph.flexml_layers[54].compute_node[2][2].in[0], compute_graph.flexml_layers[55].compute_node[2][2].in[0], compute_graph.flexml_layers[55].compute_node[2][2].in[2], compute_graph.flexml_layers[56].compute_node[2][2].in[0], compute_graph.flexml_layers[57].compute_node[2][2].in[0], compute_graph.flexml_layers[58].compute_node[2][2].in[0], compute_graph.flexml_layers[59].compute_node[2][2].in[0], compute_graph.flexml_layers[60].compute_node[2][2].in[0], compute_graph.flexml_layers[61].compute_node[2][2].in[0], compute_graph.flexml_layers[61].compute_node[2][2].in[2], compute_graph.flexml_layers[62].compute_node[2][2].in[0], compute_graph.flexml_layers[200].compute_node[2][2].in[0], compute_graph.flexml_layers[201].compute_node[2][2].in[0], compute_graph.flexml_layers[202].compute_node[2][2].in[0], compute_graph.flexml_layers[64].compute_node[2][2].in[0], compute_graph.flexml_layers[65].compute_node[2][2].in[0], compute_graph.flexml_layers[66].compute_node[2][2].in[0], compute_graph.flexml_layers[66].compute_node[2][2].in[2], compute_graph.flexml_layers[67].compute_node[2][2].in[0], compute_graph.flexml_layers[67].compute_node[2][2].in[2], compute_graph.flexml_layers[68].compute_node[2][2].in[0], compute_graph.flexml_layers[69].compute_node[2][2].in[0], compute_graph.flexml_layers[70].compute_node[2][2].in[0], compute_graph.flexml_layers[71].compute_node[2][2].in[0], compute_graph.flexml_layers[72].compute_node[2][2].in[0], compute_graph.flexml_layers[72].compute_node[2][2].in[2], compute_graph.flexml_layers[73].compute_node[2][2].in[0], compute_graph.flexml_layers[74].compute_node[2][2].in[0], compute_graph.flexml_layers[75].compute_node[2][2].in[0], compute_graph.flexml_layers[76].compute_node[2][2].in[0], compute_graph.flexml_layers[77].compute_node[2][2].in[0], compute_graph.flexml_layers[77].compute_node[2][2].in[2], compute_graph.flexml_layers[78].compute_node[2][2].in[0], compute_graph.flexml_layers[78].compute_node[2][2].in[2], compute_graph.flexml_layers[79].compute_node[2][2].in[0], compute_graph.flexml_layers[80].compute_node[2][2].in[0], compute_graph.flexml_layers[81].compute_node[2][2].in[0], compute_graph.flexml_layers[82].compute_node[2][2].in[0], compute_graph.flexml_layers[83].compute_node[2][2].in[0], compute_graph.flexml_layers[83].compute_node[2][2].in[2], compute_graph.flexml_layers[84].compute_node[2][2].in[0], compute_graph.flexml_layers[85].compute_node[2][2].in[0], compute_graph.flexml_layers[86].compute_node[2][2].in[0], compute_graph.flexml_layers[87].compute_node[2][2].in[0], compute_graph.flexml_layers[88].compute_node[2][2].in[0], compute_graph.flexml_layers[88].compute_node[2][2].in[2], compute_graph.flexml_layers[89].compute_node[2][2].in[0], compute_graph.flexml_layers[89].compute_node[2][2].in[2], compute_graph.flexml_layers[90].compute_node[2][2].in[0], compute_graph.flexml_layers[91].compute_node[2][2].in[0], compute_graph.flexml_layers[92].compute_node[2][2].in[0], compute_graph.flexml_layers[93].compute_node[2][2].in[0], compute_graph.flexml_layers[94].compute_node[2][2].in[0], compute_graph.flexml_layers[94].compute_node[2][2].in[2], compute_graph.flexml_layers[95].compute_node[2][2].in[0], compute_graph.flexml_layers[96].compute_node[2][2].in[0], compute_graph.flexml_layers[97].compute_node[2][2].in[0], compute_graph.flexml_layers[98].compute_node[2][2].in[0], compute_graph.flexml_layers[99].compute_node[2][2].in[0], compute_graph.flexml_layers[99].compute_node[2][2].in[2], compute_graph.flexml_layers[100].compute_node[2][2].in[0], compute_graph.flexml_layers[101].compute_node[2][2].in[0], compute_graph.flexml_layers[102].compute_node[2][2].in[0], compute_graph.flexml_layers[103].compute_node[2][2].in[0], compute_graph.flexml_layers[104].compute_node[2][2].in[0], compute_graph.flexml_layers[105].compute_node[2][2].in[0], compute_graph.flexml_layers[106].compute_node[2][2].in[0], compute_graph.flexml_layers[106].compute_node[2][2].in[2], compute_graph.flexml_layers[107].compute_node[2][2].in[0], compute_graph.flexml_layers[108].compute_node[2][2].in[0], compute_graph.flexml_layers[109].compute_node[2][2].in[0], compute_graph.flexml_layers[110].compute_node[2][2].in[0], compute_graph.flexml_layers[111].compute_node[2][2].in[0], compute_graph.flexml_layers[112].compute_node[2][2].in[0], compute_graph.flexml_layers[112].compute_node[2][2].in[2], compute_graph.flexml_layers[113].compute_node[2][2].in[0], compute_graph.flexml_layers[114].compute_node[2][2].in[0], compute_graph.flexml_layers[115].compute_node[2][2].in[0], compute_graph.flexml_layers[116].compute_node[2][2].in[0], compute_graph.flexml_layers[117].compute_node[2][2].in[0], compute_graph.flexml_layers[117].compute_node[2][2].in[2], compute_graph.flexml_layers[118].compute_node[2][2].in[0], compute_graph.flexml_layers[119].compute_node[2][2].in[0], compute_graph.flexml_layers[120].compute_node[2][2].in[0], compute_graph.flexml_layers[121].compute_node[2][2].in[0], compute_graph.flexml_layers[122].compute_node[2][2].in[0], compute_graph.flexml_layers[123].compute_node[2][2].in[0], compute_graph.flexml_layers[124].compute_node[2][2].in[0], compute_graph.flexml_layers[124].compute_node[2][2].in[2], compute_graph.flexml_layers[125].compute_node[2][2].in[0], compute_graph.flexml_layers[125].compute_node[2][2].in[2], compute_graph.flexml_layers[126].compute_node[2][2].in[0], compute_graph.flexml_layers[127].compute_node[2][2].in[0], compute_graph.flexml_layers[128].compute_node[2][2].in[0], compute_graph.flexml_layers[129].compute_node[2][2].in[0], compute_graph.flexml_layers[130].compute_node[2][2].in[0], compute_graph.flexml_layers[130].compute_node[2][2].in[2], compute_graph.flexml_layers[131].compute_node[2][2].in[0], compute_graph.flexml_layers[132].compute_node[2][2].in[0], compute_graph.flexml_layers[133].compute_node[2][2].in[0], compute_graph.flexml_layers[134].compute_node[2][2].in[0], compute_graph.flexml_layers[135].compute_node[2][2].in[0], compute_graph.flexml_layers[135].compute_node[2][2].in[2], compute_graph.flexml_layers[136].compute_node[2][2].in[0], compute_graph.flexml_layers[137].compute_node[2][2].in[0], compute_graph.flexml_layers[138].compute_node[2][2].in[0], compute_graph.flexml_layers[139].compute_node[2][2].in[0], compute_graph.flexml_layers[140].compute_node[2][2].in[0], compute_graph.flexml_layers[141].compute_node[2][2].in[0], compute_graph.flexml_layers[142].compute_node[2][2].in[0], compute_graph.flexml_layers[142].compute_node[2][2].in[2], compute_graph.flexml_layers[145].compute_node[2][2].in[0], compute_graph.flexml_layers[146].compute_node[2][2].in[0], compute_graph.flexml_layers[147].compute_node[2][2].in[0], compute_graph.flexml_layers[148].compute_node[2][2].in[0], compute_graph.flexml_layers[148].compute_node[2][2].in[2], compute_graph.flexml_layers[149].compute_node[2][2].in[0], compute_graph.flexml_layers[150].compute_node[2][2].in[0], compute_graph.flexml_layers[151].compute_node[2][2].in[0], compute_graph.flexml_layers[152].compute_node[2][2].in[0], compute_graph.flexml_layers[153].compute_node[2][2].in[0], compute_graph.flexml_layers[153].compute_node[2][2].in[2], compute_graph.flexml_layers[154].compute_node[2][2].in[0], compute_graph.flexml_layers[155].compute_node[2][2].in[0], compute_graph.flexml_layers[156].compute_node[2][2].in[0], compute_graph.flexml_layers[157].compute_node[2][2].in[0], compute_graph.flexml_layers[158].compute_node[2][2].in[0], compute_graph.flexml_layers[159].compute_node[2][2].in[0], compute_graph.flexml_layers[160].compute_node[2][2].in[0], compute_graph.flexml_layers[160].compute_node[2][2].in[2], compute_graph.flexml_layers[161].compute_node[2][2].in[0], compute_graph.flexml_layers[161].compute_node[2][2].in[2], compute_graph.flexml_layers[162].compute_node[2][2].in[0], compute_graph.flexml_layers[163].compute_node[2][2].in[0], compute_graph.flexml_layers[164].compute_node[2][2].in[0], compute_graph.flexml_layers[165].compute_node[2][2].in[0], compute_graph.flexml_layers[166].compute_node[2][2].in[0], compute_graph.flexml_layers[166].compute_node[2][2].in[2], compute_graph.flexml_layers[167].compute_node[2][2].in[0], compute_graph.flexml_layers[168].compute_node[2][2].in[0], compute_graph.flexml_layers[169].compute_node[2][2].in[0], compute_graph.flexml_layers[170].compute_node[2][2].in[0], compute_graph.flexml_layers[171].compute_node[2][2].in[0], compute_graph.flexml_layers[171].compute_node[2][2].in[2], compute_graph.flexml_layers[172].compute_node[2][2].in[0], compute_graph.flexml_layers[173].compute_node[2][2].in[0], compute_graph.flexml_layers[174].compute_node[2][2].in[0], compute_graph.flexml_layers[175].compute_node[2][2].in[0], compute_graph.flexml_layers[176].compute_node[2][2].in[0], compute_graph.flexml_layers[177].compute_node[2][2].in[0], compute_graph.flexml_layers[178].compute_node[2][2].in[0], compute_graph.flexml_layers[178].compute_node[2][2].in[2], compute_graph.flexml_layers[179].compute_node[2][2].in[0], compute_graph.flexml_layers[179].compute_node[2][2].in[2], compute_graph.flexml_layers[180].compute_node[2][2].in[0], compute_graph.flexml_layers[181].compute_node[2][2].in[0], compute_graph.flexml_layers[182].compute_node[2][2].in[0], compute_graph.flexml_layers[183].compute_node[2][2].in[0], compute_graph.flexml_layers[184].compute_node[2][2].in[0], compute_graph.flexml_layers[184].compute_node[2][2].in[2], compute_graph.flexml_layers[185].compute_node[2][2].in[0], compute_graph.flexml_layers[186].compute_node[2][2].in[0], compute_graph.flexml_layers[187].compute_node[2][2].in[0], compute_graph.flexml_layers[188].compute_node[2][2].in[0], compute_graph.flexml_layers[188].compute_node[2][2].in[2], compute_graph.flexml_layers[189].compute_node[2][2].in[0], compute_graph.flexml_layers[190].compute_node[2][2].in[0], compute_graph.flexml_layers[191].compute_node[2][2].in[0], compute_graph.flexml_layers[191].compute_node[2][2].in[2], compute_graph.flexml_layers[192].compute_node[2][2].in[0], compute_graph.flexml_layers[192].compute_node[2][2].in[2], compute_graph.flexml_layers[193].compute_node[2][2].in[0], compute_graph.flexml_layers[193].compute_node[2][2].in[2], compute_graph.flexml_layers[194].compute_node[2][2].in[0], compute_graph.flexml_layers[195].compute_node[2][2].in[0], compute_graph.flexml_layers[196].compute_node[2][2].in[0], compute_graph.flexml_layers[196].compute_node[2][2].in[2], compute_graph.flexml_layers[197].compute_node[2][2].in[0], compute_graph.flexml_layers[197].compute_node[2][2].in[2], compute_graph.flexml_layers[198].compute_node[2][2].in[0], compute_graph.flexml_layers[199].compute_node[2][2].in[0], compute_graph.flexml_layers[203].compute_node[2][2].in[0], compute_graph.flexml_layers[204].compute_node[2][2].in[0], compute_graph.flexml_layers[204].compute_node[2][2].in[2], compute_graph.flexml_layers[205].compute_node[2][2].in[0], compute_graph.flexml_layers[205].compute_node[2][2].in[2], compute_graph.flexml_layers[206].compute_node[2][2].in[0], compute_graph.flexml_layers[206].compute_node[2][2].in[2], compute_graph.flexml_layers[207].compute_node[2][2].in[0], compute_graph.flexml_layers[208].compute_node[2][2].in[0], compute_graph.flexml_layers[209].compute_node[2][2].in[0], compute_graph.flexml_layers[209].compute_node[2][2].in[2], compute_graph.flexml_layers[210].compute_node[2][2].in[0], compute_graph.flexml_layers[210].compute_node[2][2].in[2], compute_graph.flexml_layers[211].compute_node[2][2].in[0], compute_graph.flexml_layers[212].compute_node[2][2].in[0], compute_graph.flexml_layers[213].compute_node[2][2].in[0], compute_graph.flexml_layers[214].compute_node[2][2].in[0], compute_graph.flexml_layers[214].compute_node[2][2].in[2], compute_graph.flexml_layers[215].compute_node[2][2].in[0], compute_graph.flexml_layers[215].compute_node[2][2].in[2], compute_graph.flexml_layers[216].compute_node[2][2].in[0], compute_graph.flexml_layers[216].compute_node[2][2].in[2], compute_graph.flexml_layers[217].compute_node[2][2].in[0], compute_graph.flexml_layers[218].compute_node[2][2].in[0], compute_graph.flexml_layers[219].compute_node[2][2].in[0], compute_graph.flexml_layers[219].compute_node[2][2].in[2], compute_graph.flexml_layers[220].compute_node[2][2].in[0], compute_graph.flexml_layers[220].compute_node[2][2].in[2], compute_graph.flexml_layers[221].compute_node[2][2].in[0], compute_graph.flexml_layers[222].compute_node[2][2].in[0], compute_graph.flexml_layers[223].compute_node[2][2].in[0], compute_graph.flexml_layers[224].compute_node[2][2].in[0], compute_graph.flexml_layers[224].compute_node[2][2].in[2], compute_graph.flexml_layers[225].compute_node[2][2].in[0], compute_graph.flexml_layers[225].compute_node[2][2].in[2], compute_graph.flexml_layers[226].compute_node[2][2].in[0], compute_graph.flexml_layers[226].compute_node[2][2].in[2], compute_graph.flexml_layers[227].compute_node[2][2].in[0], compute_graph.flexml_layers[228].compute_node[2][2].in[0], compute_graph.flexml_layers[229].compute_node[2][2].in[0], compute_graph.flexml_layers[229].compute_node[2][2].in[2], compute_graph.flexml_layers[230].compute_node[2][2].in[0], compute_graph.flexml_layers[230].compute_node[2][2].in[2], compute_graph.flexml_layers[231].compute_node[2][2].in[0], compute_graph.flexml_layers[232].compute_node[2][2].in[0], compute_graph.flexml_layers[233].compute_node[2][2].in[0], compute_graph.flexml_layers[234].compute_node[2][2].in[0], compute_graph.flexml_layers[235].compute_node[2][2].in[0], compute_graph.flexml_layers[235].compute_node[2][2].in[2], compute_graph.flexml_layers[236].compute_node[2][2].in[0]) + +Column 2Row 2Channel 0 + +i14_po0(compute_graph.templated_graph_4.trans_comp_nd[2][2].out[0], compute_graph.flexml_layers[36].compute_node[2][2].out[0], compute_graph.templated_graph_231.mk[2][2].out[0], compute_graph.templated_graph_254.mk[2][2].out[0], compute_graph.templated_graph_258.compute_node[2][2].out[0], compute_graph.templated_graph_259.compute_node[2][2].out[0], compute_graph.templated_graph_260.mk[2][2].out[0], compute_graph.templated_graph_263.compute_node[2][2].out[0], compute_graph.templated_graph_266.compute_node[2][2].out[0], compute_graph.templated_graph_268.mk[2][2].out[0], compute_graph.templated_graph_273.mk[2][2].out[0], compute_graph.templated_graph_274.mk[2][2].out[0], compute_graph.flexml_layers[63].compute_node[2][2].out[0], compute_graph.templated_graph_293.mk[2][2].out[0], compute_graph.templated_graph_298.compute_node[2][2].out[0], compute_graph.templated_graph_300.compute_node[2][2].out[0], compute_graph.flexml_layers[0].compute_node[2][2].out[0], compute_graph.flexml_layers[1].compute_node[2][2].out[0], compute_graph.flexml_layers[2].compute_node[2][2].out[0], compute_graph.flexml_layers[3].compute_node[2][2].out[0], compute_graph.flexml_layers[4].compute_node[2][2].out[0], compute_graph.flexml_layers[5].compute_node[2][2].out[0], compute_graph.flexml_layers[6].compute_node[2][2].out[0], compute_graph.flexml_layers[7].compute_node[2][2].out[0], compute_graph.flexml_layers[8].compute_node[2][2].out[0], compute_graph.flexml_layers[9].compute_node[2][2].out[0], compute_graph.flexml_layers[10].compute_node[2][2].out[0], compute_graph.flexml_layers[11].compute_node[2][2].out[0], compute_graph.flexml_layers[12].compute_node[2][2].out[0], compute_graph.flexml_layers[13].compute_node[2][2].out[0], compute_graph.flexml_layers[14].compute_node[2][2].out[0], compute_graph.flexml_layers[15].compute_node[2][2].out[0], compute_graph.flexml_layers[16].compute_node[2][2].out[0], compute_graph.flexml_layers[17].compute_node[2][2].out[0], compute_graph.flexml_layers[18].compute_node[2][2].out[0], compute_graph.flexml_layers[19].compute_node[2][2].out[0], compute_graph.flexml_layers[20].compute_node[2][2].out[0], compute_graph.flexml_layers[21].compute_node[2][2].out[0], compute_graph.flexml_layers[22].compute_node[2][2].out[0], compute_graph.flexml_layers[23].compute_node[2][2].out[0], compute_graph.flexml_layers[24].compute_node[2][2].out[0], compute_graph.flexml_layers[25].compute_node[2][2].out[0], compute_graph.flexml_layers[26].compute_node[2][2].out[0], compute_graph.flexml_layers[27].compute_node[2][2].out[0], compute_graph.flexml_layers[28].compute_node[2][2].out[0], compute_graph.flexml_layers[29].compute_node[2][2].out[0], compute_graph.flexml_layers[30].compute_node[2][2].out[0], compute_graph.flexml_layers[31].compute_node[2][2].out[0], compute_graph.flexml_layers[32].compute_node[2][2].out[0], compute_graph.flexml_layers[33].compute_node[2][2].out[0], compute_graph.flexml_layers[34].compute_node[2][2].out[0], compute_graph.flexml_layers[35].compute_node[2][2].out[0], compute_graph.flexml_layers[143].compute_node[2][2].out[0], compute_graph.flexml_layers[144].compute_node[2][2].out[0], compute_graph.flexml_layers[37].compute_node[2][2].out[0], compute_graph.flexml_layers[38].compute_node[2][2].out[0], compute_graph.flexml_layers[39].compute_node[2][2].out[0], compute_graph.flexml_layers[40].compute_node[2][2].out[0], compute_graph.flexml_layers[41].compute_node[2][2].out[0], compute_graph.flexml_layers[42].compute_node[2][2].out[0], compute_graph.flexml_layers[43].compute_node[2][2].out[0], compute_graph.flexml_layers[44].compute_node[2][2].out[0], compute_graph.flexml_layers[45].compute_node[2][2].out[0], compute_graph.flexml_layers[46].compute_node[2][2].out[0], compute_graph.flexml_layers[47].compute_node[2][2].out[0], compute_graph.flexml_layers[48].compute_node[2][2].out[0], compute_graph.flexml_layers[49].compute_node[2][2].out[0], compute_graph.flexml_layers[50].compute_node[2][2].out[0], compute_graph.flexml_layers[51].compute_node[2][2].out[0], compute_graph.flexml_layers[52].compute_node[2][2].out[0], compute_graph.flexml_layers[53].compute_node[2][2].out[0], compute_graph.flexml_layers[54].compute_node[2][2].out[0], compute_graph.flexml_layers[55].compute_node[2][2].out[0], compute_graph.flexml_layers[56].compute_node[2][2].out[0], compute_graph.flexml_layers[57].compute_node[2][2].out[0], compute_graph.flexml_layers[58].compute_node[2][2].out[0], compute_graph.flexml_layers[59].compute_node[2][2].out[0], compute_graph.flexml_layers[60].compute_node[2][2].out[0], compute_graph.flexml_layers[61].compute_node[2][2].out[0], compute_graph.flexml_layers[62].compute_node[2][2].out[0], compute_graph.flexml_layers[200].compute_node[2][2].out[0], compute_graph.flexml_layers[201].compute_node[2][2].out[0], compute_graph.flexml_layers[202].compute_node[2][2].out[0], compute_graph.flexml_layers[64].compute_node[2][2].out[0], compute_graph.flexml_layers[65].compute_node[2][2].out[0], compute_graph.flexml_layers[66].compute_node[2][2].out[0], compute_graph.flexml_layers[67].compute_node[2][2].out[0], compute_graph.flexml_layers[68].compute_node[2][2].out[0], compute_graph.flexml_layers[69].compute_node[2][2].out[0], compute_graph.flexml_layers[70].compute_node[2][2].out[0], compute_graph.flexml_layers[71].compute_node[2][2].out[0], compute_graph.flexml_layers[72].compute_node[2][2].out[0], compute_graph.flexml_layers[73].compute_node[2][2].out[0], compute_graph.flexml_layers[74].compute_node[2][2].out[0], compute_graph.flexml_layers[75].compute_node[2][2].out[0], compute_graph.flexml_layers[76].compute_node[2][2].out[0], compute_graph.flexml_layers[77].compute_node[2][2].out[0], compute_graph.flexml_layers[78].compute_node[2][2].out[0], compute_graph.flexml_layers[79].compute_node[2][2].out[0], compute_graph.flexml_layers[80].compute_node[2][2].out[0], compute_graph.flexml_layers[81].compute_node[2][2].out[0], compute_graph.flexml_layers[82].compute_node[2][2].out[0], compute_graph.flexml_layers[83].compute_node[2][2].out[0], compute_graph.flexml_layers[84].compute_node[2][2].out[0], compute_graph.flexml_layers[85].compute_node[2][2].out[0], compute_graph.flexml_layers[86].compute_node[2][2].out[0], compute_graph.flexml_layers[87].compute_node[2][2].out[0], compute_graph.flexml_layers[88].compute_node[2][2].out[0], compute_graph.flexml_layers[89].compute_node[2][2].out[0], compute_graph.flexml_layers[90].compute_node[2][2].out[0], compute_graph.flexml_layers[91].compute_node[2][2].out[0], compute_graph.flexml_layers[92].compute_node[2][2].out[0], compute_graph.flexml_layers[93].compute_node[2][2].out[0], compute_graph.flexml_layers[94].compute_node[2][2].out[0], compute_graph.flexml_layers[95].compute_node[2][2].out[0], compute_graph.flexml_layers[96].compute_node[2][2].out[0], compute_graph.flexml_layers[97].compute_node[2][2].out[0], compute_graph.flexml_layers[98].compute_node[2][2].out[0], compute_graph.flexml_layers[99].compute_node[2][2].out[0], compute_graph.flexml_layers[100].compute_node[2][2].out[0], compute_graph.flexml_layers[101].compute_node[2][2].out[0], compute_graph.flexml_layers[102].compute_node[2][2].out[0], compute_graph.flexml_layers[103].compute_node[2][2].out[0], compute_graph.flexml_layers[104].compute_node[2][2].out[0], compute_graph.flexml_layers[105].compute_node[2][2].out[0], compute_graph.flexml_layers[106].compute_node[2][2].out[0], compute_graph.flexml_layers[107].compute_node[2][2].out[0], compute_graph.flexml_layers[108].compute_node[2][2].out[0], compute_graph.flexml_layers[109].compute_node[2][2].out[0], compute_graph.flexml_layers[110].compute_node[2][2].out[0], compute_graph.flexml_layers[111].compute_node[2][2].out[0], compute_graph.flexml_layers[112].compute_node[2][2].out[0], compute_graph.flexml_layers[113].compute_node[2][2].out[0], compute_graph.flexml_layers[114].compute_node[2][2].out[0], compute_graph.flexml_layers[115].compute_node[2][2].out[0], compute_graph.flexml_layers[116].compute_node[2][2].out[0], compute_graph.flexml_layers[117].compute_node[2][2].out[0], compute_graph.flexml_layers[118].compute_node[2][2].out[0], compute_graph.flexml_layers[119].compute_node[2][2].out[0], compute_graph.flexml_layers[120].compute_node[2][2].out[0], compute_graph.flexml_layers[121].compute_node[2][2].out[0], compute_graph.flexml_layers[122].compute_node[2][2].out[0], compute_graph.flexml_layers[123].compute_node[2][2].out[0], compute_graph.flexml_layers[124].compute_node[2][2].out[0], compute_graph.flexml_layers[125].compute_node[2][2].out[0], compute_graph.flexml_layers[126].compute_node[2][2].out[0], compute_graph.flexml_layers[127].compute_node[2][2].out[0], compute_graph.flexml_layers[128].compute_node[2][2].out[0], compute_graph.flexml_layers[129].compute_node[2][2].out[0], compute_graph.flexml_layers[130].compute_node[2][2].out[0], compute_graph.flexml_layers[131].compute_node[2][2].out[0], compute_graph.flexml_layers[132].compute_node[2][2].out[0], compute_graph.flexml_layers[133].compute_node[2][2].out[0], compute_graph.flexml_layers[134].compute_node[2][2].out[0], compute_graph.flexml_layers[135].compute_node[2][2].out[0], compute_graph.flexml_layers[136].compute_node[2][2].out[0], compute_graph.flexml_layers[137].compute_node[2][2].out[0], compute_graph.flexml_layers[138].compute_node[2][2].out[0], compute_graph.flexml_layers[139].compute_node[2][2].out[0], compute_graph.flexml_layers[140].compute_node[2][2].out[0], compute_graph.flexml_layers[141].compute_node[2][2].out[0], compute_graph.flexml_layers[142].compute_node[2][2].out[0], compute_graph.flexml_layers[145].compute_node[2][2].out[0], compute_graph.flexml_layers[146].compute_node[2][2].out[0], compute_graph.flexml_layers[147].compute_node[2][2].out[0], compute_graph.flexml_layers[148].compute_node[2][2].out[0], compute_graph.flexml_layers[149].compute_node[2][2].out[0], compute_graph.flexml_layers[150].compute_node[2][2].out[0], compute_graph.flexml_layers[151].compute_node[2][2].out[0], compute_graph.flexml_layers[152].compute_node[2][2].out[0], compute_graph.flexml_layers[153].compute_node[2][2].out[0], compute_graph.flexml_layers[154].compute_node[2][2].out[0], compute_graph.flexml_layers[155].compute_node[2][2].out[0], compute_graph.flexml_layers[156].compute_node[2][2].out[0], compute_graph.flexml_layers[157].compute_node[2][2].out[0], compute_graph.flexml_layers[158].compute_node[2][2].out[0], compute_graph.flexml_layers[159].compute_node[2][2].out[0], compute_graph.flexml_layers[160].compute_node[2][2].out[0], compute_graph.flexml_layers[161].compute_node[2][2].out[0], compute_graph.flexml_layers[162].compute_node[2][2].out[0], compute_graph.flexml_layers[163].compute_node[2][2].out[0], compute_graph.flexml_layers[164].compute_node[2][2].out[0], compute_graph.flexml_layers[165].compute_node[2][2].out[0], compute_graph.flexml_layers[166].compute_node[2][2].out[0], compute_graph.flexml_layers[167].compute_node[2][2].out[0], compute_graph.flexml_layers[168].compute_node[2][2].out[0], compute_graph.flexml_layers[169].compute_node[2][2].out[0], compute_graph.flexml_layers[170].compute_node[2][2].out[0], compute_graph.flexml_layers[171].compute_node[2][2].out[0], compute_graph.flexml_layers[172].compute_node[2][2].out[0], compute_graph.flexml_layers[173].compute_node[2][2].out[0], compute_graph.flexml_layers[174].compute_node[2][2].out[0], compute_graph.flexml_layers[175].compute_node[2][2].out[0], compute_graph.flexml_layers[176].compute_node[2][2].out[0], compute_graph.flexml_layers[177].compute_node[2][2].out[0], compute_graph.flexml_layers[178].compute_node[2][2].out[0], compute_graph.flexml_layers[179].compute_node[2][2].out[0], compute_graph.flexml_layers[180].compute_node[2][2].out[0], compute_graph.flexml_layers[181].compute_node[2][2].out[0], compute_graph.flexml_layers[182].compute_node[2][2].out[0], compute_graph.flexml_layers[183].compute_node[2][2].out[0], compute_graph.flexml_layers[184].compute_node[2][2].out[0], compute_graph.flexml_layers[185].compute_node[2][2].out[0], compute_graph.flexml_layers[186].compute_node[2][2].out[0], compute_graph.flexml_layers[187].compute_node[2][2].out[0], compute_graph.flexml_layers[188].compute_node[2][2].out[0], compute_graph.flexml_layers[189].compute_node[2][2].out[0], compute_graph.flexml_layers[190].compute_node[2][2].out[0], compute_graph.flexml_layers[191].compute_node[2][2].out[0], compute_graph.flexml_layers[192].compute_node[2][2].out[0], compute_graph.flexml_layers[193].compute_node[2][2].out[0], compute_graph.flexml_layers[194].compute_node[2][2].out[0], compute_graph.flexml_layers[195].compute_node[2][2].out[0], compute_graph.flexml_layers[196].compute_node[2][2].out[0], compute_graph.flexml_layers[197].compute_node[2][2].out[0], compute_graph.flexml_layers[198].compute_node[2][2].out[0], compute_graph.flexml_layers[199].compute_node[2][2].out[0], compute_graph.flexml_layers[203].compute_node[2][2].out[0], compute_graph.flexml_layers[204].compute_node[2][2].out[0], compute_graph.flexml_layers[205].compute_node[2][2].out[0], compute_graph.flexml_layers[206].compute_node[2][2].out[0], compute_graph.flexml_layers[207].compute_node[2][2].out[0], compute_graph.flexml_layers[208].compute_node[2][2].out[0], compute_graph.flexml_layers[209].compute_node[2][2].out[0], compute_graph.flexml_layers[210].compute_node[2][2].out[0], compute_graph.flexml_layers[211].compute_node[2][2].out[0], compute_graph.flexml_layers[212].compute_node[2][2].out[0], compute_graph.flexml_layers[213].compute_node[2][2].out[0], compute_graph.flexml_layers[214].compute_node[2][2].out[0], compute_graph.flexml_layers[215].compute_node[2][2].out[0], compute_graph.flexml_layers[216].compute_node[2][2].out[0], compute_graph.flexml_layers[217].compute_node[2][2].out[0], compute_graph.flexml_layers[218].compute_node[2][2].out[0], compute_graph.flexml_layers[219].compute_node[2][2].out[0], compute_graph.flexml_layers[220].compute_node[2][2].out[0], compute_graph.flexml_layers[221].compute_node[2][2].out[0], compute_graph.flexml_layers[222].compute_node[2][2].out[0], compute_graph.flexml_layers[223].compute_node[2][2].out[0], compute_graph.flexml_layers[224].compute_node[2][2].out[0], compute_graph.flexml_layers[225].compute_node[2][2].out[0], compute_graph.flexml_layers[226].compute_node[2][2].out[0], compute_graph.flexml_layers[227].compute_node[2][2].out[0], compute_graph.flexml_layers[228].compute_node[2][2].out[0], compute_graph.flexml_layers[229].compute_node[2][2].out[0], compute_graph.flexml_layers[230].compute_node[2][2].out[0], compute_graph.flexml_layers[231].compute_node[2][2].out[0], compute_graph.flexml_layers[232].compute_node[2][2].out[0], compute_graph.flexml_layers[233].compute_node[2][2].out[0], compute_graph.flexml_layers[234].compute_node[2][2].out[0], compute_graph.flexml_layers[235].compute_node[2][2].out[0], compute_graph.flexml_layers[236].compute_node[2][2].out[0]) + +Column 2Row 2Channel 1 + +i14_pi1(compute_graph.flexml_layers[36].compute_node[2][2].in[1], compute_graph.templated_graph_260.mk[2][2].in[1], compute_graph.templated_graph_268.mk[2][2].in[1], compute_graph.templated_graph_273.mk[2][2].in[1], compute_graph.flexml_layers[3].compute_node[2][2].in[1], compute_graph.flexml_layers[8].compute_node[2][2].in[1], compute_graph.flexml_layers[9].compute_node[2][2].in[1], compute_graph.flexml_layers[10].compute_node[2][2].in[1], compute_graph.flexml_layers[11].compute_node[2][2].in[1], compute_graph.flexml_layers[12].compute_node[2][2].in[1], compute_graph.flexml_layers[13].compute_node[2][2].in[1], compute_graph.flexml_layers[14].compute_node[2][2].in[1], compute_graph.flexml_layers[15].compute_node[2][2].in[1], compute_graph.flexml_layers[16].compute_node[2][2].in[1], compute_graph.flexml_layers[17].compute_node[2][2].in[1], compute_graph.flexml_layers[19].compute_node[2][2].in[1], compute_graph.flexml_layers[20].compute_node[2][2].in[1], compute_graph.flexml_layers[25].compute_node[2][2].in[1], compute_graph.flexml_layers[26].compute_node[2][2].in[1], compute_graph.flexml_layers[27].compute_node[2][2].in[1], compute_graph.flexml_layers[29].compute_node[2][2].in[1], compute_graph.flexml_layers[30].compute_node[2][2].in[1], compute_graph.flexml_layers[35].compute_node[2][2].in[1], compute_graph.flexml_layers[143].compute_node[2][2].in[1], compute_graph.flexml_layers[144].compute_node[2][2].in[1], compute_graph.flexml_layers[37].compute_node[2][2].in[1], compute_graph.flexml_layers[39].compute_node[2][2].in[1], compute_graph.flexml_layers[40].compute_node[2][2].in[1], compute_graph.flexml_layers[45].compute_node[2][2].in[1], compute_graph.flexml_layers[46].compute_node[2][2].in[1], compute_graph.flexml_layers[51].compute_node[2][2].in[1], compute_graph.flexml_layers[56].compute_node[2][2].in[1], compute_graph.flexml_layers[57].compute_node[2][2].in[1], compute_graph.flexml_layers[62].compute_node[2][2].in[1], compute_graph.flexml_layers[201].compute_node[2][2].in[1], compute_graph.flexml_layers[202].compute_node[2][2].in[1], compute_graph.flexml_layers[67].compute_node[2][2].in[1], compute_graph.flexml_layers[68].compute_node[2][2].in[1], compute_graph.flexml_layers[73].compute_node[2][2].in[1], compute_graph.flexml_layers[78].compute_node[2][2].in[1], compute_graph.flexml_layers[79].compute_node[2][2].in[1], compute_graph.flexml_layers[84].compute_node[2][2].in[1], compute_graph.flexml_layers[89].compute_node[2][2].in[1], compute_graph.flexml_layers[90].compute_node[2][2].in[1], compute_graph.flexml_layers[95].compute_node[2][2].in[1], compute_graph.flexml_layers[101].compute_node[2][2].in[1], compute_graph.flexml_layers[102].compute_node[2][2].in[1], compute_graph.flexml_layers[107].compute_node[2][2].in[1], compute_graph.flexml_layers[108].compute_node[2][2].in[1], compute_graph.flexml_layers[113].compute_node[2][2].in[1], compute_graph.flexml_layers[119].compute_node[2][2].in[1], compute_graph.flexml_layers[120].compute_node[2][2].in[1], compute_graph.flexml_layers[125].compute_node[2][2].in[1], compute_graph.flexml_layers[126].compute_node[2][2].in[1], compute_graph.flexml_layers[131].compute_node[2][2].in[1], compute_graph.flexml_layers[137].compute_node[2][2].in[1], compute_graph.flexml_layers[138].compute_node[2][2].in[1], compute_graph.flexml_layers[149].compute_node[2][2].in[1], compute_graph.flexml_layers[155].compute_node[2][2].in[1], compute_graph.flexml_layers[156].compute_node[2][2].in[1], compute_graph.flexml_layers[161].compute_node[2][2].in[1], compute_graph.flexml_layers[162].compute_node[2][2].in[1], compute_graph.flexml_layers[167].compute_node[2][2].in[1], compute_graph.flexml_layers[173].compute_node[2][2].in[1], compute_graph.flexml_layers[174].compute_node[2][2].in[1], compute_graph.flexml_layers[179].compute_node[2][2].in[1], compute_graph.flexml_layers[180].compute_node[2][2].in[1], compute_graph.flexml_layers[186].compute_node[2][2].in[1], compute_graph.flexml_layers[188].compute_node[2][2].in[1], compute_graph.flexml_layers[189].compute_node[2][2].in[1], compute_graph.flexml_layers[194].compute_node[2][2].in[1], compute_graph.flexml_layers[207].compute_node[2][2].in[1], compute_graph.flexml_layers[211].compute_node[2][2].in[1], compute_graph.flexml_layers[212].compute_node[2][2].in[1], compute_graph.flexml_layers[217].compute_node[2][2].in[1], compute_graph.flexml_layers[221].compute_node[2][2].in[1], compute_graph.flexml_layers[222].compute_node[2][2].in[1], compute_graph.flexml_layers[227].compute_node[2][2].in[1], compute_graph.flexml_layers[231].compute_node[2][2].in[1], compute_graph.flexml_layers[232].compute_node[2][2].in[1], compute_graph.flexml_layers[233].compute_node[2][2].in[1]) + +Column 2Row 3Channel 0 + +i15_pi0(compute_graph.templated_graph_4.trans_comp_nd[2][3].in[0], compute_graph.flexml_layers[36].compute_node[2][3].in[0], compute_graph.templated_graph_231.mk[2][3].in[0], compute_graph.templated_graph_254.mk[2][3].in[0], compute_graph.templated_graph_258.compute_node[2][3].in[0], compute_graph.templated_graph_259.compute_node[2][3].in[0], compute_graph.templated_graph_260.mk[2][3].in[0], compute_graph.templated_graph_263.compute_node[2][3].in[0], compute_graph.templated_graph_266.compute_node[2][3].in[0], compute_graph.templated_graph_268.mk[2][3].in[0], compute_graph.templated_graph_273.mk[2][3].in[0], compute_graph.templated_graph_274.mk[2][3].in[0], compute_graph.flexml_layers[63].compute_node[2][3].in[0], compute_graph.templated_graph_293.mk[2][3].in[0], compute_graph.templated_graph_298.compute_node[2][3].in[0], compute_graph.templated_graph_300.compute_node[2][3].in[0], compute_graph.flexml_layers[0].compute_node[2][3].in[0], compute_graph.flexml_layers[1].compute_node[2][3].in[0], compute_graph.flexml_layers[1].compute_node[2][3].in[2], compute_graph.flexml_layers[2].compute_node[2][3].in[0], compute_graph.flexml_layers[2].compute_node[2][3].in[2], compute_graph.flexml_layers[3].compute_node[2][3].in[0], compute_graph.flexml_layers[4].compute_node[2][3].in[0], compute_graph.flexml_layers[5].compute_node[2][3].in[0], compute_graph.flexml_layers[6].compute_node[2][3].in[0], compute_graph.flexml_layers[7].compute_node[2][3].in[0], compute_graph.flexml_layers[7].compute_node[2][3].in[2], compute_graph.flexml_layers[8].compute_node[2][3].in[0], compute_graph.flexml_layers[9].compute_node[2][3].in[0], compute_graph.flexml_layers[9].compute_node[2][3].in[2], compute_graph.flexml_layers[10].compute_node[2][3].in[0], compute_graph.flexml_layers[11].compute_node[2][3].in[0], compute_graph.flexml_layers[12].compute_node[2][3].in[0], compute_graph.flexml_layers[13].compute_node[2][3].in[0], compute_graph.flexml_layers[14].compute_node[2][3].in[0], compute_graph.flexml_layers[15].compute_node[2][3].in[0], compute_graph.flexml_layers[15].compute_node[2][3].in[2], compute_graph.flexml_layers[16].compute_node[2][3].in[0], compute_graph.flexml_layers[17].compute_node[2][3].in[0], compute_graph.flexml_layers[18].compute_node[2][3].in[0], compute_graph.flexml_layers[19].compute_node[2][3].in[0], compute_graph.flexml_layers[20].compute_node[2][3].in[0], compute_graph.flexml_layers[21].compute_node[2][3].in[0], compute_graph.flexml_layers[22].compute_node[2][3].in[0], compute_graph.flexml_layers[23].compute_node[2][3].in[0], compute_graph.flexml_layers[24].compute_node[2][3].in[0], compute_graph.flexml_layers[24].compute_node[2][3].in[2], compute_graph.flexml_layers[25].compute_node[2][3].in[0], compute_graph.flexml_layers[26].compute_node[2][3].in[0], compute_graph.flexml_layers[27].compute_node[2][3].in[0], compute_graph.flexml_layers[28].compute_node[2][3].in[0], compute_graph.flexml_layers[29].compute_node[2][3].in[0], compute_graph.flexml_layers[30].compute_node[2][3].in[0], compute_graph.flexml_layers[31].compute_node[2][3].in[0], compute_graph.flexml_layers[32].compute_node[2][3].in[0], compute_graph.flexml_layers[33].compute_node[2][3].in[0], compute_graph.flexml_layers[34].compute_node[2][3].in[0], compute_graph.flexml_layers[34].compute_node[2][3].in[2], compute_graph.flexml_layers[35].compute_node[2][3].in[0], compute_graph.flexml_layers[35].compute_node[2][3].in[2], compute_graph.flexml_layers[143].compute_node[2][3].in[0], compute_graph.flexml_layers[144].compute_node[2][3].in[0], compute_graph.flexml_layers[37].compute_node[2][3].in[0], compute_graph.flexml_layers[38].compute_node[2][3].in[0], compute_graph.flexml_layers[39].compute_node[2][3].in[0], compute_graph.flexml_layers[40].compute_node[2][3].in[0], compute_graph.flexml_layers[41].compute_node[2][3].in[0], compute_graph.flexml_layers[42].compute_node[2][3].in[0], compute_graph.flexml_layers[43].compute_node[2][3].in[0], compute_graph.flexml_layers[44].compute_node[2][3].in[0], compute_graph.flexml_layers[44].compute_node[2][3].in[2], compute_graph.flexml_layers[45].compute_node[2][3].in[0], compute_graph.flexml_layers[45].compute_node[2][3].in[2], compute_graph.flexml_layers[46].compute_node[2][3].in[0], compute_graph.flexml_layers[47].compute_node[2][3].in[0], compute_graph.flexml_layers[48].compute_node[2][3].in[0], compute_graph.flexml_layers[49].compute_node[2][3].in[0], compute_graph.flexml_layers[50].compute_node[2][3].in[0], compute_graph.flexml_layers[50].compute_node[2][3].in[2], compute_graph.flexml_layers[51].compute_node[2][3].in[0], compute_graph.flexml_layers[52].compute_node[2][3].in[0], compute_graph.flexml_layers[53].compute_node[2][3].in[0], compute_graph.flexml_layers[54].compute_node[2][3].in[0], compute_graph.flexml_layers[55].compute_node[2][3].in[0], compute_graph.flexml_layers[55].compute_node[2][3].in[2], compute_graph.flexml_layers[56].compute_node[2][3].in[0], compute_graph.flexml_layers[57].compute_node[2][3].in[0], compute_graph.flexml_layers[58].compute_node[2][3].in[0], compute_graph.flexml_layers[59].compute_node[2][3].in[0], compute_graph.flexml_layers[60].compute_node[2][3].in[0], compute_graph.flexml_layers[61].compute_node[2][3].in[0], compute_graph.flexml_layers[61].compute_node[2][3].in[2], compute_graph.flexml_layers[62].compute_node[2][3].in[0], compute_graph.flexml_layers[200].compute_node[2][3].in[0], compute_graph.flexml_layers[201].compute_node[2][3].in[0], compute_graph.flexml_layers[202].compute_node[2][3].in[0], compute_graph.flexml_layers[64].compute_node[2][3].in[0], compute_graph.flexml_layers[65].compute_node[2][3].in[0], compute_graph.flexml_layers[66].compute_node[2][3].in[0], compute_graph.flexml_layers[66].compute_node[2][3].in[2], compute_graph.flexml_layers[67].compute_node[2][3].in[0], compute_graph.flexml_layers[67].compute_node[2][3].in[2], compute_graph.flexml_layers[68].compute_node[2][3].in[0], compute_graph.flexml_layers[69].compute_node[2][3].in[0], compute_graph.flexml_layers[70].compute_node[2][3].in[0], compute_graph.flexml_layers[71].compute_node[2][3].in[0], compute_graph.flexml_layers[72].compute_node[2][3].in[0], compute_graph.flexml_layers[72].compute_node[2][3].in[2], compute_graph.flexml_layers[73].compute_node[2][3].in[0], compute_graph.flexml_layers[74].compute_node[2][3].in[0], compute_graph.flexml_layers[75].compute_node[2][3].in[0], compute_graph.flexml_layers[76].compute_node[2][3].in[0], compute_graph.flexml_layers[77].compute_node[2][3].in[0], compute_graph.flexml_layers[77].compute_node[2][3].in[2], compute_graph.flexml_layers[78].compute_node[2][3].in[0], compute_graph.flexml_layers[78].compute_node[2][3].in[2], compute_graph.flexml_layers[79].compute_node[2][3].in[0], compute_graph.flexml_layers[80].compute_node[2][3].in[0], compute_graph.flexml_layers[81].compute_node[2][3].in[0], compute_graph.flexml_layers[82].compute_node[2][3].in[0], compute_graph.flexml_layers[83].compute_node[2][3].in[0], compute_graph.flexml_layers[83].compute_node[2][3].in[2], compute_graph.flexml_layers[84].compute_node[2][3].in[0], compute_graph.flexml_layers[85].compute_node[2][3].in[0], compute_graph.flexml_layers[86].compute_node[2][3].in[0], compute_graph.flexml_layers[87].compute_node[2][3].in[0], compute_graph.flexml_layers[88].compute_node[2][3].in[0], compute_graph.flexml_layers[88].compute_node[2][3].in[2], compute_graph.flexml_layers[89].compute_node[2][3].in[0], compute_graph.flexml_layers[89].compute_node[2][3].in[2], compute_graph.flexml_layers[90].compute_node[2][3].in[0], compute_graph.flexml_layers[91].compute_node[2][3].in[0], compute_graph.flexml_layers[92].compute_node[2][3].in[0], compute_graph.flexml_layers[93].compute_node[2][3].in[0], compute_graph.flexml_layers[94].compute_node[2][3].in[0], compute_graph.flexml_layers[94].compute_node[2][3].in[2], compute_graph.flexml_layers[95].compute_node[2][3].in[0], compute_graph.flexml_layers[96].compute_node[2][3].in[0], compute_graph.flexml_layers[97].compute_node[2][3].in[0], compute_graph.flexml_layers[98].compute_node[2][3].in[0], compute_graph.flexml_layers[99].compute_node[2][3].in[0], compute_graph.flexml_layers[99].compute_node[2][3].in[2], compute_graph.flexml_layers[100].compute_node[2][3].in[0], compute_graph.flexml_layers[101].compute_node[2][3].in[0], compute_graph.flexml_layers[102].compute_node[2][3].in[0], compute_graph.flexml_layers[103].compute_node[2][3].in[0], compute_graph.flexml_layers[104].compute_node[2][3].in[0], compute_graph.flexml_layers[105].compute_node[2][3].in[0], compute_graph.flexml_layers[106].compute_node[2][3].in[0], compute_graph.flexml_layers[106].compute_node[2][3].in[2], compute_graph.flexml_layers[107].compute_node[2][3].in[0], compute_graph.flexml_layers[108].compute_node[2][3].in[0], compute_graph.flexml_layers[109].compute_node[2][3].in[0], compute_graph.flexml_layers[110].compute_node[2][3].in[0], compute_graph.flexml_layers[111].compute_node[2][3].in[0], compute_graph.flexml_layers[112].compute_node[2][3].in[0], compute_graph.flexml_layers[112].compute_node[2][3].in[2], compute_graph.flexml_layers[113].compute_node[2][3].in[0], compute_graph.flexml_layers[114].compute_node[2][3].in[0], compute_graph.flexml_layers[115].compute_node[2][3].in[0], compute_graph.flexml_layers[116].compute_node[2][3].in[0], compute_graph.flexml_layers[117].compute_node[2][3].in[0], compute_graph.flexml_layers[117].compute_node[2][3].in[2], compute_graph.flexml_layers[118].compute_node[2][3].in[0], compute_graph.flexml_layers[119].compute_node[2][3].in[0], compute_graph.flexml_layers[120].compute_node[2][3].in[0], compute_graph.flexml_layers[121].compute_node[2][3].in[0], compute_graph.flexml_layers[122].compute_node[2][3].in[0], compute_graph.flexml_layers[123].compute_node[2][3].in[0], compute_graph.flexml_layers[124].compute_node[2][3].in[0], compute_graph.flexml_layers[124].compute_node[2][3].in[2], compute_graph.flexml_layers[125].compute_node[2][3].in[0], compute_graph.flexml_layers[125].compute_node[2][3].in[2], compute_graph.flexml_layers[126].compute_node[2][3].in[0], compute_graph.flexml_layers[127].compute_node[2][3].in[0], compute_graph.flexml_layers[128].compute_node[2][3].in[0], compute_graph.flexml_layers[129].compute_node[2][3].in[0], compute_graph.flexml_layers[130].compute_node[2][3].in[0], compute_graph.flexml_layers[130].compute_node[2][3].in[2], compute_graph.flexml_layers[131].compute_node[2][3].in[0], compute_graph.flexml_layers[132].compute_node[2][3].in[0], compute_graph.flexml_layers[133].compute_node[2][3].in[0], compute_graph.flexml_layers[134].compute_node[2][3].in[0], compute_graph.flexml_layers[135].compute_node[2][3].in[0], compute_graph.flexml_layers[135].compute_node[2][3].in[2], compute_graph.flexml_layers[136].compute_node[2][3].in[0], compute_graph.flexml_layers[137].compute_node[2][3].in[0], compute_graph.flexml_layers[138].compute_node[2][3].in[0], compute_graph.flexml_layers[139].compute_node[2][3].in[0], compute_graph.flexml_layers[140].compute_node[2][3].in[0], compute_graph.flexml_layers[141].compute_node[2][3].in[0], compute_graph.flexml_layers[142].compute_node[2][3].in[0], compute_graph.flexml_layers[142].compute_node[2][3].in[2], compute_graph.flexml_layers[145].compute_node[2][3].in[0], compute_graph.flexml_layers[146].compute_node[2][3].in[0], compute_graph.flexml_layers[147].compute_node[2][3].in[0], compute_graph.flexml_layers[148].compute_node[2][3].in[0], compute_graph.flexml_layers[148].compute_node[2][3].in[2], compute_graph.flexml_layers[149].compute_node[2][3].in[0], compute_graph.flexml_layers[150].compute_node[2][3].in[0], compute_graph.flexml_layers[151].compute_node[2][3].in[0], compute_graph.flexml_layers[152].compute_node[2][3].in[0], compute_graph.flexml_layers[153].compute_node[2][3].in[0], compute_graph.flexml_layers[153].compute_node[2][3].in[2], compute_graph.flexml_layers[154].compute_node[2][3].in[0], compute_graph.flexml_layers[155].compute_node[2][3].in[0], compute_graph.flexml_layers[156].compute_node[2][3].in[0], compute_graph.flexml_layers[157].compute_node[2][3].in[0], compute_graph.flexml_layers[158].compute_node[2][3].in[0], compute_graph.flexml_layers[159].compute_node[2][3].in[0], compute_graph.flexml_layers[160].compute_node[2][3].in[0], compute_graph.flexml_layers[160].compute_node[2][3].in[2], compute_graph.flexml_layers[161].compute_node[2][3].in[0], compute_graph.flexml_layers[161].compute_node[2][3].in[2], compute_graph.flexml_layers[162].compute_node[2][3].in[0], compute_graph.flexml_layers[163].compute_node[2][3].in[0], compute_graph.flexml_layers[164].compute_node[2][3].in[0], compute_graph.flexml_layers[165].compute_node[2][3].in[0], compute_graph.flexml_layers[166].compute_node[2][3].in[0], compute_graph.flexml_layers[166].compute_node[2][3].in[2], compute_graph.flexml_layers[167].compute_node[2][3].in[0], compute_graph.flexml_layers[168].compute_node[2][3].in[0], compute_graph.flexml_layers[169].compute_node[2][3].in[0], compute_graph.flexml_layers[170].compute_node[2][3].in[0], compute_graph.flexml_layers[171].compute_node[2][3].in[0], compute_graph.flexml_layers[171].compute_node[2][3].in[2], compute_graph.flexml_layers[172].compute_node[2][3].in[0], compute_graph.flexml_layers[173].compute_node[2][3].in[0], compute_graph.flexml_layers[174].compute_node[2][3].in[0], compute_graph.flexml_layers[175].compute_node[2][3].in[0], compute_graph.flexml_layers[176].compute_node[2][3].in[0], compute_graph.flexml_layers[177].compute_node[2][3].in[0], compute_graph.flexml_layers[178].compute_node[2][3].in[0], compute_graph.flexml_layers[178].compute_node[2][3].in[2], compute_graph.flexml_layers[179].compute_node[2][3].in[0], compute_graph.flexml_layers[179].compute_node[2][3].in[2], compute_graph.flexml_layers[180].compute_node[2][3].in[0], compute_graph.flexml_layers[181].compute_node[2][3].in[0], compute_graph.flexml_layers[182].compute_node[2][3].in[0], compute_graph.flexml_layers[183].compute_node[2][3].in[0], compute_graph.flexml_layers[184].compute_node[2][3].in[0], compute_graph.flexml_layers[184].compute_node[2][3].in[2], compute_graph.flexml_layers[185].compute_node[2][3].in[0], compute_graph.flexml_layers[186].compute_node[2][3].in[0], compute_graph.flexml_layers[187].compute_node[2][3].in[0], compute_graph.flexml_layers[188].compute_node[2][3].in[0], compute_graph.flexml_layers[188].compute_node[2][3].in[2], compute_graph.flexml_layers[189].compute_node[2][3].in[0], compute_graph.flexml_layers[190].compute_node[2][3].in[0], compute_graph.flexml_layers[191].compute_node[2][3].in[0], compute_graph.flexml_layers[191].compute_node[2][3].in[2], compute_graph.flexml_layers[192].compute_node[2][3].in[0], compute_graph.flexml_layers[192].compute_node[2][3].in[2], compute_graph.flexml_layers[193].compute_node[2][3].in[0], compute_graph.flexml_layers[193].compute_node[2][3].in[2], compute_graph.flexml_layers[194].compute_node[2][3].in[0], compute_graph.flexml_layers[195].compute_node[2][3].in[0], compute_graph.flexml_layers[196].compute_node[2][3].in[0], compute_graph.flexml_layers[196].compute_node[2][3].in[2], compute_graph.flexml_layers[197].compute_node[2][3].in[0], compute_graph.flexml_layers[197].compute_node[2][3].in[2], compute_graph.flexml_layers[198].compute_node[2][3].in[0], compute_graph.flexml_layers[199].compute_node[2][3].in[0], compute_graph.flexml_layers[203].compute_node[2][3].in[0], compute_graph.flexml_layers[204].compute_node[2][3].in[0], compute_graph.flexml_layers[204].compute_node[2][3].in[2], compute_graph.flexml_layers[205].compute_node[2][3].in[0], compute_graph.flexml_layers[205].compute_node[2][3].in[2], compute_graph.flexml_layers[206].compute_node[2][3].in[0], compute_graph.flexml_layers[206].compute_node[2][3].in[2], compute_graph.flexml_layers[207].compute_node[2][3].in[0], compute_graph.flexml_layers[208].compute_node[2][3].in[0], compute_graph.flexml_layers[209].compute_node[2][3].in[0], compute_graph.flexml_layers[209].compute_node[2][3].in[2], compute_graph.flexml_layers[210].compute_node[2][3].in[0], compute_graph.flexml_layers[210].compute_node[2][3].in[2], compute_graph.flexml_layers[211].compute_node[2][3].in[0], compute_graph.flexml_layers[212].compute_node[2][3].in[0], compute_graph.flexml_layers[213].compute_node[2][3].in[0], compute_graph.flexml_layers[214].compute_node[2][3].in[0], compute_graph.flexml_layers[214].compute_node[2][3].in[2], compute_graph.flexml_layers[215].compute_node[2][3].in[0], compute_graph.flexml_layers[215].compute_node[2][3].in[2], compute_graph.flexml_layers[216].compute_node[2][3].in[0], compute_graph.flexml_layers[216].compute_node[2][3].in[2], compute_graph.flexml_layers[217].compute_node[2][3].in[0], compute_graph.flexml_layers[218].compute_node[2][3].in[0], compute_graph.flexml_layers[219].compute_node[2][3].in[0], compute_graph.flexml_layers[219].compute_node[2][3].in[2], compute_graph.flexml_layers[220].compute_node[2][3].in[0], compute_graph.flexml_layers[220].compute_node[2][3].in[2], compute_graph.flexml_layers[221].compute_node[2][3].in[0], compute_graph.flexml_layers[222].compute_node[2][3].in[0], compute_graph.flexml_layers[223].compute_node[2][3].in[0], compute_graph.flexml_layers[224].compute_node[2][3].in[0], compute_graph.flexml_layers[224].compute_node[2][3].in[2], compute_graph.flexml_layers[225].compute_node[2][3].in[0], compute_graph.flexml_layers[225].compute_node[2][3].in[2], compute_graph.flexml_layers[226].compute_node[2][3].in[0], compute_graph.flexml_layers[226].compute_node[2][3].in[2], compute_graph.flexml_layers[227].compute_node[2][3].in[0], compute_graph.flexml_layers[228].compute_node[2][3].in[0], compute_graph.flexml_layers[229].compute_node[2][3].in[0], compute_graph.flexml_layers[229].compute_node[2][3].in[2], compute_graph.flexml_layers[230].compute_node[2][3].in[0], compute_graph.flexml_layers[230].compute_node[2][3].in[2], compute_graph.flexml_layers[231].compute_node[2][3].in[0], compute_graph.flexml_layers[232].compute_node[2][3].in[0], compute_graph.flexml_layers[233].compute_node[2][3].in[0], compute_graph.flexml_layers[234].compute_node[2][3].in[0], compute_graph.flexml_layers[235].compute_node[2][3].in[0], compute_graph.flexml_layers[235].compute_node[2][3].in[2], compute_graph.flexml_layers[236].compute_node[2][3].in[0]) + +Column 2Row 3Channel 0 + +i15_po0(compute_graph.templated_graph_4.trans_comp_nd[2][3].out[0], compute_graph.flexml_layers[36].compute_node[2][3].out[0], compute_graph.templated_graph_231.mk[2][3].out[0], compute_graph.templated_graph_254.mk[2][3].out[0], compute_graph.templated_graph_258.compute_node[2][3].out[0], compute_graph.templated_graph_259.compute_node[2][3].out[0], compute_graph.templated_graph_260.mk[2][3].out[0], compute_graph.templated_graph_263.compute_node[2][3].out[0], compute_graph.templated_graph_266.compute_node[2][3].out[0], compute_graph.templated_graph_268.mk[2][3].out[0], compute_graph.templated_graph_273.mk[2][3].out[0], compute_graph.templated_graph_274.mk[2][3].out[0], compute_graph.flexml_layers[63].compute_node[2][3].out[0], compute_graph.templated_graph_293.mk[2][3].out[0], compute_graph.templated_graph_298.compute_node[2][3].out[0], compute_graph.templated_graph_300.compute_node[2][3].out[0], compute_graph.flexml_layers[0].compute_node[2][3].out[0], compute_graph.flexml_layers[1].compute_node[2][3].out[0], compute_graph.flexml_layers[2].compute_node[2][3].out[0], compute_graph.flexml_layers[3].compute_node[2][3].out[0], compute_graph.flexml_layers[4].compute_node[2][3].out[0], compute_graph.flexml_layers[5].compute_node[2][3].out[0], compute_graph.flexml_layers[6].compute_node[2][3].out[0], compute_graph.flexml_layers[7].compute_node[2][3].out[0], compute_graph.flexml_layers[8].compute_node[2][3].out[0], compute_graph.flexml_layers[9].compute_node[2][3].out[0], compute_graph.flexml_layers[10].compute_node[2][3].out[0], compute_graph.flexml_layers[11].compute_node[2][3].out[0], compute_graph.flexml_layers[12].compute_node[2][3].out[0], compute_graph.flexml_layers[13].compute_node[2][3].out[0], compute_graph.flexml_layers[14].compute_node[2][3].out[0], compute_graph.flexml_layers[15].compute_node[2][3].out[0], compute_graph.flexml_layers[16].compute_node[2][3].out[0], compute_graph.flexml_layers[17].compute_node[2][3].out[0], compute_graph.flexml_layers[18].compute_node[2][3].out[0], compute_graph.flexml_layers[19].compute_node[2][3].out[0], compute_graph.flexml_layers[20].compute_node[2][3].out[0], compute_graph.flexml_layers[21].compute_node[2][3].out[0], compute_graph.flexml_layers[22].compute_node[2][3].out[0], compute_graph.flexml_layers[23].compute_node[2][3].out[0], compute_graph.flexml_layers[24].compute_node[2][3].out[0], compute_graph.flexml_layers[25].compute_node[2][3].out[0], compute_graph.flexml_layers[26].compute_node[2][3].out[0], compute_graph.flexml_layers[27].compute_node[2][3].out[0], compute_graph.flexml_layers[28].compute_node[2][3].out[0], compute_graph.flexml_layers[29].compute_node[2][3].out[0], compute_graph.flexml_layers[30].compute_node[2][3].out[0], compute_graph.flexml_layers[31].compute_node[2][3].out[0], compute_graph.flexml_layers[32].compute_node[2][3].out[0], compute_graph.flexml_layers[33].compute_node[2][3].out[0], compute_graph.flexml_layers[34].compute_node[2][3].out[0], compute_graph.flexml_layers[35].compute_node[2][3].out[0], compute_graph.flexml_layers[143].compute_node[2][3].out[0], compute_graph.flexml_layers[144].compute_node[2][3].out[0], compute_graph.flexml_layers[37].compute_node[2][3].out[0], compute_graph.flexml_layers[38].compute_node[2][3].out[0], compute_graph.flexml_layers[39].compute_node[2][3].out[0], compute_graph.flexml_layers[40].compute_node[2][3].out[0], compute_graph.flexml_layers[41].compute_node[2][3].out[0], compute_graph.flexml_layers[42].compute_node[2][3].out[0], compute_graph.flexml_layers[43].compute_node[2][3].out[0], compute_graph.flexml_layers[44].compute_node[2][3].out[0], compute_graph.flexml_layers[45].compute_node[2][3].out[0], compute_graph.flexml_layers[46].compute_node[2][3].out[0], compute_graph.flexml_layers[47].compute_node[2][3].out[0], compute_graph.flexml_layers[48].compute_node[2][3].out[0], compute_graph.flexml_layers[49].compute_node[2][3].out[0], compute_graph.flexml_layers[50].compute_node[2][3].out[0], compute_graph.flexml_layers[51].compute_node[2][3].out[0], compute_graph.flexml_layers[52].compute_node[2][3].out[0], compute_graph.flexml_layers[53].compute_node[2][3].out[0], compute_graph.flexml_layers[54].compute_node[2][3].out[0], compute_graph.flexml_layers[55].compute_node[2][3].out[0], compute_graph.flexml_layers[56].compute_node[2][3].out[0], compute_graph.flexml_layers[57].compute_node[2][3].out[0], compute_graph.flexml_layers[58].compute_node[2][3].out[0], compute_graph.flexml_layers[59].compute_node[2][3].out[0], compute_graph.flexml_layers[60].compute_node[2][3].out[0], compute_graph.flexml_layers[61].compute_node[2][3].out[0], compute_graph.flexml_layers[62].compute_node[2][3].out[0], compute_graph.flexml_layers[200].compute_node[2][3].out[0], compute_graph.flexml_layers[201].compute_node[2][3].out[0], compute_graph.flexml_layers[202].compute_node[2][3].out[0], compute_graph.flexml_layers[64].compute_node[2][3].out[0], compute_graph.flexml_layers[65].compute_node[2][3].out[0], compute_graph.flexml_layers[66].compute_node[2][3].out[0], compute_graph.flexml_layers[67].compute_node[2][3].out[0], compute_graph.flexml_layers[68].compute_node[2][3].out[0], compute_graph.flexml_layers[69].compute_node[2][3].out[0], compute_graph.flexml_layers[70].compute_node[2][3].out[0], compute_graph.flexml_layers[71].compute_node[2][3].out[0], compute_graph.flexml_layers[72].compute_node[2][3].out[0], compute_graph.flexml_layers[73].compute_node[2][3].out[0], compute_graph.flexml_layers[74].compute_node[2][3].out[0], compute_graph.flexml_layers[75].compute_node[2][3].out[0], compute_graph.flexml_layers[76].compute_node[2][3].out[0], compute_graph.flexml_layers[77].compute_node[2][3].out[0], compute_graph.flexml_layers[78].compute_node[2][3].out[0], compute_graph.flexml_layers[79].compute_node[2][3].out[0], compute_graph.flexml_layers[80].compute_node[2][3].out[0], compute_graph.flexml_layers[81].compute_node[2][3].out[0], compute_graph.flexml_layers[82].compute_node[2][3].out[0], compute_graph.flexml_layers[83].compute_node[2][3].out[0], compute_graph.flexml_layers[84].compute_node[2][3].out[0], compute_graph.flexml_layers[85].compute_node[2][3].out[0], compute_graph.flexml_layers[86].compute_node[2][3].out[0], compute_graph.flexml_layers[87].compute_node[2][3].out[0], compute_graph.flexml_layers[88].compute_node[2][3].out[0], compute_graph.flexml_layers[89].compute_node[2][3].out[0], compute_graph.flexml_layers[90].compute_node[2][3].out[0], compute_graph.flexml_layers[91].compute_node[2][3].out[0], compute_graph.flexml_layers[92].compute_node[2][3].out[0], compute_graph.flexml_layers[93].compute_node[2][3].out[0], compute_graph.flexml_layers[94].compute_node[2][3].out[0], compute_graph.flexml_layers[95].compute_node[2][3].out[0], compute_graph.flexml_layers[96].compute_node[2][3].out[0], compute_graph.flexml_layers[97].compute_node[2][3].out[0], compute_graph.flexml_layers[98].compute_node[2][3].out[0], compute_graph.flexml_layers[99].compute_node[2][3].out[0], compute_graph.flexml_layers[100].compute_node[2][3].out[0], compute_graph.flexml_layers[101].compute_node[2][3].out[0], compute_graph.flexml_layers[102].compute_node[2][3].out[0], compute_graph.flexml_layers[103].compute_node[2][3].out[0], compute_graph.flexml_layers[104].compute_node[2][3].out[0], compute_graph.flexml_layers[105].compute_node[2][3].out[0], compute_graph.flexml_layers[106].compute_node[2][3].out[0], compute_graph.flexml_layers[107].compute_node[2][3].out[0], compute_graph.flexml_layers[108].compute_node[2][3].out[0], compute_graph.flexml_layers[109].compute_node[2][3].out[0], compute_graph.flexml_layers[110].compute_node[2][3].out[0], compute_graph.flexml_layers[111].compute_node[2][3].out[0], compute_graph.flexml_layers[112].compute_node[2][3].out[0], compute_graph.flexml_layers[113].compute_node[2][3].out[0], compute_graph.flexml_layers[114].compute_node[2][3].out[0], compute_graph.flexml_layers[115].compute_node[2][3].out[0], compute_graph.flexml_layers[116].compute_node[2][3].out[0], compute_graph.flexml_layers[117].compute_node[2][3].out[0], compute_graph.flexml_layers[118].compute_node[2][3].out[0], compute_graph.flexml_layers[119].compute_node[2][3].out[0], compute_graph.flexml_layers[120].compute_node[2][3].out[0], compute_graph.flexml_layers[121].compute_node[2][3].out[0], compute_graph.flexml_layers[122].compute_node[2][3].out[0], compute_graph.flexml_layers[123].compute_node[2][3].out[0], compute_graph.flexml_layers[124].compute_node[2][3].out[0], compute_graph.flexml_layers[125].compute_node[2][3].out[0], compute_graph.flexml_layers[126].compute_node[2][3].out[0], compute_graph.flexml_layers[127].compute_node[2][3].out[0], compute_graph.flexml_layers[128].compute_node[2][3].out[0], compute_graph.flexml_layers[129].compute_node[2][3].out[0], compute_graph.flexml_layers[130].compute_node[2][3].out[0], compute_graph.flexml_layers[131].compute_node[2][3].out[0], compute_graph.flexml_layers[132].compute_node[2][3].out[0], compute_graph.flexml_layers[133].compute_node[2][3].out[0], compute_graph.flexml_layers[134].compute_node[2][3].out[0], compute_graph.flexml_layers[135].compute_node[2][3].out[0], compute_graph.flexml_layers[136].compute_node[2][3].out[0], compute_graph.flexml_layers[137].compute_node[2][3].out[0], compute_graph.flexml_layers[138].compute_node[2][3].out[0], compute_graph.flexml_layers[139].compute_node[2][3].out[0], compute_graph.flexml_layers[140].compute_node[2][3].out[0], compute_graph.flexml_layers[141].compute_node[2][3].out[0], compute_graph.flexml_layers[142].compute_node[2][3].out[0], compute_graph.flexml_layers[145].compute_node[2][3].out[0], compute_graph.flexml_layers[146].compute_node[2][3].out[0], compute_graph.flexml_layers[147].compute_node[2][3].out[0], compute_graph.flexml_layers[148].compute_node[2][3].out[0], compute_graph.flexml_layers[149].compute_node[2][3].out[0], compute_graph.flexml_layers[150].compute_node[2][3].out[0], compute_graph.flexml_layers[151].compute_node[2][3].out[0], compute_graph.flexml_layers[152].compute_node[2][3].out[0], compute_graph.flexml_layers[153].compute_node[2][3].out[0], compute_graph.flexml_layers[154].compute_node[2][3].out[0], compute_graph.flexml_layers[155].compute_node[2][3].out[0], compute_graph.flexml_layers[156].compute_node[2][3].out[0], compute_graph.flexml_layers[157].compute_node[2][3].out[0], compute_graph.flexml_layers[158].compute_node[2][3].out[0], compute_graph.flexml_layers[159].compute_node[2][3].out[0], compute_graph.flexml_layers[160].compute_node[2][3].out[0], compute_graph.flexml_layers[161].compute_node[2][3].out[0], compute_graph.flexml_layers[162].compute_node[2][3].out[0], compute_graph.flexml_layers[163].compute_node[2][3].out[0], compute_graph.flexml_layers[164].compute_node[2][3].out[0], compute_graph.flexml_layers[165].compute_node[2][3].out[0], compute_graph.flexml_layers[166].compute_node[2][3].out[0], compute_graph.flexml_layers[167].compute_node[2][3].out[0], compute_graph.flexml_layers[168].compute_node[2][3].out[0], compute_graph.flexml_layers[169].compute_node[2][3].out[0], compute_graph.flexml_layers[170].compute_node[2][3].out[0], compute_graph.flexml_layers[171].compute_node[2][3].out[0], compute_graph.flexml_layers[172].compute_node[2][3].out[0], compute_graph.flexml_layers[173].compute_node[2][3].out[0], compute_graph.flexml_layers[174].compute_node[2][3].out[0], compute_graph.flexml_layers[175].compute_node[2][3].out[0], compute_graph.flexml_layers[176].compute_node[2][3].out[0], compute_graph.flexml_layers[177].compute_node[2][3].out[0], compute_graph.flexml_layers[178].compute_node[2][3].out[0], compute_graph.flexml_layers[179].compute_node[2][3].out[0], compute_graph.flexml_layers[180].compute_node[2][3].out[0], compute_graph.flexml_layers[181].compute_node[2][3].out[0], compute_graph.flexml_layers[182].compute_node[2][3].out[0], compute_graph.flexml_layers[183].compute_node[2][3].out[0], compute_graph.flexml_layers[184].compute_node[2][3].out[0], compute_graph.flexml_layers[185].compute_node[2][3].out[0], compute_graph.flexml_layers[186].compute_node[2][3].out[0], compute_graph.flexml_layers[187].compute_node[2][3].out[0], compute_graph.flexml_layers[188].compute_node[2][3].out[0], compute_graph.flexml_layers[189].compute_node[2][3].out[0], compute_graph.flexml_layers[190].compute_node[2][3].out[0], compute_graph.flexml_layers[191].compute_node[2][3].out[0], compute_graph.flexml_layers[192].compute_node[2][3].out[0], compute_graph.flexml_layers[193].compute_node[2][3].out[0], compute_graph.flexml_layers[194].compute_node[2][3].out[0], compute_graph.flexml_layers[195].compute_node[2][3].out[0], compute_graph.flexml_layers[196].compute_node[2][3].out[0], compute_graph.flexml_layers[197].compute_node[2][3].out[0], compute_graph.flexml_layers[198].compute_node[2][3].out[0], compute_graph.flexml_layers[199].compute_node[2][3].out[0], compute_graph.flexml_layers[203].compute_node[2][3].out[0], compute_graph.flexml_layers[204].compute_node[2][3].out[0], compute_graph.flexml_layers[205].compute_node[2][3].out[0], compute_graph.flexml_layers[206].compute_node[2][3].out[0], compute_graph.flexml_layers[207].compute_node[2][3].out[0], compute_graph.flexml_layers[208].compute_node[2][3].out[0], compute_graph.flexml_layers[209].compute_node[2][3].out[0], compute_graph.flexml_layers[210].compute_node[2][3].out[0], compute_graph.flexml_layers[211].compute_node[2][3].out[0], compute_graph.flexml_layers[212].compute_node[2][3].out[0], compute_graph.flexml_layers[213].compute_node[2][3].out[0], compute_graph.flexml_layers[214].compute_node[2][3].out[0], compute_graph.flexml_layers[215].compute_node[2][3].out[0], compute_graph.flexml_layers[216].compute_node[2][3].out[0], compute_graph.flexml_layers[217].compute_node[2][3].out[0], compute_graph.flexml_layers[218].compute_node[2][3].out[0], compute_graph.flexml_layers[219].compute_node[2][3].out[0], compute_graph.flexml_layers[220].compute_node[2][3].out[0], compute_graph.flexml_layers[221].compute_node[2][3].out[0], compute_graph.flexml_layers[222].compute_node[2][3].out[0], compute_graph.flexml_layers[223].compute_node[2][3].out[0], compute_graph.flexml_layers[224].compute_node[2][3].out[0], compute_graph.flexml_layers[225].compute_node[2][3].out[0], compute_graph.flexml_layers[226].compute_node[2][3].out[0], compute_graph.flexml_layers[227].compute_node[2][3].out[0], compute_graph.flexml_layers[228].compute_node[2][3].out[0], compute_graph.flexml_layers[229].compute_node[2][3].out[0], compute_graph.flexml_layers[230].compute_node[2][3].out[0], compute_graph.flexml_layers[231].compute_node[2][3].out[0], compute_graph.flexml_layers[232].compute_node[2][3].out[0], compute_graph.flexml_layers[233].compute_node[2][3].out[0], compute_graph.flexml_layers[234].compute_node[2][3].out[0], compute_graph.flexml_layers[235].compute_node[2][3].out[0], compute_graph.flexml_layers[236].compute_node[2][3].out[0]) + +Column 2Row 3Channel 1 + +i15_pi1(compute_graph.flexml_layers[36].compute_node[2][3].in[1], compute_graph.templated_graph_260.mk[2][3].in[1], compute_graph.templated_graph_268.mk[2][3].in[1], compute_graph.templated_graph_273.mk[2][3].in[1], compute_graph.flexml_layers[3].compute_node[2][3].in[1], compute_graph.flexml_layers[8].compute_node[2][3].in[1], compute_graph.flexml_layers[9].compute_node[2][3].in[1], compute_graph.flexml_layers[10].compute_node[2][3].in[1], compute_graph.flexml_layers[11].compute_node[2][3].in[1], compute_graph.flexml_layers[12].compute_node[2][3].in[1], compute_graph.flexml_layers[13].compute_node[2][3].in[1], compute_graph.flexml_layers[14].compute_node[2][3].in[1], compute_graph.flexml_layers[15].compute_node[2][3].in[1], compute_graph.flexml_layers[16].compute_node[2][3].in[1], compute_graph.flexml_layers[17].compute_node[2][3].in[1], compute_graph.flexml_layers[19].compute_node[2][3].in[1], compute_graph.flexml_layers[20].compute_node[2][3].in[1], compute_graph.flexml_layers[25].compute_node[2][3].in[1], compute_graph.flexml_layers[26].compute_node[2][3].in[1], compute_graph.flexml_layers[27].compute_node[2][3].in[1], compute_graph.flexml_layers[29].compute_node[2][3].in[1], compute_graph.flexml_layers[30].compute_node[2][3].in[1], compute_graph.flexml_layers[35].compute_node[2][3].in[1], compute_graph.flexml_layers[143].compute_node[2][3].in[1], compute_graph.flexml_layers[144].compute_node[2][3].in[1], compute_graph.flexml_layers[37].compute_node[2][3].in[1], compute_graph.flexml_layers[39].compute_node[2][3].in[1], compute_graph.flexml_layers[40].compute_node[2][3].in[1], compute_graph.flexml_layers[45].compute_node[2][3].in[1], compute_graph.flexml_layers[46].compute_node[2][3].in[1], compute_graph.flexml_layers[51].compute_node[2][3].in[1], compute_graph.flexml_layers[56].compute_node[2][3].in[1], compute_graph.flexml_layers[57].compute_node[2][3].in[1], compute_graph.flexml_layers[62].compute_node[2][3].in[1], compute_graph.flexml_layers[201].compute_node[2][3].in[1], compute_graph.flexml_layers[202].compute_node[2][3].in[1], compute_graph.flexml_layers[67].compute_node[2][3].in[1], compute_graph.flexml_layers[68].compute_node[2][3].in[1], compute_graph.flexml_layers[73].compute_node[2][3].in[1], compute_graph.flexml_layers[78].compute_node[2][3].in[1], compute_graph.flexml_layers[79].compute_node[2][3].in[1], compute_graph.flexml_layers[84].compute_node[2][3].in[1], compute_graph.flexml_layers[89].compute_node[2][3].in[1], compute_graph.flexml_layers[90].compute_node[2][3].in[1], compute_graph.flexml_layers[95].compute_node[2][3].in[1], compute_graph.flexml_layers[101].compute_node[2][3].in[1], compute_graph.flexml_layers[102].compute_node[2][3].in[1], compute_graph.flexml_layers[107].compute_node[2][3].in[1], compute_graph.flexml_layers[108].compute_node[2][3].in[1], compute_graph.flexml_layers[113].compute_node[2][3].in[1], compute_graph.flexml_layers[119].compute_node[2][3].in[1], compute_graph.flexml_layers[120].compute_node[2][3].in[1], compute_graph.flexml_layers[125].compute_node[2][3].in[1], compute_graph.flexml_layers[126].compute_node[2][3].in[1], compute_graph.flexml_layers[131].compute_node[2][3].in[1], compute_graph.flexml_layers[137].compute_node[2][3].in[1], compute_graph.flexml_layers[138].compute_node[2][3].in[1], compute_graph.flexml_layers[149].compute_node[2][3].in[1], compute_graph.flexml_layers[155].compute_node[2][3].in[1], compute_graph.flexml_layers[156].compute_node[2][3].in[1], compute_graph.flexml_layers[161].compute_node[2][3].in[1], compute_graph.flexml_layers[162].compute_node[2][3].in[1], compute_graph.flexml_layers[167].compute_node[2][3].in[1], compute_graph.flexml_layers[173].compute_node[2][3].in[1], compute_graph.flexml_layers[174].compute_node[2][3].in[1], compute_graph.flexml_layers[179].compute_node[2][3].in[1], compute_graph.flexml_layers[180].compute_node[2][3].in[1], compute_graph.flexml_layers[186].compute_node[2][3].in[1], compute_graph.flexml_layers[188].compute_node[2][3].in[1], compute_graph.flexml_layers[189].compute_node[2][3].in[1], compute_graph.flexml_layers[194].compute_node[2][3].in[1], compute_graph.flexml_layers[207].compute_node[2][3].in[1], compute_graph.flexml_layers[211].compute_node[2][3].in[1], compute_graph.flexml_layers[212].compute_node[2][3].in[1], compute_graph.flexml_layers[217].compute_node[2][3].in[1], compute_graph.flexml_layers[221].compute_node[2][3].in[1], compute_graph.flexml_layers[222].compute_node[2][3].in[1], compute_graph.flexml_layers[227].compute_node[2][3].in[1], compute_graph.flexml_layers[231].compute_node[2][3].in[1], compute_graph.flexml_layers[232].compute_node[2][3].in[1], compute_graph.flexml_layers[233].compute_node[2][3].in[1]) + +Column 3Row 0Channel 0 + +i16_pi0(compute_graph.templated_graph_4.trans_comp_nd[3][0].in[0], compute_graph.templated_graph_231.mk[3][0].in[0], compute_graph.templated_graph_254.mk[3][0].in[0], compute_graph.templated_graph_258.compute_node[3][0].in[0], compute_graph.templated_graph_259.compute_node[3][0].in[0], compute_graph.templated_graph_260.mk[3][0].in[0], compute_graph.templated_graph_263.compute_node[3][0].in[0], compute_graph.templated_graph_266.compute_node[3][0].in[0], compute_graph.templated_graph_268.mk[3][0].in[0], compute_graph.templated_graph_273.mk[3][0].in[0], compute_graph.templated_graph_274.mk[3][0].in[0], compute_graph.flexml_layers[63].compute_node[3][0].in[0], compute_graph.templated_graph_293.mk[3][0].in[0], compute_graph.templated_graph_298.compute_node[3][0].in[0], compute_graph.templated_graph_300.compute_node[3][0].in[0], compute_graph.flexml_layers[0].compute_node[3][0].in[0], compute_graph.flexml_layers[1].compute_node[3][0].in[0], compute_graph.flexml_layers[1].compute_node[3][0].in[2], compute_graph.flexml_layers[2].compute_node[3][0].in[0], compute_graph.flexml_layers[2].compute_node[3][0].in[2], compute_graph.flexml_layers[3].compute_node[3][0].in[0], compute_graph.flexml_layers[4].compute_node[3][0].in[0], compute_graph.flexml_layers[5].compute_node[3][0].in[0], compute_graph.flexml_layers[6].compute_node[3][0].in[0], compute_graph.flexml_layers[7].compute_node[3][0].in[0], compute_graph.flexml_layers[7].compute_node[3][0].in[2], compute_graph.flexml_layers[8].compute_node[3][0].in[0], compute_graph.flexml_layers[9].compute_node[3][0].in[0], compute_graph.flexml_layers[9].compute_node[3][0].in[2], compute_graph.flexml_layers[10].compute_node[3][0].in[0], compute_graph.flexml_layers[11].compute_node[3][0].in[0], compute_graph.flexml_layers[12].compute_node[3][0].in[0], compute_graph.flexml_layers[13].compute_node[3][0].in[0], compute_graph.flexml_layers[14].compute_node[3][0].in[0], compute_graph.flexml_layers[15].compute_node[3][0].in[0], compute_graph.flexml_layers[15].compute_node[3][0].in[2], compute_graph.flexml_layers[16].compute_node[3][0].in[0], compute_graph.flexml_layers[17].compute_node[3][0].in[0], compute_graph.flexml_layers[18].compute_node[3][0].in[0], compute_graph.flexml_layers[19].compute_node[3][0].in[0], compute_graph.flexml_layers[20].compute_node[3][0].in[0], compute_graph.flexml_layers[21].compute_node[3][0].in[0], compute_graph.flexml_layers[22].compute_node[3][0].in[0], compute_graph.flexml_layers[23].compute_node[3][0].in[0], compute_graph.flexml_layers[24].compute_node[3][0].in[0], compute_graph.flexml_layers[24].compute_node[3][0].in[2], compute_graph.flexml_layers[25].compute_node[3][0].in[0], compute_graph.flexml_layers[26].compute_node[3][0].in[0], compute_graph.flexml_layers[27].compute_node[3][0].in[0], compute_graph.flexml_layers[28].compute_node[3][0].in[0], compute_graph.flexml_layers[29].compute_node[3][0].in[0], compute_graph.flexml_layers[30].compute_node[3][0].in[0], compute_graph.flexml_layers[31].compute_node[3][0].in[0], compute_graph.flexml_layers[32].compute_node[3][0].in[0], compute_graph.flexml_layers[33].compute_node[3][0].in[0], compute_graph.flexml_layers[34].compute_node[3][0].in[0], compute_graph.flexml_layers[34].compute_node[3][0].in[2], compute_graph.flexml_layers[35].compute_node[3][0].in[0], compute_graph.flexml_layers[35].compute_node[3][0].in[2], compute_graph.flexml_layers[143].compute_node[3][0].in[0], compute_graph.flexml_layers[144].compute_node[3][0].in[0], compute_graph.flexml_layers[36].compute_node[3][0].in[0], compute_graph.flexml_layers[37].compute_node[3][0].in[0], compute_graph.flexml_layers[38].compute_node[3][0].in[0], compute_graph.flexml_layers[39].compute_node[3][0].in[0], compute_graph.flexml_layers[40].compute_node[3][0].in[0], compute_graph.flexml_layers[41].compute_node[3][0].in[0], compute_graph.flexml_layers[42].compute_node[3][0].in[0], compute_graph.flexml_layers[43].compute_node[3][0].in[0], compute_graph.flexml_layers[44].compute_node[3][0].in[0], compute_graph.flexml_layers[44].compute_node[3][0].in[2], compute_graph.flexml_layers[45].compute_node[3][0].in[0], compute_graph.flexml_layers[45].compute_node[3][0].in[2], compute_graph.flexml_layers[46].compute_node[3][0].in[0], compute_graph.flexml_layers[47].compute_node[3][0].in[0], compute_graph.flexml_layers[48].compute_node[3][0].in[0], compute_graph.flexml_layers[49].compute_node[3][0].in[0], compute_graph.flexml_layers[50].compute_node[3][0].in[0], compute_graph.flexml_layers[50].compute_node[3][0].in[2], compute_graph.flexml_layers[51].compute_node[3][0].in[0], compute_graph.flexml_layers[52].compute_node[3][0].in[0], compute_graph.flexml_layers[53].compute_node[3][0].in[0], compute_graph.flexml_layers[54].compute_node[3][0].in[0], compute_graph.flexml_layers[55].compute_node[3][0].in[0], compute_graph.flexml_layers[55].compute_node[3][0].in[2], compute_graph.flexml_layers[56].compute_node[3][0].in[0], compute_graph.flexml_layers[57].compute_node[3][0].in[0], compute_graph.flexml_layers[58].compute_node[3][0].in[0], compute_graph.flexml_layers[59].compute_node[3][0].in[0], compute_graph.flexml_layers[60].compute_node[3][0].in[0], compute_graph.flexml_layers[61].compute_node[3][0].in[0], compute_graph.flexml_layers[61].compute_node[3][0].in[2], compute_graph.flexml_layers[62].compute_node[3][0].in[0], compute_graph.flexml_layers[200].compute_node[3][0].in[0], compute_graph.flexml_layers[201].compute_node[3][0].in[0], compute_graph.flexml_layers[202].compute_node[3][0].in[0], compute_graph.flexml_layers[64].compute_node[3][0].in[0], compute_graph.flexml_layers[65].compute_node[3][0].in[0], compute_graph.flexml_layers[66].compute_node[3][0].in[0], compute_graph.flexml_layers[66].compute_node[3][0].in[2], compute_graph.flexml_layers[67].compute_node[3][0].in[0], compute_graph.flexml_layers[67].compute_node[3][0].in[2], compute_graph.flexml_layers[68].compute_node[3][0].in[0], compute_graph.flexml_layers[69].compute_node[3][0].in[0], compute_graph.flexml_layers[70].compute_node[3][0].in[0], compute_graph.flexml_layers[71].compute_node[3][0].in[0], compute_graph.flexml_layers[72].compute_node[3][0].in[0], compute_graph.flexml_layers[72].compute_node[3][0].in[2], compute_graph.flexml_layers[73].compute_node[3][0].in[0], compute_graph.flexml_layers[74].compute_node[3][0].in[0], compute_graph.flexml_layers[75].compute_node[3][0].in[0], compute_graph.flexml_layers[76].compute_node[3][0].in[0], compute_graph.flexml_layers[77].compute_node[3][0].in[0], compute_graph.flexml_layers[77].compute_node[3][0].in[2], compute_graph.flexml_layers[78].compute_node[3][0].in[0], compute_graph.flexml_layers[78].compute_node[3][0].in[2], compute_graph.flexml_layers[79].compute_node[3][0].in[0], compute_graph.flexml_layers[80].compute_node[3][0].in[0], compute_graph.flexml_layers[81].compute_node[3][0].in[0], compute_graph.flexml_layers[82].compute_node[3][0].in[0], compute_graph.flexml_layers[83].compute_node[3][0].in[0], compute_graph.flexml_layers[83].compute_node[3][0].in[2], compute_graph.flexml_layers[84].compute_node[3][0].in[0], compute_graph.flexml_layers[85].compute_node[3][0].in[0], compute_graph.flexml_layers[86].compute_node[3][0].in[0], compute_graph.flexml_layers[87].compute_node[3][0].in[0], compute_graph.flexml_layers[88].compute_node[3][0].in[0], compute_graph.flexml_layers[88].compute_node[3][0].in[2], compute_graph.flexml_layers[89].compute_node[3][0].in[0], compute_graph.flexml_layers[89].compute_node[3][0].in[2], compute_graph.flexml_layers[90].compute_node[3][0].in[0], compute_graph.flexml_layers[91].compute_node[3][0].in[0], compute_graph.flexml_layers[92].compute_node[3][0].in[0], compute_graph.flexml_layers[93].compute_node[3][0].in[0], compute_graph.flexml_layers[94].compute_node[3][0].in[0], compute_graph.flexml_layers[94].compute_node[3][0].in[2], compute_graph.flexml_layers[95].compute_node[3][0].in[0], compute_graph.flexml_layers[96].compute_node[3][0].in[0], compute_graph.flexml_layers[97].compute_node[3][0].in[0], compute_graph.flexml_layers[98].compute_node[3][0].in[0], compute_graph.flexml_layers[99].compute_node[3][0].in[0], compute_graph.flexml_layers[99].compute_node[3][0].in[2], compute_graph.flexml_layers[100].compute_node[3][0].in[0], compute_graph.flexml_layers[101].compute_node[3][0].in[0], compute_graph.flexml_layers[102].compute_node[3][0].in[0], compute_graph.flexml_layers[103].compute_node[3][0].in[0], compute_graph.flexml_layers[104].compute_node[3][0].in[0], compute_graph.flexml_layers[105].compute_node[3][0].in[0], compute_graph.flexml_layers[106].compute_node[3][0].in[0], compute_graph.flexml_layers[106].compute_node[3][0].in[2], compute_graph.flexml_layers[107].compute_node[3][0].in[0], compute_graph.flexml_layers[108].compute_node[3][0].in[0], compute_graph.flexml_layers[109].compute_node[3][0].in[0], compute_graph.flexml_layers[110].compute_node[3][0].in[0], compute_graph.flexml_layers[111].compute_node[3][0].in[0], compute_graph.flexml_layers[112].compute_node[3][0].in[0], compute_graph.flexml_layers[112].compute_node[3][0].in[2], compute_graph.flexml_layers[113].compute_node[3][0].in[0], compute_graph.flexml_layers[114].compute_node[3][0].in[0], compute_graph.flexml_layers[115].compute_node[3][0].in[0], compute_graph.flexml_layers[116].compute_node[3][0].in[0], compute_graph.flexml_layers[117].compute_node[3][0].in[0], compute_graph.flexml_layers[117].compute_node[3][0].in[2], compute_graph.flexml_layers[118].compute_node[3][0].in[0], compute_graph.flexml_layers[119].compute_node[3][0].in[0], compute_graph.flexml_layers[120].compute_node[3][0].in[0], compute_graph.flexml_layers[121].compute_node[3][0].in[0], compute_graph.flexml_layers[122].compute_node[3][0].in[0], compute_graph.flexml_layers[123].compute_node[3][0].in[0], compute_graph.flexml_layers[124].compute_node[3][0].in[0], compute_graph.flexml_layers[124].compute_node[3][0].in[2], compute_graph.flexml_layers[125].compute_node[3][0].in[0], compute_graph.flexml_layers[125].compute_node[3][0].in[2], compute_graph.flexml_layers[126].compute_node[3][0].in[0], compute_graph.flexml_layers[127].compute_node[3][0].in[0], compute_graph.flexml_layers[128].compute_node[3][0].in[0], compute_graph.flexml_layers[129].compute_node[3][0].in[0], compute_graph.flexml_layers[130].compute_node[3][0].in[0], compute_graph.flexml_layers[130].compute_node[3][0].in[2], compute_graph.flexml_layers[131].compute_node[3][0].in[0], compute_graph.flexml_layers[132].compute_node[3][0].in[0], compute_graph.flexml_layers[133].compute_node[3][0].in[0], compute_graph.flexml_layers[134].compute_node[3][0].in[0], compute_graph.flexml_layers[135].compute_node[3][0].in[0], compute_graph.flexml_layers[135].compute_node[3][0].in[2], compute_graph.flexml_layers[136].compute_node[3][0].in[0], compute_graph.flexml_layers[137].compute_node[3][0].in[0], compute_graph.flexml_layers[138].compute_node[3][0].in[0], compute_graph.flexml_layers[139].compute_node[3][0].in[0], compute_graph.flexml_layers[140].compute_node[3][0].in[0], compute_graph.flexml_layers[141].compute_node[3][0].in[0], compute_graph.flexml_layers[142].compute_node[3][0].in[0], compute_graph.flexml_layers[142].compute_node[3][0].in[2], compute_graph.flexml_layers[145].compute_node[3][0].in[0], compute_graph.flexml_layers[146].compute_node[3][0].in[0], compute_graph.flexml_layers[147].compute_node[3][0].in[0], compute_graph.flexml_layers[148].compute_node[3][0].in[0], compute_graph.flexml_layers[148].compute_node[3][0].in[2], compute_graph.flexml_layers[149].compute_node[3][0].in[0], compute_graph.flexml_layers[150].compute_node[3][0].in[0], compute_graph.flexml_layers[151].compute_node[3][0].in[0], compute_graph.flexml_layers[152].compute_node[3][0].in[0], compute_graph.flexml_layers[153].compute_node[3][0].in[0], compute_graph.flexml_layers[153].compute_node[3][0].in[2], compute_graph.flexml_layers[154].compute_node[3][0].in[0], compute_graph.flexml_layers[155].compute_node[3][0].in[0], compute_graph.flexml_layers[156].compute_node[3][0].in[0], compute_graph.flexml_layers[157].compute_node[3][0].in[0], compute_graph.flexml_layers[158].compute_node[3][0].in[0], compute_graph.flexml_layers[159].compute_node[3][0].in[0], compute_graph.flexml_layers[160].compute_node[3][0].in[0], compute_graph.flexml_layers[160].compute_node[3][0].in[2], compute_graph.flexml_layers[161].compute_node[3][0].in[0], compute_graph.flexml_layers[161].compute_node[3][0].in[2], compute_graph.flexml_layers[162].compute_node[3][0].in[0], compute_graph.flexml_layers[163].compute_node[3][0].in[0], compute_graph.flexml_layers[164].compute_node[3][0].in[0], compute_graph.flexml_layers[165].compute_node[3][0].in[0], compute_graph.flexml_layers[166].compute_node[3][0].in[0], compute_graph.flexml_layers[166].compute_node[3][0].in[2], compute_graph.flexml_layers[167].compute_node[3][0].in[0], compute_graph.flexml_layers[168].compute_node[3][0].in[0], compute_graph.flexml_layers[169].compute_node[3][0].in[0], compute_graph.flexml_layers[170].compute_node[3][0].in[0], compute_graph.flexml_layers[171].compute_node[3][0].in[0], compute_graph.flexml_layers[171].compute_node[3][0].in[2], compute_graph.flexml_layers[172].compute_node[3][0].in[0], compute_graph.flexml_layers[173].compute_node[3][0].in[0], compute_graph.flexml_layers[174].compute_node[3][0].in[0], compute_graph.flexml_layers[175].compute_node[3][0].in[0], compute_graph.flexml_layers[176].compute_node[3][0].in[0], compute_graph.flexml_layers[177].compute_node[3][0].in[0], compute_graph.flexml_layers[178].compute_node[3][0].in[0], compute_graph.flexml_layers[178].compute_node[3][0].in[2], compute_graph.flexml_layers[179].compute_node[3][0].in[0], compute_graph.flexml_layers[179].compute_node[3][0].in[2], compute_graph.flexml_layers[180].compute_node[3][0].in[0], compute_graph.flexml_layers[181].compute_node[3][0].in[0], compute_graph.flexml_layers[182].compute_node[3][0].in[0], compute_graph.flexml_layers[183].compute_node[3][0].in[0], compute_graph.flexml_layers[184].compute_node[3][0].in[0], compute_graph.flexml_layers[184].compute_node[3][0].in[2], compute_graph.flexml_layers[185].compute_node[3][0].in[0], compute_graph.flexml_layers[186].compute_node[3][0].in[0], compute_graph.flexml_layers[187].compute_node[3][0].in[0], compute_graph.flexml_layers[188].compute_node[3][0].in[0], compute_graph.flexml_layers[188].compute_node[3][0].in[2], compute_graph.flexml_layers[189].compute_node[3][0].in[0], compute_graph.flexml_layers[190].compute_node[3][0].in[0], compute_graph.flexml_layers[191].compute_node[3][0].in[0], compute_graph.flexml_layers[191].compute_node[3][0].in[2], compute_graph.flexml_layers[192].compute_node[3][0].in[0], compute_graph.flexml_layers[192].compute_node[3][0].in[2], compute_graph.flexml_layers[193].compute_node[3][0].in[0], compute_graph.flexml_layers[193].compute_node[3][0].in[2], compute_graph.flexml_layers[194].compute_node[3][0].in[0], compute_graph.flexml_layers[195].compute_node[3][0].in[0], compute_graph.flexml_layers[196].compute_node[3][0].in[0], compute_graph.flexml_layers[196].compute_node[3][0].in[2], compute_graph.flexml_layers[197].compute_node[3][0].in[0], compute_graph.flexml_layers[197].compute_node[3][0].in[2], compute_graph.flexml_layers[198].compute_node[3][0].in[0], compute_graph.flexml_layers[199].compute_node[3][0].in[0], compute_graph.flexml_layers[203].compute_node[3][0].in[0], compute_graph.flexml_layers[204].compute_node[3][0].in[0], compute_graph.flexml_layers[204].compute_node[3][0].in[2], compute_graph.flexml_layers[205].compute_node[3][0].in[0], compute_graph.flexml_layers[205].compute_node[3][0].in[2], compute_graph.flexml_layers[206].compute_node[3][0].in[0], compute_graph.flexml_layers[206].compute_node[3][0].in[2], compute_graph.flexml_layers[207].compute_node[3][0].in[0], compute_graph.flexml_layers[208].compute_node[3][0].in[0], compute_graph.flexml_layers[209].compute_node[3][0].in[0], compute_graph.flexml_layers[209].compute_node[3][0].in[2], compute_graph.flexml_layers[210].compute_node[3][0].in[0], compute_graph.flexml_layers[210].compute_node[3][0].in[2], compute_graph.flexml_layers[211].compute_node[3][0].in[0], compute_graph.flexml_layers[212].compute_node[3][0].in[0], compute_graph.flexml_layers[213].compute_node[3][0].in[0], compute_graph.flexml_layers[214].compute_node[3][0].in[0], compute_graph.flexml_layers[214].compute_node[3][0].in[2], compute_graph.flexml_layers[215].compute_node[3][0].in[0], compute_graph.flexml_layers[215].compute_node[3][0].in[2], compute_graph.flexml_layers[216].compute_node[3][0].in[0], compute_graph.flexml_layers[216].compute_node[3][0].in[2], compute_graph.flexml_layers[217].compute_node[3][0].in[0], compute_graph.flexml_layers[218].compute_node[3][0].in[0], compute_graph.flexml_layers[219].compute_node[3][0].in[0], compute_graph.flexml_layers[219].compute_node[3][0].in[2], compute_graph.flexml_layers[220].compute_node[3][0].in[0], compute_graph.flexml_layers[220].compute_node[3][0].in[2], compute_graph.flexml_layers[221].compute_node[3][0].in[0], compute_graph.flexml_layers[222].compute_node[3][0].in[0], compute_graph.flexml_layers[223].compute_node[3][0].in[0], compute_graph.flexml_layers[224].compute_node[3][0].in[0], compute_graph.flexml_layers[224].compute_node[3][0].in[2], compute_graph.flexml_layers[225].compute_node[3][0].in[0], compute_graph.flexml_layers[225].compute_node[3][0].in[2], compute_graph.flexml_layers[226].compute_node[3][0].in[0], compute_graph.flexml_layers[226].compute_node[3][0].in[2], compute_graph.flexml_layers[227].compute_node[3][0].in[0], compute_graph.flexml_layers[228].compute_node[3][0].in[0], compute_graph.flexml_layers[229].compute_node[3][0].in[0], compute_graph.flexml_layers[229].compute_node[3][0].in[2], compute_graph.flexml_layers[230].compute_node[3][0].in[0], compute_graph.flexml_layers[230].compute_node[3][0].in[2], compute_graph.flexml_layers[231].compute_node[3][0].in[0], compute_graph.flexml_layers[232].compute_node[3][0].in[0], compute_graph.flexml_layers[233].compute_node[3][0].in[0], compute_graph.flexml_layers[234].compute_node[3][0].in[0], compute_graph.flexml_layers[235].compute_node[3][0].in[0], compute_graph.flexml_layers[235].compute_node[3][0].in[2], compute_graph.flexml_layers[236].compute_node[3][0].in[0]) + +Column 3Row 0Channel 0 + +i16_po0(compute_graph.templated_graph_4.trans_comp_nd[3][0].out[0], compute_graph.templated_graph_231.mk[3][0].out[0], compute_graph.templated_graph_254.mk[3][0].out[0], compute_graph.templated_graph_258.compute_node[3][0].out[0], compute_graph.templated_graph_259.compute_node[3][0].out[0], compute_graph.templated_graph_260.mk[3][0].out[0], compute_graph.templated_graph_263.compute_node[3][0].out[0], compute_graph.templated_graph_266.compute_node[3][0].out[0], compute_graph.templated_graph_268.mk[3][0].out[0], compute_graph.templated_graph_273.mk[3][0].out[0], compute_graph.templated_graph_274.mk[3][0].out[0], compute_graph.flexml_layers[63].compute_node[3][0].out[0], compute_graph.templated_graph_293.mk[3][0].out[0], compute_graph.templated_graph_298.compute_node[3][0].out[0], compute_graph.templated_graph_300.compute_node[3][0].out[0], compute_graph.flexml_layers[0].compute_node[3][0].out[0], compute_graph.flexml_layers[1].compute_node[3][0].out[0], compute_graph.flexml_layers[2].compute_node[3][0].out[0], compute_graph.flexml_layers[3].compute_node[3][0].out[0], compute_graph.flexml_layers[4].compute_node[3][0].out[0], compute_graph.flexml_layers[5].compute_node[3][0].out[0], compute_graph.flexml_layers[6].compute_node[3][0].out[0], compute_graph.flexml_layers[7].compute_node[3][0].out[0], compute_graph.flexml_layers[8].compute_node[3][0].out[0], compute_graph.flexml_layers[9].compute_node[3][0].out[0], compute_graph.flexml_layers[10].compute_node[3][0].out[0], compute_graph.flexml_layers[11].compute_node[3][0].out[0], compute_graph.flexml_layers[12].compute_node[3][0].out[0], compute_graph.flexml_layers[13].compute_node[3][0].out[0], compute_graph.flexml_layers[14].compute_node[3][0].out[0], compute_graph.flexml_layers[15].compute_node[3][0].out[0], compute_graph.flexml_layers[16].compute_node[3][0].out[0], compute_graph.flexml_layers[17].compute_node[3][0].out[0], compute_graph.flexml_layers[18].compute_node[3][0].out[0], compute_graph.flexml_layers[19].compute_node[3][0].out[0], compute_graph.flexml_layers[20].compute_node[3][0].out[0], compute_graph.flexml_layers[21].compute_node[3][0].out[0], compute_graph.flexml_layers[22].compute_node[3][0].out[0], compute_graph.flexml_layers[23].compute_node[3][0].out[0], compute_graph.flexml_layers[24].compute_node[3][0].out[0], compute_graph.flexml_layers[25].compute_node[3][0].out[0], compute_graph.flexml_layers[26].compute_node[3][0].out[0], compute_graph.flexml_layers[27].compute_node[3][0].out[0], compute_graph.flexml_layers[28].compute_node[3][0].out[0], compute_graph.flexml_layers[29].compute_node[3][0].out[0], compute_graph.flexml_layers[30].compute_node[3][0].out[0], compute_graph.flexml_layers[31].compute_node[3][0].out[0], compute_graph.flexml_layers[32].compute_node[3][0].out[0], compute_graph.flexml_layers[33].compute_node[3][0].out[0], compute_graph.flexml_layers[34].compute_node[3][0].out[0], compute_graph.flexml_layers[35].compute_node[3][0].out[0], compute_graph.flexml_layers[143].compute_node[3][0].out[0], compute_graph.flexml_layers[144].compute_node[3][0].out[0], compute_graph.flexml_layers[36].compute_node[3][0].out[0], compute_graph.flexml_layers[37].compute_node[3][0].out[0], compute_graph.flexml_layers[38].compute_node[3][0].out[0], compute_graph.flexml_layers[39].compute_node[3][0].out[0], compute_graph.flexml_layers[40].compute_node[3][0].out[0], compute_graph.flexml_layers[41].compute_node[3][0].out[0], compute_graph.flexml_layers[42].compute_node[3][0].out[0], compute_graph.flexml_layers[43].compute_node[3][0].out[0], compute_graph.flexml_layers[44].compute_node[3][0].out[0], compute_graph.flexml_layers[45].compute_node[3][0].out[0], compute_graph.flexml_layers[46].compute_node[3][0].out[0], compute_graph.flexml_layers[47].compute_node[3][0].out[0], compute_graph.flexml_layers[48].compute_node[3][0].out[0], compute_graph.flexml_layers[49].compute_node[3][0].out[0], compute_graph.flexml_layers[50].compute_node[3][0].out[0], compute_graph.flexml_layers[51].compute_node[3][0].out[0], compute_graph.flexml_layers[52].compute_node[3][0].out[0], compute_graph.flexml_layers[53].compute_node[3][0].out[0], compute_graph.flexml_layers[54].compute_node[3][0].out[0], compute_graph.flexml_layers[55].compute_node[3][0].out[0], compute_graph.flexml_layers[56].compute_node[3][0].out[0], compute_graph.flexml_layers[57].compute_node[3][0].out[0], compute_graph.flexml_layers[58].compute_node[3][0].out[0], compute_graph.flexml_layers[59].compute_node[3][0].out[0], compute_graph.flexml_layers[60].compute_node[3][0].out[0], compute_graph.flexml_layers[61].compute_node[3][0].out[0], compute_graph.flexml_layers[62].compute_node[3][0].out[0], compute_graph.flexml_layers[200].compute_node[3][0].out[0], compute_graph.flexml_layers[201].compute_node[3][0].out[0], compute_graph.flexml_layers[64].compute_node[3][0].out[0], compute_graph.flexml_layers[65].compute_node[3][0].out[0], compute_graph.flexml_layers[66].compute_node[3][0].out[0], compute_graph.flexml_layers[67].compute_node[3][0].out[0], compute_graph.flexml_layers[68].compute_node[3][0].out[0], compute_graph.flexml_layers[69].compute_node[3][0].out[0], compute_graph.flexml_layers[70].compute_node[3][0].out[0], compute_graph.flexml_layers[71].compute_node[3][0].out[0], compute_graph.flexml_layers[72].compute_node[3][0].out[0], compute_graph.flexml_layers[73].compute_node[3][0].out[0], compute_graph.flexml_layers[74].compute_node[3][0].out[0], compute_graph.flexml_layers[75].compute_node[3][0].out[0], compute_graph.flexml_layers[76].compute_node[3][0].out[0], compute_graph.flexml_layers[77].compute_node[3][0].out[0], compute_graph.flexml_layers[78].compute_node[3][0].out[0], compute_graph.flexml_layers[79].compute_node[3][0].out[0], compute_graph.flexml_layers[80].compute_node[3][0].out[0], compute_graph.flexml_layers[81].compute_node[3][0].out[0], compute_graph.flexml_layers[82].compute_node[3][0].out[0], compute_graph.flexml_layers[83].compute_node[3][0].out[0], compute_graph.flexml_layers[84].compute_node[3][0].out[0], compute_graph.flexml_layers[85].compute_node[3][0].out[0], compute_graph.flexml_layers[86].compute_node[3][0].out[0], compute_graph.flexml_layers[87].compute_node[3][0].out[0], compute_graph.flexml_layers[88].compute_node[3][0].out[0], compute_graph.flexml_layers[89].compute_node[3][0].out[0], compute_graph.flexml_layers[90].compute_node[3][0].out[0], compute_graph.flexml_layers[91].compute_node[3][0].out[0], compute_graph.flexml_layers[92].compute_node[3][0].out[0], compute_graph.flexml_layers[93].compute_node[3][0].out[0], compute_graph.flexml_layers[94].compute_node[3][0].out[0], compute_graph.flexml_layers[95].compute_node[3][0].out[0], compute_graph.flexml_layers[96].compute_node[3][0].out[0], compute_graph.flexml_layers[97].compute_node[3][0].out[0], compute_graph.flexml_layers[98].compute_node[3][0].out[0], compute_graph.flexml_layers[99].compute_node[3][0].out[0], compute_graph.flexml_layers[100].compute_node[3][0].out[0], compute_graph.flexml_layers[101].compute_node[3][0].out[0], compute_graph.flexml_layers[102].compute_node[3][0].out[0], compute_graph.flexml_layers[103].compute_node[3][0].out[0], compute_graph.flexml_layers[104].compute_node[3][0].out[0], compute_graph.flexml_layers[105].compute_node[3][0].out[0], compute_graph.flexml_layers[106].compute_node[3][0].out[0], compute_graph.flexml_layers[107].compute_node[3][0].out[0], compute_graph.flexml_layers[108].compute_node[3][0].out[0], compute_graph.flexml_layers[109].compute_node[3][0].out[0], compute_graph.flexml_layers[110].compute_node[3][0].out[0], compute_graph.flexml_layers[111].compute_node[3][0].out[0], compute_graph.flexml_layers[112].compute_node[3][0].out[0], compute_graph.flexml_layers[113].compute_node[3][0].out[0], compute_graph.flexml_layers[114].compute_node[3][0].out[0], compute_graph.flexml_layers[115].compute_node[3][0].out[0], compute_graph.flexml_layers[116].compute_node[3][0].out[0], compute_graph.flexml_layers[117].compute_node[3][0].out[0], compute_graph.flexml_layers[118].compute_node[3][0].out[0], compute_graph.flexml_layers[119].compute_node[3][0].out[0], compute_graph.flexml_layers[120].compute_node[3][0].out[0], compute_graph.flexml_layers[121].compute_node[3][0].out[0], compute_graph.flexml_layers[122].compute_node[3][0].out[0], compute_graph.flexml_layers[123].compute_node[3][0].out[0], compute_graph.flexml_layers[124].compute_node[3][0].out[0], compute_graph.flexml_layers[125].compute_node[3][0].out[0], compute_graph.flexml_layers[126].compute_node[3][0].out[0], compute_graph.flexml_layers[127].compute_node[3][0].out[0], compute_graph.flexml_layers[128].compute_node[3][0].out[0], compute_graph.flexml_layers[129].compute_node[3][0].out[0], compute_graph.flexml_layers[130].compute_node[3][0].out[0], compute_graph.flexml_layers[131].compute_node[3][0].out[0], compute_graph.flexml_layers[132].compute_node[3][0].out[0], compute_graph.flexml_layers[133].compute_node[3][0].out[0], compute_graph.flexml_layers[134].compute_node[3][0].out[0], compute_graph.flexml_layers[135].compute_node[3][0].out[0], compute_graph.flexml_layers[136].compute_node[3][0].out[0], compute_graph.flexml_layers[137].compute_node[3][0].out[0], compute_graph.flexml_layers[138].compute_node[3][0].out[0], compute_graph.flexml_layers[139].compute_node[3][0].out[0], compute_graph.flexml_layers[140].compute_node[3][0].out[0], compute_graph.flexml_layers[141].compute_node[3][0].out[0], compute_graph.flexml_layers[142].compute_node[3][0].out[0], compute_graph.flexml_layers[145].compute_node[3][0].out[0], compute_graph.flexml_layers[146].compute_node[3][0].out[0], compute_graph.flexml_layers[147].compute_node[3][0].out[0], compute_graph.flexml_layers[148].compute_node[3][0].out[0], compute_graph.flexml_layers[149].compute_node[3][0].out[0], compute_graph.flexml_layers[150].compute_node[3][0].out[0], compute_graph.flexml_layers[151].compute_node[3][0].out[0], compute_graph.flexml_layers[152].compute_node[3][0].out[0], compute_graph.flexml_layers[153].compute_node[3][0].out[0], compute_graph.flexml_layers[154].compute_node[3][0].out[0], compute_graph.flexml_layers[155].compute_node[3][0].out[0], compute_graph.flexml_layers[156].compute_node[3][0].out[0], compute_graph.flexml_layers[157].compute_node[3][0].out[0], compute_graph.flexml_layers[158].compute_node[3][0].out[0], compute_graph.flexml_layers[159].compute_node[3][0].out[0], compute_graph.flexml_layers[160].compute_node[3][0].out[0], compute_graph.flexml_layers[161].compute_node[3][0].out[0], compute_graph.flexml_layers[162].compute_node[3][0].out[0], compute_graph.flexml_layers[163].compute_node[3][0].out[0], compute_graph.flexml_layers[164].compute_node[3][0].out[0], compute_graph.flexml_layers[165].compute_node[3][0].out[0], compute_graph.flexml_layers[166].compute_node[3][0].out[0], compute_graph.flexml_layers[167].compute_node[3][0].out[0], compute_graph.flexml_layers[168].compute_node[3][0].out[0], compute_graph.flexml_layers[169].compute_node[3][0].out[0], compute_graph.flexml_layers[170].compute_node[3][0].out[0], compute_graph.flexml_layers[171].compute_node[3][0].out[0], compute_graph.flexml_layers[172].compute_node[3][0].out[0], compute_graph.flexml_layers[173].compute_node[3][0].out[0], compute_graph.flexml_layers[174].compute_node[3][0].out[0], compute_graph.flexml_layers[175].compute_node[3][0].out[0], compute_graph.flexml_layers[176].compute_node[3][0].out[0], compute_graph.flexml_layers[177].compute_node[3][0].out[0], compute_graph.flexml_layers[178].compute_node[3][0].out[0], compute_graph.flexml_layers[179].compute_node[3][0].out[0], compute_graph.flexml_layers[180].compute_node[3][0].out[0], compute_graph.flexml_layers[181].compute_node[3][0].out[0], compute_graph.flexml_layers[182].compute_node[3][0].out[0], compute_graph.flexml_layers[183].compute_node[3][0].out[0], compute_graph.flexml_layers[184].compute_node[3][0].out[0], compute_graph.flexml_layers[185].compute_node[3][0].out[0], compute_graph.flexml_layers[186].compute_node[3][0].out[0], compute_graph.flexml_layers[187].compute_node[3][0].out[0], compute_graph.flexml_layers[188].compute_node[3][0].out[0], compute_graph.flexml_layers[189].compute_node[3][0].out[0], compute_graph.flexml_layers[190].compute_node[3][0].out[0], compute_graph.flexml_layers[191].compute_node[3][0].out[0], compute_graph.flexml_layers[192].compute_node[3][0].out[0], compute_graph.flexml_layers[193].compute_node[3][0].out[0], compute_graph.flexml_layers[194].compute_node[3][0].out[0], compute_graph.flexml_layers[195].compute_node[3][0].out[0], compute_graph.flexml_layers[196].compute_node[3][0].out[0], compute_graph.flexml_layers[197].compute_node[3][0].out[0], compute_graph.flexml_layers[198].compute_node[3][0].out[0], compute_graph.flexml_layers[199].compute_node[3][0].out[0], compute_graph.flexml_layers[202].compute_node[3][0].out[0], compute_graph.flexml_layers[203].compute_node[3][0].out[0], compute_graph.flexml_layers[204].compute_node[3][0].out[0], compute_graph.flexml_layers[205].compute_node[3][0].out[0], compute_graph.flexml_layers[206].compute_node[3][0].out[0], compute_graph.flexml_layers[207].compute_node[3][0].out[0], compute_graph.flexml_layers[208].compute_node[3][0].out[0], compute_graph.flexml_layers[209].compute_node[3][0].out[0], compute_graph.flexml_layers[210].compute_node[3][0].out[0], compute_graph.flexml_layers[211].compute_node[3][0].out[0], compute_graph.flexml_layers[212].compute_node[3][0].out[0], compute_graph.flexml_layers[213].compute_node[3][0].out[0], compute_graph.flexml_layers[214].compute_node[3][0].out[0], compute_graph.flexml_layers[215].compute_node[3][0].out[0], compute_graph.flexml_layers[216].compute_node[3][0].out[0], compute_graph.flexml_layers[217].compute_node[3][0].out[0], compute_graph.flexml_layers[218].compute_node[3][0].out[0], compute_graph.flexml_layers[219].compute_node[3][0].out[0], compute_graph.flexml_layers[220].compute_node[3][0].out[0], compute_graph.flexml_layers[221].compute_node[3][0].out[0], compute_graph.flexml_layers[222].compute_node[3][0].out[0], compute_graph.flexml_layers[223].compute_node[3][0].out[0], compute_graph.flexml_layers[224].compute_node[3][0].out[0], compute_graph.flexml_layers[225].compute_node[3][0].out[0], compute_graph.flexml_layers[226].compute_node[3][0].out[0], compute_graph.flexml_layers[227].compute_node[3][0].out[0], compute_graph.flexml_layers[228].compute_node[3][0].out[0], compute_graph.flexml_layers[229].compute_node[3][0].out[0], compute_graph.flexml_layers[230].compute_node[3][0].out[0], compute_graph.flexml_layers[231].compute_node[3][0].out[0], compute_graph.flexml_layers[232].compute_node[3][0].out[0], compute_graph.flexml_layers[233].compute_node[3][0].out[0], compute_graph.flexml_layers[234].compute_node[3][0].out[0], compute_graph.flexml_layers[235].compute_node[3][0].out[0], compute_graph.flexml_layers[236].compute_node[3][0].out[0]) + +Column 3Row 0Channel 1 + +i16_pi1(compute_graph.templated_graph_260.mk[3][0].in[1], compute_graph.templated_graph_268.mk[3][0].in[1], compute_graph.templated_graph_273.mk[3][0].in[1], compute_graph.flexml_layers[3].compute_node[3][0].in[1], compute_graph.flexml_layers[8].compute_node[3][0].in[1], compute_graph.flexml_layers[9].compute_node[3][0].in[1], compute_graph.flexml_layers[10].compute_node[3][0].in[1], compute_graph.flexml_layers[11].compute_node[3][0].in[1], compute_graph.flexml_layers[12].compute_node[3][0].in[1], compute_graph.flexml_layers[13].compute_node[3][0].in[1], compute_graph.flexml_layers[14].compute_node[3][0].in[1], compute_graph.flexml_layers[15].compute_node[3][0].in[1], compute_graph.flexml_layers[16].compute_node[3][0].in[1], compute_graph.flexml_layers[17].compute_node[3][0].in[1], compute_graph.flexml_layers[19].compute_node[3][0].in[1], compute_graph.flexml_layers[20].compute_node[3][0].in[1], compute_graph.flexml_layers[25].compute_node[3][0].in[1], compute_graph.flexml_layers[26].compute_node[3][0].in[1], compute_graph.flexml_layers[27].compute_node[3][0].in[1], compute_graph.flexml_layers[29].compute_node[3][0].in[1], compute_graph.flexml_layers[30].compute_node[3][0].in[1], compute_graph.flexml_layers[35].compute_node[3][0].in[1], compute_graph.flexml_layers[143].compute_node[3][0].in[1], compute_graph.flexml_layers[144].compute_node[3][0].in[1], compute_graph.flexml_layers[36].compute_node[3][0].in[1], compute_graph.flexml_layers[37].compute_node[3][0].in[1], compute_graph.flexml_layers[39].compute_node[3][0].in[1], compute_graph.flexml_layers[40].compute_node[3][0].in[1], compute_graph.flexml_layers[45].compute_node[3][0].in[1], compute_graph.flexml_layers[46].compute_node[3][0].in[1], compute_graph.flexml_layers[51].compute_node[3][0].in[1], compute_graph.flexml_layers[56].compute_node[3][0].in[1], compute_graph.flexml_layers[57].compute_node[3][0].in[1], compute_graph.flexml_layers[62].compute_node[3][0].in[1], compute_graph.flexml_layers[201].compute_node[3][0].in[1], compute_graph.flexml_layers[202].compute_node[3][0].in[1], compute_graph.flexml_layers[67].compute_node[3][0].in[1], compute_graph.flexml_layers[68].compute_node[3][0].in[1], compute_graph.flexml_layers[73].compute_node[3][0].in[1], compute_graph.flexml_layers[78].compute_node[3][0].in[1], compute_graph.flexml_layers[79].compute_node[3][0].in[1], compute_graph.flexml_layers[84].compute_node[3][0].in[1], compute_graph.flexml_layers[89].compute_node[3][0].in[1], compute_graph.flexml_layers[90].compute_node[3][0].in[1], compute_graph.flexml_layers[95].compute_node[3][0].in[1], compute_graph.flexml_layers[101].compute_node[3][0].in[1], compute_graph.flexml_layers[102].compute_node[3][0].in[1], compute_graph.flexml_layers[107].compute_node[3][0].in[1], compute_graph.flexml_layers[108].compute_node[3][0].in[1], compute_graph.flexml_layers[113].compute_node[3][0].in[1], compute_graph.flexml_layers[119].compute_node[3][0].in[1], compute_graph.flexml_layers[120].compute_node[3][0].in[1], compute_graph.flexml_layers[125].compute_node[3][0].in[1], compute_graph.flexml_layers[126].compute_node[3][0].in[1], compute_graph.flexml_layers[131].compute_node[3][0].in[1], compute_graph.flexml_layers[137].compute_node[3][0].in[1], compute_graph.flexml_layers[138].compute_node[3][0].in[1], compute_graph.flexml_layers[149].compute_node[3][0].in[1], compute_graph.flexml_layers[155].compute_node[3][0].in[1], compute_graph.flexml_layers[156].compute_node[3][0].in[1], compute_graph.flexml_layers[161].compute_node[3][0].in[1], compute_graph.flexml_layers[162].compute_node[3][0].in[1], compute_graph.flexml_layers[167].compute_node[3][0].in[1], compute_graph.flexml_layers[173].compute_node[3][0].in[1], compute_graph.flexml_layers[174].compute_node[3][0].in[1], compute_graph.flexml_layers[179].compute_node[3][0].in[1], compute_graph.flexml_layers[180].compute_node[3][0].in[1], compute_graph.flexml_layers[186].compute_node[3][0].in[1], compute_graph.flexml_layers[188].compute_node[3][0].in[1], compute_graph.flexml_layers[189].compute_node[3][0].in[1], compute_graph.flexml_layers[194].compute_node[3][0].in[1], compute_graph.flexml_layers[207].compute_node[3][0].in[1], compute_graph.flexml_layers[211].compute_node[3][0].in[1], compute_graph.flexml_layers[212].compute_node[3][0].in[1], compute_graph.flexml_layers[217].compute_node[3][0].in[1], compute_graph.flexml_layers[221].compute_node[3][0].in[1], compute_graph.flexml_layers[222].compute_node[3][0].in[1], compute_graph.flexml_layers[227].compute_node[3][0].in[1], compute_graph.flexml_layers[231].compute_node[3][0].in[1], compute_graph.flexml_layers[232].compute_node[3][0].in[1], compute_graph.flexml_layers[233].compute_node[3][0].in[1]) + +Column 3Row 1Channel 0 + +i17_pi0(compute_graph.templated_graph_4.trans_comp_nd[3][1].in[0], compute_graph.flexml_layers[35].compute_node[3][1].in[2], compute_graph.templated_graph_231.mk[3][1].in[0], compute_graph.templated_graph_254.mk[3][1].in[0], compute_graph.templated_graph_258.compute_node[3][1].in[0], compute_graph.templated_graph_259.compute_node[3][1].in[0], compute_graph.templated_graph_260.mk[3][1].in[0], compute_graph.templated_graph_263.compute_node[3][1].in[0], compute_graph.templated_graph_266.compute_node[3][1].in[0], compute_graph.templated_graph_268.mk[3][1].in[0], compute_graph.templated_graph_273.mk[3][1].in[0], compute_graph.templated_graph_274.mk[3][1].in[0], compute_graph.flexml_layers[63].compute_node[3][1].in[0], compute_graph.templated_graph_293.mk[3][1].in[0], compute_graph.templated_graph_298.compute_node[3][1].in[0], compute_graph.templated_graph_300.compute_node[3][1].in[0], compute_graph.flexml_layers[0].compute_node[3][1].in[0], compute_graph.flexml_layers[1].compute_node[3][1].in[0], compute_graph.flexml_layers[1].compute_node[3][1].in[2], compute_graph.flexml_layers[2].compute_node[3][1].in[0], compute_graph.flexml_layers[2].compute_node[3][1].in[2], compute_graph.flexml_layers[3].compute_node[3][1].in[0], compute_graph.flexml_layers[4].compute_node[3][1].in[0], compute_graph.flexml_layers[5].compute_node[3][1].in[0], compute_graph.flexml_layers[6].compute_node[3][1].in[0], compute_graph.flexml_layers[7].compute_node[3][1].in[0], compute_graph.flexml_layers[7].compute_node[3][1].in[2], compute_graph.flexml_layers[8].compute_node[3][1].in[0], compute_graph.flexml_layers[9].compute_node[3][1].in[0], compute_graph.flexml_layers[9].compute_node[3][1].in[2], compute_graph.flexml_layers[10].compute_node[3][1].in[0], compute_graph.flexml_layers[11].compute_node[3][1].in[0], compute_graph.flexml_layers[12].compute_node[3][1].in[0], compute_graph.flexml_layers[13].compute_node[3][1].in[0], compute_graph.flexml_layers[14].compute_node[3][1].in[0], compute_graph.flexml_layers[15].compute_node[3][1].in[0], compute_graph.flexml_layers[15].compute_node[3][1].in[2], compute_graph.flexml_layers[16].compute_node[3][1].in[0], compute_graph.flexml_layers[17].compute_node[3][1].in[0], compute_graph.flexml_layers[18].compute_node[3][1].in[0], compute_graph.flexml_layers[19].compute_node[3][1].in[0], compute_graph.flexml_layers[20].compute_node[3][1].in[0], compute_graph.flexml_layers[21].compute_node[3][1].in[0], compute_graph.flexml_layers[22].compute_node[3][1].in[0], compute_graph.flexml_layers[23].compute_node[3][1].in[0], compute_graph.flexml_layers[24].compute_node[3][1].in[0], compute_graph.flexml_layers[24].compute_node[3][1].in[2], compute_graph.flexml_layers[25].compute_node[3][1].in[0], compute_graph.flexml_layers[26].compute_node[3][1].in[0], compute_graph.flexml_layers[27].compute_node[3][1].in[0], compute_graph.flexml_layers[28].compute_node[3][1].in[0], compute_graph.flexml_layers[29].compute_node[3][1].in[0], compute_graph.flexml_layers[30].compute_node[3][1].in[0], compute_graph.flexml_layers[31].compute_node[3][1].in[0], compute_graph.flexml_layers[32].compute_node[3][1].in[0], compute_graph.flexml_layers[33].compute_node[3][1].in[0], compute_graph.flexml_layers[34].compute_node[3][1].in[0], compute_graph.flexml_layers[34].compute_node[3][1].in[2], compute_graph.flexml_layers[35].compute_node[3][1].in[0], compute_graph.flexml_layers[143].compute_node[3][1].in[0], compute_graph.flexml_layers[144].compute_node[3][1].in[0], compute_graph.flexml_layers[36].compute_node[3][1].in[0], compute_graph.flexml_layers[37].compute_node[3][1].in[0], compute_graph.flexml_layers[38].compute_node[3][1].in[0], compute_graph.flexml_layers[39].compute_node[3][1].in[0], compute_graph.flexml_layers[40].compute_node[3][1].in[0], compute_graph.flexml_layers[41].compute_node[3][1].in[0], compute_graph.flexml_layers[42].compute_node[3][1].in[0], compute_graph.flexml_layers[43].compute_node[3][1].in[0], compute_graph.flexml_layers[44].compute_node[3][1].in[0], compute_graph.flexml_layers[44].compute_node[3][1].in[2], compute_graph.flexml_layers[45].compute_node[3][1].in[0], compute_graph.flexml_layers[45].compute_node[3][1].in[2], compute_graph.flexml_layers[46].compute_node[3][1].in[0], compute_graph.flexml_layers[47].compute_node[3][1].in[0], compute_graph.flexml_layers[48].compute_node[3][1].in[0], compute_graph.flexml_layers[49].compute_node[3][1].in[0], compute_graph.flexml_layers[50].compute_node[3][1].in[0], compute_graph.flexml_layers[50].compute_node[3][1].in[2], compute_graph.flexml_layers[51].compute_node[3][1].in[0], compute_graph.flexml_layers[52].compute_node[3][1].in[0], compute_graph.flexml_layers[53].compute_node[3][1].in[0], compute_graph.flexml_layers[54].compute_node[3][1].in[0], compute_graph.flexml_layers[55].compute_node[3][1].in[0], compute_graph.flexml_layers[55].compute_node[3][1].in[2], compute_graph.flexml_layers[56].compute_node[3][1].in[0], compute_graph.flexml_layers[57].compute_node[3][1].in[0], compute_graph.flexml_layers[58].compute_node[3][1].in[0], compute_graph.flexml_layers[59].compute_node[3][1].in[0], compute_graph.flexml_layers[60].compute_node[3][1].in[0], compute_graph.flexml_layers[61].compute_node[3][1].in[0], compute_graph.flexml_layers[61].compute_node[3][1].in[2], compute_graph.flexml_layers[62].compute_node[3][1].in[0], compute_graph.flexml_layers[200].compute_node[3][1].in[0], compute_graph.flexml_layers[201].compute_node[3][1].in[0], compute_graph.flexml_layers[64].compute_node[3][1].in[0], compute_graph.flexml_layers[65].compute_node[3][1].in[0], compute_graph.flexml_layers[66].compute_node[3][1].in[0], compute_graph.flexml_layers[66].compute_node[3][1].in[2], compute_graph.flexml_layers[67].compute_node[3][1].in[0], compute_graph.flexml_layers[67].compute_node[3][1].in[2], compute_graph.flexml_layers[68].compute_node[3][1].in[0], compute_graph.flexml_layers[69].compute_node[3][1].in[0], compute_graph.flexml_layers[70].compute_node[3][1].in[0], compute_graph.flexml_layers[71].compute_node[3][1].in[0], compute_graph.flexml_layers[72].compute_node[3][1].in[0], compute_graph.flexml_layers[72].compute_node[3][1].in[2], compute_graph.flexml_layers[73].compute_node[3][1].in[0], compute_graph.flexml_layers[74].compute_node[3][1].in[0], compute_graph.flexml_layers[75].compute_node[3][1].in[0], compute_graph.flexml_layers[76].compute_node[3][1].in[0], compute_graph.flexml_layers[77].compute_node[3][1].in[0], compute_graph.flexml_layers[77].compute_node[3][1].in[2], compute_graph.flexml_layers[78].compute_node[3][1].in[0], compute_graph.flexml_layers[78].compute_node[3][1].in[2], compute_graph.flexml_layers[79].compute_node[3][1].in[0], compute_graph.flexml_layers[80].compute_node[3][1].in[0], compute_graph.flexml_layers[81].compute_node[3][1].in[0], compute_graph.flexml_layers[82].compute_node[3][1].in[0], compute_graph.flexml_layers[83].compute_node[3][1].in[0], compute_graph.flexml_layers[83].compute_node[3][1].in[2], compute_graph.flexml_layers[84].compute_node[3][1].in[0], compute_graph.flexml_layers[85].compute_node[3][1].in[0], compute_graph.flexml_layers[86].compute_node[3][1].in[0], compute_graph.flexml_layers[87].compute_node[3][1].in[0], compute_graph.flexml_layers[88].compute_node[3][1].in[0], compute_graph.flexml_layers[88].compute_node[3][1].in[2], compute_graph.flexml_layers[89].compute_node[3][1].in[0], compute_graph.flexml_layers[89].compute_node[3][1].in[2], compute_graph.flexml_layers[90].compute_node[3][1].in[0], compute_graph.flexml_layers[91].compute_node[3][1].in[0], compute_graph.flexml_layers[92].compute_node[3][1].in[0], compute_graph.flexml_layers[93].compute_node[3][1].in[0], compute_graph.flexml_layers[94].compute_node[3][1].in[0], compute_graph.flexml_layers[94].compute_node[3][1].in[2], compute_graph.flexml_layers[95].compute_node[3][1].in[0], compute_graph.flexml_layers[96].compute_node[3][1].in[0], compute_graph.flexml_layers[97].compute_node[3][1].in[0], compute_graph.flexml_layers[98].compute_node[3][1].in[0], compute_graph.flexml_layers[99].compute_node[3][1].in[0], compute_graph.flexml_layers[99].compute_node[3][1].in[2], compute_graph.flexml_layers[100].compute_node[3][1].in[0], compute_graph.flexml_layers[101].compute_node[3][1].in[0], compute_graph.flexml_layers[102].compute_node[3][1].in[0], compute_graph.flexml_layers[103].compute_node[3][1].in[0], compute_graph.flexml_layers[104].compute_node[3][1].in[0], compute_graph.flexml_layers[105].compute_node[3][1].in[0], compute_graph.flexml_layers[106].compute_node[3][1].in[0], compute_graph.flexml_layers[106].compute_node[3][1].in[2], compute_graph.flexml_layers[107].compute_node[3][1].in[0], compute_graph.flexml_layers[108].compute_node[3][1].in[0], compute_graph.flexml_layers[109].compute_node[3][1].in[0], compute_graph.flexml_layers[110].compute_node[3][1].in[0], compute_graph.flexml_layers[111].compute_node[3][1].in[0], compute_graph.flexml_layers[112].compute_node[3][1].in[0], compute_graph.flexml_layers[112].compute_node[3][1].in[2], compute_graph.flexml_layers[113].compute_node[3][1].in[0], compute_graph.flexml_layers[114].compute_node[3][1].in[0], compute_graph.flexml_layers[115].compute_node[3][1].in[0], compute_graph.flexml_layers[116].compute_node[3][1].in[0], compute_graph.flexml_layers[117].compute_node[3][1].in[0], compute_graph.flexml_layers[117].compute_node[3][1].in[2], compute_graph.flexml_layers[118].compute_node[3][1].in[0], compute_graph.flexml_layers[119].compute_node[3][1].in[0], compute_graph.flexml_layers[120].compute_node[3][1].in[0], compute_graph.flexml_layers[121].compute_node[3][1].in[0], compute_graph.flexml_layers[122].compute_node[3][1].in[0], compute_graph.flexml_layers[123].compute_node[3][1].in[0], compute_graph.flexml_layers[124].compute_node[3][1].in[0], compute_graph.flexml_layers[124].compute_node[3][1].in[2], compute_graph.flexml_layers[125].compute_node[3][1].in[0], compute_graph.flexml_layers[125].compute_node[3][1].in[2], compute_graph.flexml_layers[126].compute_node[3][1].in[0], compute_graph.flexml_layers[127].compute_node[3][1].in[0], compute_graph.flexml_layers[128].compute_node[3][1].in[0], compute_graph.flexml_layers[129].compute_node[3][1].in[0], compute_graph.flexml_layers[130].compute_node[3][1].in[0], compute_graph.flexml_layers[130].compute_node[3][1].in[2], compute_graph.flexml_layers[131].compute_node[3][1].in[0], compute_graph.flexml_layers[132].compute_node[3][1].in[0], compute_graph.flexml_layers[133].compute_node[3][1].in[0], compute_graph.flexml_layers[134].compute_node[3][1].in[0], compute_graph.flexml_layers[135].compute_node[3][1].in[0], compute_graph.flexml_layers[135].compute_node[3][1].in[2], compute_graph.flexml_layers[136].compute_node[3][1].in[0], compute_graph.flexml_layers[137].compute_node[3][1].in[0], compute_graph.flexml_layers[138].compute_node[3][1].in[0], compute_graph.flexml_layers[139].compute_node[3][1].in[0], compute_graph.flexml_layers[140].compute_node[3][1].in[0], compute_graph.flexml_layers[141].compute_node[3][1].in[0], compute_graph.flexml_layers[142].compute_node[3][1].in[0], compute_graph.flexml_layers[142].compute_node[3][1].in[2], compute_graph.flexml_layers[145].compute_node[3][1].in[0], compute_graph.flexml_layers[146].compute_node[3][1].in[0], compute_graph.flexml_layers[147].compute_node[3][1].in[0], compute_graph.flexml_layers[148].compute_node[3][1].in[0], compute_graph.flexml_layers[148].compute_node[3][1].in[2], compute_graph.flexml_layers[149].compute_node[3][1].in[0], compute_graph.flexml_layers[150].compute_node[3][1].in[0], compute_graph.flexml_layers[151].compute_node[3][1].in[0], compute_graph.flexml_layers[152].compute_node[3][1].in[0], compute_graph.flexml_layers[153].compute_node[3][1].in[0], compute_graph.flexml_layers[153].compute_node[3][1].in[2], compute_graph.flexml_layers[154].compute_node[3][1].in[0], compute_graph.flexml_layers[155].compute_node[3][1].in[0], compute_graph.flexml_layers[156].compute_node[3][1].in[0], compute_graph.flexml_layers[157].compute_node[3][1].in[0], compute_graph.flexml_layers[158].compute_node[3][1].in[0], compute_graph.flexml_layers[159].compute_node[3][1].in[0], compute_graph.flexml_layers[160].compute_node[3][1].in[0], compute_graph.flexml_layers[160].compute_node[3][1].in[2], compute_graph.flexml_layers[161].compute_node[3][1].in[0], compute_graph.flexml_layers[161].compute_node[3][1].in[2], compute_graph.flexml_layers[162].compute_node[3][1].in[0], compute_graph.flexml_layers[163].compute_node[3][1].in[0], compute_graph.flexml_layers[164].compute_node[3][1].in[0], compute_graph.flexml_layers[165].compute_node[3][1].in[0], compute_graph.flexml_layers[166].compute_node[3][1].in[0], compute_graph.flexml_layers[166].compute_node[3][1].in[2], compute_graph.flexml_layers[167].compute_node[3][1].in[0], compute_graph.flexml_layers[168].compute_node[3][1].in[0], compute_graph.flexml_layers[169].compute_node[3][1].in[0], compute_graph.flexml_layers[170].compute_node[3][1].in[0], compute_graph.flexml_layers[171].compute_node[3][1].in[0], compute_graph.flexml_layers[171].compute_node[3][1].in[2], compute_graph.flexml_layers[172].compute_node[3][1].in[0], compute_graph.flexml_layers[173].compute_node[3][1].in[0], compute_graph.flexml_layers[174].compute_node[3][1].in[0], compute_graph.flexml_layers[175].compute_node[3][1].in[0], compute_graph.flexml_layers[176].compute_node[3][1].in[0], compute_graph.flexml_layers[177].compute_node[3][1].in[0], compute_graph.flexml_layers[178].compute_node[3][1].in[0], compute_graph.flexml_layers[178].compute_node[3][1].in[2], compute_graph.flexml_layers[179].compute_node[3][1].in[0], compute_graph.flexml_layers[179].compute_node[3][1].in[2], compute_graph.flexml_layers[180].compute_node[3][1].in[0], compute_graph.flexml_layers[181].compute_node[3][1].in[0], compute_graph.flexml_layers[182].compute_node[3][1].in[0], compute_graph.flexml_layers[183].compute_node[3][1].in[0], compute_graph.flexml_layers[184].compute_node[3][1].in[0], compute_graph.flexml_layers[184].compute_node[3][1].in[2], compute_graph.flexml_layers[185].compute_node[3][1].in[0], compute_graph.flexml_layers[186].compute_node[3][1].in[0], compute_graph.flexml_layers[187].compute_node[3][1].in[0], compute_graph.flexml_layers[188].compute_node[3][1].in[0], compute_graph.flexml_layers[188].compute_node[3][1].in[2], compute_graph.flexml_layers[189].compute_node[3][1].in[0], compute_graph.flexml_layers[190].compute_node[3][1].in[0], compute_graph.flexml_layers[191].compute_node[3][1].in[0], compute_graph.flexml_layers[191].compute_node[3][1].in[2], compute_graph.flexml_layers[192].compute_node[3][1].in[0], compute_graph.flexml_layers[192].compute_node[3][1].in[2], compute_graph.flexml_layers[193].compute_node[3][1].in[0], compute_graph.flexml_layers[193].compute_node[3][1].in[2], compute_graph.flexml_layers[194].compute_node[3][1].in[0], compute_graph.flexml_layers[195].compute_node[3][1].in[0], compute_graph.flexml_layers[196].compute_node[3][1].in[0], compute_graph.flexml_layers[196].compute_node[3][1].in[2], compute_graph.flexml_layers[197].compute_node[3][1].in[0], compute_graph.flexml_layers[197].compute_node[3][1].in[2], compute_graph.flexml_layers[198].compute_node[3][1].in[0], compute_graph.flexml_layers[199].compute_node[3][1].in[0], compute_graph.flexml_layers[202].compute_node[3][1].in[0], compute_graph.flexml_layers[203].compute_node[3][1].in[0], compute_graph.flexml_layers[204].compute_node[3][1].in[0], compute_graph.flexml_layers[204].compute_node[3][1].in[2], compute_graph.flexml_layers[205].compute_node[3][1].in[0], compute_graph.flexml_layers[205].compute_node[3][1].in[2], compute_graph.flexml_layers[206].compute_node[3][1].in[0], compute_graph.flexml_layers[206].compute_node[3][1].in[2], compute_graph.flexml_layers[207].compute_node[3][1].in[0], compute_graph.flexml_layers[208].compute_node[3][1].in[0], compute_graph.flexml_layers[209].compute_node[3][1].in[0], compute_graph.flexml_layers[209].compute_node[3][1].in[2], compute_graph.flexml_layers[210].compute_node[3][1].in[0], compute_graph.flexml_layers[210].compute_node[3][1].in[2], compute_graph.flexml_layers[211].compute_node[3][1].in[0], compute_graph.flexml_layers[212].compute_node[3][1].in[0], compute_graph.flexml_layers[213].compute_node[3][1].in[0], compute_graph.flexml_layers[214].compute_node[3][1].in[0], compute_graph.flexml_layers[214].compute_node[3][1].in[2], compute_graph.flexml_layers[215].compute_node[3][1].in[0], compute_graph.flexml_layers[215].compute_node[3][1].in[2], compute_graph.flexml_layers[216].compute_node[3][1].in[0], compute_graph.flexml_layers[216].compute_node[3][1].in[2], compute_graph.flexml_layers[217].compute_node[3][1].in[0], compute_graph.flexml_layers[218].compute_node[3][1].in[0], compute_graph.flexml_layers[219].compute_node[3][1].in[0], compute_graph.flexml_layers[219].compute_node[3][1].in[2], compute_graph.flexml_layers[220].compute_node[3][1].in[0], compute_graph.flexml_layers[220].compute_node[3][1].in[2], compute_graph.flexml_layers[221].compute_node[3][1].in[0], compute_graph.flexml_layers[222].compute_node[3][1].in[0], compute_graph.flexml_layers[223].compute_node[3][1].in[0], compute_graph.flexml_layers[224].compute_node[3][1].in[0], compute_graph.flexml_layers[224].compute_node[3][1].in[2], compute_graph.flexml_layers[225].compute_node[3][1].in[0], compute_graph.flexml_layers[225].compute_node[3][1].in[2], compute_graph.flexml_layers[226].compute_node[3][1].in[0], compute_graph.flexml_layers[226].compute_node[3][1].in[2], compute_graph.flexml_layers[227].compute_node[3][1].in[0], compute_graph.flexml_layers[228].compute_node[3][1].in[0], compute_graph.flexml_layers[229].compute_node[3][1].in[0], compute_graph.flexml_layers[229].compute_node[3][1].in[2], compute_graph.flexml_layers[230].compute_node[3][1].in[0], compute_graph.flexml_layers[230].compute_node[3][1].in[2], compute_graph.flexml_layers[231].compute_node[3][1].in[0], compute_graph.flexml_layers[232].compute_node[3][1].in[0], compute_graph.flexml_layers[233].compute_node[3][1].in[0], compute_graph.flexml_layers[234].compute_node[3][1].in[0], compute_graph.flexml_layers[235].compute_node[3][1].in[0], compute_graph.flexml_layers[235].compute_node[3][1].in[2], compute_graph.flexml_layers[236].compute_node[3][1].in[0]) + +Column 3Row 1Channel 0 + +i17_po0(compute_graph.templated_graph_4.trans_comp_nd[3][1].out[0], compute_graph.flexml_layers[35].compute_node[3][1].out[0], compute_graph.templated_graph_231.mk[3][1].out[0], compute_graph.templated_graph_254.mk[3][1].out[0], compute_graph.templated_graph_258.compute_node[3][1].out[0], compute_graph.templated_graph_259.compute_node[3][1].out[0], compute_graph.templated_graph_260.mk[3][1].out[0], compute_graph.templated_graph_263.compute_node[3][1].out[0], compute_graph.templated_graph_266.compute_node[3][1].out[0], compute_graph.templated_graph_268.mk[3][1].out[0], compute_graph.templated_graph_273.mk[3][1].out[0], compute_graph.templated_graph_274.mk[3][1].out[0], compute_graph.flexml_layers[63].compute_node[3][1].out[0], compute_graph.templated_graph_293.mk[3][1].out[0], compute_graph.templated_graph_298.compute_node[3][1].out[0], compute_graph.templated_graph_300.compute_node[3][1].out[0], compute_graph.flexml_layers[0].compute_node[3][1].out[0], compute_graph.flexml_layers[1].compute_node[3][1].out[0], compute_graph.flexml_layers[2].compute_node[3][1].out[0], compute_graph.flexml_layers[3].compute_node[3][1].out[0], compute_graph.flexml_layers[4].compute_node[3][1].out[0], compute_graph.flexml_layers[5].compute_node[3][1].out[0], compute_graph.flexml_layers[6].compute_node[3][1].out[0], compute_graph.flexml_layers[7].compute_node[3][1].out[0], compute_graph.flexml_layers[8].compute_node[3][1].out[0], compute_graph.flexml_layers[9].compute_node[3][1].out[0], compute_graph.flexml_layers[10].compute_node[3][1].out[0], compute_graph.flexml_layers[11].compute_node[3][1].out[0], compute_graph.flexml_layers[12].compute_node[3][1].out[0], compute_graph.flexml_layers[13].compute_node[3][1].out[0], compute_graph.flexml_layers[14].compute_node[3][1].out[0], compute_graph.flexml_layers[15].compute_node[3][1].out[0], compute_graph.flexml_layers[16].compute_node[3][1].out[0], compute_graph.flexml_layers[17].compute_node[3][1].out[0], compute_graph.flexml_layers[18].compute_node[3][1].out[0], compute_graph.flexml_layers[19].compute_node[3][1].out[0], compute_graph.flexml_layers[20].compute_node[3][1].out[0], compute_graph.flexml_layers[21].compute_node[3][1].out[0], compute_graph.flexml_layers[22].compute_node[3][1].out[0], compute_graph.flexml_layers[23].compute_node[3][1].out[0], compute_graph.flexml_layers[24].compute_node[3][1].out[0], compute_graph.flexml_layers[25].compute_node[3][1].out[0], compute_graph.flexml_layers[26].compute_node[3][1].out[0], compute_graph.flexml_layers[27].compute_node[3][1].out[0], compute_graph.flexml_layers[28].compute_node[3][1].out[0], compute_graph.flexml_layers[29].compute_node[3][1].out[0], compute_graph.flexml_layers[30].compute_node[3][1].out[0], compute_graph.flexml_layers[31].compute_node[3][1].out[0], compute_graph.flexml_layers[32].compute_node[3][1].out[0], compute_graph.flexml_layers[33].compute_node[3][1].out[0], compute_graph.flexml_layers[34].compute_node[3][1].out[0], compute_graph.flexml_layers[143].compute_node[3][1].out[0], compute_graph.flexml_layers[144].compute_node[3][1].out[0], compute_graph.flexml_layers[36].compute_node[3][1].out[0], compute_graph.flexml_layers[37].compute_node[3][1].out[0], compute_graph.flexml_layers[38].compute_node[3][1].out[0], compute_graph.flexml_layers[39].compute_node[3][1].out[0], compute_graph.flexml_layers[40].compute_node[3][1].out[0], compute_graph.flexml_layers[41].compute_node[3][1].out[0], compute_graph.flexml_layers[42].compute_node[3][1].out[0], compute_graph.flexml_layers[43].compute_node[3][1].out[0], compute_graph.flexml_layers[44].compute_node[3][1].out[0], compute_graph.flexml_layers[45].compute_node[3][1].out[0], compute_graph.flexml_layers[46].compute_node[3][1].out[0], compute_graph.flexml_layers[47].compute_node[3][1].out[0], compute_graph.flexml_layers[48].compute_node[3][1].out[0], compute_graph.flexml_layers[49].compute_node[3][1].out[0], compute_graph.flexml_layers[50].compute_node[3][1].out[0], compute_graph.flexml_layers[51].compute_node[3][1].out[0], compute_graph.flexml_layers[52].compute_node[3][1].out[0], compute_graph.flexml_layers[53].compute_node[3][1].out[0], compute_graph.flexml_layers[54].compute_node[3][1].out[0], compute_graph.flexml_layers[55].compute_node[3][1].out[0], compute_graph.flexml_layers[56].compute_node[3][1].out[0], compute_graph.flexml_layers[57].compute_node[3][1].out[0], compute_graph.flexml_layers[58].compute_node[3][1].out[0], compute_graph.flexml_layers[59].compute_node[3][1].out[0], compute_graph.flexml_layers[60].compute_node[3][1].out[0], compute_graph.flexml_layers[61].compute_node[3][1].out[0], compute_graph.flexml_layers[62].compute_node[3][1].out[0], compute_graph.flexml_layers[200].compute_node[3][1].out[0], compute_graph.flexml_layers[201].compute_node[3][1].out[0], compute_graph.flexml_layers[64].compute_node[3][1].out[0], compute_graph.flexml_layers[65].compute_node[3][1].out[0], compute_graph.flexml_layers[66].compute_node[3][1].out[0], compute_graph.flexml_layers[67].compute_node[3][1].out[0], compute_graph.flexml_layers[68].compute_node[3][1].out[0], compute_graph.flexml_layers[69].compute_node[3][1].out[0], compute_graph.flexml_layers[70].compute_node[3][1].out[0], compute_graph.flexml_layers[71].compute_node[3][1].out[0], compute_graph.flexml_layers[72].compute_node[3][1].out[0], compute_graph.flexml_layers[73].compute_node[3][1].out[0], compute_graph.flexml_layers[74].compute_node[3][1].out[0], compute_graph.flexml_layers[75].compute_node[3][1].out[0], compute_graph.flexml_layers[76].compute_node[3][1].out[0], compute_graph.flexml_layers[77].compute_node[3][1].out[0], compute_graph.flexml_layers[78].compute_node[3][1].out[0], compute_graph.flexml_layers[79].compute_node[3][1].out[0], compute_graph.flexml_layers[80].compute_node[3][1].out[0], compute_graph.flexml_layers[81].compute_node[3][1].out[0], compute_graph.flexml_layers[82].compute_node[3][1].out[0], compute_graph.flexml_layers[83].compute_node[3][1].out[0], compute_graph.flexml_layers[84].compute_node[3][1].out[0], compute_graph.flexml_layers[85].compute_node[3][1].out[0], compute_graph.flexml_layers[86].compute_node[3][1].out[0], compute_graph.flexml_layers[87].compute_node[3][1].out[0], compute_graph.flexml_layers[88].compute_node[3][1].out[0], compute_graph.flexml_layers[89].compute_node[3][1].out[0], compute_graph.flexml_layers[90].compute_node[3][1].out[0], compute_graph.flexml_layers[91].compute_node[3][1].out[0], compute_graph.flexml_layers[92].compute_node[3][1].out[0], compute_graph.flexml_layers[93].compute_node[3][1].out[0], compute_graph.flexml_layers[94].compute_node[3][1].out[0], compute_graph.flexml_layers[95].compute_node[3][1].out[0], compute_graph.flexml_layers[96].compute_node[3][1].out[0], compute_graph.flexml_layers[97].compute_node[3][1].out[0], compute_graph.flexml_layers[98].compute_node[3][1].out[0], compute_graph.flexml_layers[99].compute_node[3][1].out[0], compute_graph.flexml_layers[100].compute_node[3][1].out[0], compute_graph.flexml_layers[101].compute_node[3][1].out[0], compute_graph.flexml_layers[102].compute_node[3][1].out[0], compute_graph.flexml_layers[103].compute_node[3][1].out[0], compute_graph.flexml_layers[104].compute_node[3][1].out[0], compute_graph.flexml_layers[105].compute_node[3][1].out[0], compute_graph.flexml_layers[106].compute_node[3][1].out[0], compute_graph.flexml_layers[107].compute_node[3][1].out[0], compute_graph.flexml_layers[108].compute_node[3][1].out[0], compute_graph.flexml_layers[109].compute_node[3][1].out[0], compute_graph.flexml_layers[110].compute_node[3][1].out[0], compute_graph.flexml_layers[111].compute_node[3][1].out[0], compute_graph.flexml_layers[112].compute_node[3][1].out[0], compute_graph.flexml_layers[113].compute_node[3][1].out[0], compute_graph.flexml_layers[114].compute_node[3][1].out[0], compute_graph.flexml_layers[115].compute_node[3][1].out[0], compute_graph.flexml_layers[116].compute_node[3][1].out[0], compute_graph.flexml_layers[117].compute_node[3][1].out[0], compute_graph.flexml_layers[118].compute_node[3][1].out[0], compute_graph.flexml_layers[119].compute_node[3][1].out[0], compute_graph.flexml_layers[120].compute_node[3][1].out[0], compute_graph.flexml_layers[121].compute_node[3][1].out[0], compute_graph.flexml_layers[122].compute_node[3][1].out[0], compute_graph.flexml_layers[123].compute_node[3][1].out[0], compute_graph.flexml_layers[124].compute_node[3][1].out[0], compute_graph.flexml_layers[125].compute_node[3][1].out[0], compute_graph.flexml_layers[126].compute_node[3][1].out[0], compute_graph.flexml_layers[127].compute_node[3][1].out[0], compute_graph.flexml_layers[128].compute_node[3][1].out[0], compute_graph.flexml_layers[129].compute_node[3][1].out[0], compute_graph.flexml_layers[130].compute_node[3][1].out[0], compute_graph.flexml_layers[131].compute_node[3][1].out[0], compute_graph.flexml_layers[132].compute_node[3][1].out[0], compute_graph.flexml_layers[133].compute_node[3][1].out[0], compute_graph.flexml_layers[134].compute_node[3][1].out[0], compute_graph.flexml_layers[135].compute_node[3][1].out[0], compute_graph.flexml_layers[136].compute_node[3][1].out[0], compute_graph.flexml_layers[137].compute_node[3][1].out[0], compute_graph.flexml_layers[138].compute_node[3][1].out[0], compute_graph.flexml_layers[139].compute_node[3][1].out[0], compute_graph.flexml_layers[140].compute_node[3][1].out[0], compute_graph.flexml_layers[141].compute_node[3][1].out[0], compute_graph.flexml_layers[142].compute_node[3][1].out[0], compute_graph.flexml_layers[145].compute_node[3][1].out[0], compute_graph.flexml_layers[146].compute_node[3][1].out[0], compute_graph.flexml_layers[147].compute_node[3][1].out[0], compute_graph.flexml_layers[148].compute_node[3][1].out[0], compute_graph.flexml_layers[149].compute_node[3][1].out[0], compute_graph.flexml_layers[150].compute_node[3][1].out[0], compute_graph.flexml_layers[151].compute_node[3][1].out[0], compute_graph.flexml_layers[152].compute_node[3][1].out[0], compute_graph.flexml_layers[153].compute_node[3][1].out[0], compute_graph.flexml_layers[154].compute_node[3][1].out[0], compute_graph.flexml_layers[155].compute_node[3][1].out[0], compute_graph.flexml_layers[156].compute_node[3][1].out[0], compute_graph.flexml_layers[157].compute_node[3][1].out[0], compute_graph.flexml_layers[158].compute_node[3][1].out[0], compute_graph.flexml_layers[159].compute_node[3][1].out[0], compute_graph.flexml_layers[160].compute_node[3][1].out[0], compute_graph.flexml_layers[161].compute_node[3][1].out[0], compute_graph.flexml_layers[162].compute_node[3][1].out[0], compute_graph.flexml_layers[163].compute_node[3][1].out[0], compute_graph.flexml_layers[164].compute_node[3][1].out[0], compute_graph.flexml_layers[165].compute_node[3][1].out[0], compute_graph.flexml_layers[166].compute_node[3][1].out[0], compute_graph.flexml_layers[167].compute_node[3][1].out[0], compute_graph.flexml_layers[168].compute_node[3][1].out[0], compute_graph.flexml_layers[169].compute_node[3][1].out[0], compute_graph.flexml_layers[170].compute_node[3][1].out[0], compute_graph.flexml_layers[171].compute_node[3][1].out[0], compute_graph.flexml_layers[172].compute_node[3][1].out[0], compute_graph.flexml_layers[173].compute_node[3][1].out[0], compute_graph.flexml_layers[174].compute_node[3][1].out[0], compute_graph.flexml_layers[175].compute_node[3][1].out[0], compute_graph.flexml_layers[176].compute_node[3][1].out[0], compute_graph.flexml_layers[177].compute_node[3][1].out[0], compute_graph.flexml_layers[178].compute_node[3][1].out[0], compute_graph.flexml_layers[179].compute_node[3][1].out[0], compute_graph.flexml_layers[180].compute_node[3][1].out[0], compute_graph.flexml_layers[181].compute_node[3][1].out[0], compute_graph.flexml_layers[182].compute_node[3][1].out[0], compute_graph.flexml_layers[183].compute_node[3][1].out[0], compute_graph.flexml_layers[184].compute_node[3][1].out[0], compute_graph.flexml_layers[185].compute_node[3][1].out[0], compute_graph.flexml_layers[186].compute_node[3][1].out[0], compute_graph.flexml_layers[187].compute_node[3][1].out[0], compute_graph.flexml_layers[188].compute_node[3][1].out[0], compute_graph.flexml_layers[189].compute_node[3][1].out[0], compute_graph.flexml_layers[190].compute_node[3][1].out[0], compute_graph.flexml_layers[191].compute_node[3][1].out[0], compute_graph.flexml_layers[192].compute_node[3][1].out[0], compute_graph.flexml_layers[193].compute_node[3][1].out[0], compute_graph.flexml_layers[194].compute_node[3][1].out[0], compute_graph.flexml_layers[195].compute_node[3][1].out[0], compute_graph.flexml_layers[196].compute_node[3][1].out[0], compute_graph.flexml_layers[197].compute_node[3][1].out[0], compute_graph.flexml_layers[198].compute_node[3][1].out[0], compute_graph.flexml_layers[199].compute_node[3][1].out[0], compute_graph.flexml_layers[202].compute_node[3][1].out[0], compute_graph.flexml_layers[203].compute_node[3][1].out[0], compute_graph.flexml_layers[204].compute_node[3][1].out[0], compute_graph.flexml_layers[205].compute_node[3][1].out[0], compute_graph.flexml_layers[206].compute_node[3][1].out[0], compute_graph.flexml_layers[207].compute_node[3][1].out[0], compute_graph.flexml_layers[208].compute_node[3][1].out[0], compute_graph.flexml_layers[209].compute_node[3][1].out[0], compute_graph.flexml_layers[210].compute_node[3][1].out[0], compute_graph.flexml_layers[211].compute_node[3][1].out[0], compute_graph.flexml_layers[212].compute_node[3][1].out[0], compute_graph.flexml_layers[213].compute_node[3][1].out[0], compute_graph.flexml_layers[214].compute_node[3][1].out[0], compute_graph.flexml_layers[215].compute_node[3][1].out[0], compute_graph.flexml_layers[216].compute_node[3][1].out[0], compute_graph.flexml_layers[217].compute_node[3][1].out[0], compute_graph.flexml_layers[218].compute_node[3][1].out[0], compute_graph.flexml_layers[219].compute_node[3][1].out[0], compute_graph.flexml_layers[220].compute_node[3][1].out[0], compute_graph.flexml_layers[221].compute_node[3][1].out[0], compute_graph.flexml_layers[222].compute_node[3][1].out[0], compute_graph.flexml_layers[223].compute_node[3][1].out[0], compute_graph.flexml_layers[224].compute_node[3][1].out[0], compute_graph.flexml_layers[225].compute_node[3][1].out[0], compute_graph.flexml_layers[226].compute_node[3][1].out[0], compute_graph.flexml_layers[227].compute_node[3][1].out[0], compute_graph.flexml_layers[228].compute_node[3][1].out[0], compute_graph.flexml_layers[229].compute_node[3][1].out[0], compute_graph.flexml_layers[230].compute_node[3][1].out[0], compute_graph.flexml_layers[231].compute_node[3][1].out[0], compute_graph.flexml_layers[232].compute_node[3][1].out[0], compute_graph.flexml_layers[233].compute_node[3][1].out[0], compute_graph.flexml_layers[234].compute_node[3][1].out[0], compute_graph.flexml_layers[235].compute_node[3][1].out[0], compute_graph.flexml_layers[236].compute_node[3][1].out[0]) + +Column 3Row 1Channel 1 + +i17_pi1(compute_graph.templated_graph_260.mk[3][1].in[1], compute_graph.templated_graph_268.mk[3][1].in[1], compute_graph.templated_graph_273.mk[3][1].in[1], compute_graph.flexml_layers[3].compute_node[3][1].in[1], compute_graph.flexml_layers[8].compute_node[3][1].in[1], compute_graph.flexml_layers[9].compute_node[3][1].in[1], compute_graph.flexml_layers[10].compute_node[3][1].in[1], compute_graph.flexml_layers[11].compute_node[3][1].in[1], compute_graph.flexml_layers[12].compute_node[3][1].in[1], compute_graph.flexml_layers[13].compute_node[3][1].in[1], compute_graph.flexml_layers[14].compute_node[3][1].in[1], compute_graph.flexml_layers[15].compute_node[3][1].in[1], compute_graph.flexml_layers[16].compute_node[3][1].in[1], compute_graph.flexml_layers[17].compute_node[3][1].in[1], compute_graph.flexml_layers[19].compute_node[3][1].in[1], compute_graph.flexml_layers[20].compute_node[3][1].in[1], compute_graph.flexml_layers[25].compute_node[3][1].in[1], compute_graph.flexml_layers[26].compute_node[3][1].in[1], compute_graph.flexml_layers[27].compute_node[3][1].in[1], compute_graph.flexml_layers[29].compute_node[3][1].in[1], compute_graph.flexml_layers[30].compute_node[3][1].in[1], compute_graph.flexml_layers[35].compute_node[3][1].in[1], compute_graph.flexml_layers[143].compute_node[3][1].in[1], compute_graph.flexml_layers[144].compute_node[3][1].in[1], compute_graph.flexml_layers[36].compute_node[3][1].in[1], compute_graph.flexml_layers[37].compute_node[3][1].in[1], compute_graph.flexml_layers[39].compute_node[3][1].in[1], compute_graph.flexml_layers[40].compute_node[3][1].in[1], compute_graph.flexml_layers[45].compute_node[3][1].in[1], compute_graph.flexml_layers[46].compute_node[3][1].in[1], compute_graph.flexml_layers[51].compute_node[3][1].in[1], compute_graph.flexml_layers[56].compute_node[3][1].in[1], compute_graph.flexml_layers[57].compute_node[3][1].in[1], compute_graph.flexml_layers[62].compute_node[3][1].in[1], compute_graph.flexml_layers[201].compute_node[3][1].in[1], compute_graph.flexml_layers[67].compute_node[3][1].in[1], compute_graph.flexml_layers[68].compute_node[3][1].in[1], compute_graph.flexml_layers[73].compute_node[3][1].in[1], compute_graph.flexml_layers[78].compute_node[3][1].in[1], compute_graph.flexml_layers[79].compute_node[3][1].in[1], compute_graph.flexml_layers[84].compute_node[3][1].in[1], compute_graph.flexml_layers[89].compute_node[3][1].in[1], compute_graph.flexml_layers[90].compute_node[3][1].in[1], compute_graph.flexml_layers[95].compute_node[3][1].in[1], compute_graph.flexml_layers[101].compute_node[3][1].in[1], compute_graph.flexml_layers[102].compute_node[3][1].in[1], compute_graph.flexml_layers[107].compute_node[3][1].in[1], compute_graph.flexml_layers[108].compute_node[3][1].in[1], compute_graph.flexml_layers[113].compute_node[3][1].in[1], compute_graph.flexml_layers[119].compute_node[3][1].in[1], compute_graph.flexml_layers[120].compute_node[3][1].in[1], compute_graph.flexml_layers[125].compute_node[3][1].in[1], compute_graph.flexml_layers[126].compute_node[3][1].in[1], compute_graph.flexml_layers[131].compute_node[3][1].in[1], compute_graph.flexml_layers[137].compute_node[3][1].in[1], compute_graph.flexml_layers[138].compute_node[3][1].in[1], compute_graph.flexml_layers[149].compute_node[3][1].in[1], compute_graph.flexml_layers[155].compute_node[3][1].in[1], compute_graph.flexml_layers[156].compute_node[3][1].in[1], compute_graph.flexml_layers[161].compute_node[3][1].in[1], compute_graph.flexml_layers[162].compute_node[3][1].in[1], compute_graph.flexml_layers[167].compute_node[3][1].in[1], compute_graph.flexml_layers[173].compute_node[3][1].in[1], compute_graph.flexml_layers[174].compute_node[3][1].in[1], compute_graph.flexml_layers[179].compute_node[3][1].in[1], compute_graph.flexml_layers[180].compute_node[3][1].in[1], compute_graph.flexml_layers[186].compute_node[3][1].in[1], compute_graph.flexml_layers[188].compute_node[3][1].in[1], compute_graph.flexml_layers[189].compute_node[3][1].in[1], compute_graph.flexml_layers[194].compute_node[3][1].in[1], compute_graph.flexml_layers[202].compute_node[3][1].in[1], compute_graph.flexml_layers[207].compute_node[3][1].in[1], compute_graph.flexml_layers[211].compute_node[3][1].in[1], compute_graph.flexml_layers[212].compute_node[3][1].in[1], compute_graph.flexml_layers[217].compute_node[3][1].in[1], compute_graph.flexml_layers[221].compute_node[3][1].in[1], compute_graph.flexml_layers[222].compute_node[3][1].in[1], compute_graph.flexml_layers[227].compute_node[3][1].in[1], compute_graph.flexml_layers[231].compute_node[3][1].in[1], compute_graph.flexml_layers[232].compute_node[3][1].in[1], compute_graph.flexml_layers[233].compute_node[3][1].in[1]) + +Column 3Row 2Channel 0 + +i18_pi0(compute_graph.templated_graph_4.trans_comp_nd[3][2].in[0], compute_graph.flexml_layers[35].compute_node[3][2].in[0], compute_graph.flexml_layers[35].compute_node[3][2].in[2], compute_graph.templated_graph_231.mk[3][2].in[0], compute_graph.templated_graph_254.mk[3][2].in[0], compute_graph.templated_graph_258.compute_node[3][2].in[0], compute_graph.templated_graph_259.compute_node[3][2].in[0], compute_graph.templated_graph_260.mk[3][2].in[0], compute_graph.templated_graph_263.compute_node[3][2].in[0], compute_graph.templated_graph_266.compute_node[3][2].in[0], compute_graph.templated_graph_268.mk[3][2].in[0], compute_graph.templated_graph_273.mk[3][2].in[0], compute_graph.templated_graph_274.mk[3][2].in[0], compute_graph.flexml_layers[62].compute_node[3][2].in[0], compute_graph.flexml_layers[63].compute_node[3][2].in[0], compute_graph.templated_graph_293.mk[3][2].in[0], compute_graph.templated_graph_298.compute_node[3][2].in[0], compute_graph.templated_graph_300.compute_node[3][2].in[0], compute_graph.flexml_layers[0].compute_node[3][2].in[0], compute_graph.flexml_layers[1].compute_node[3][2].in[0], compute_graph.flexml_layers[1].compute_node[3][2].in[2], compute_graph.flexml_layers[2].compute_node[3][2].in[0], compute_graph.flexml_layers[2].compute_node[3][2].in[2], compute_graph.flexml_layers[3].compute_node[3][2].in[0], compute_graph.flexml_layers[4].compute_node[3][2].in[0], compute_graph.flexml_layers[5].compute_node[3][2].in[0], compute_graph.flexml_layers[6].compute_node[3][2].in[0], compute_graph.flexml_layers[7].compute_node[3][2].in[0], compute_graph.flexml_layers[7].compute_node[3][2].in[2], compute_graph.flexml_layers[8].compute_node[3][2].in[0], compute_graph.flexml_layers[9].compute_node[3][2].in[0], compute_graph.flexml_layers[9].compute_node[3][2].in[2], compute_graph.flexml_layers[10].compute_node[3][2].in[0], compute_graph.flexml_layers[11].compute_node[3][2].in[0], compute_graph.flexml_layers[12].compute_node[3][2].in[0], compute_graph.flexml_layers[13].compute_node[3][2].in[0], compute_graph.flexml_layers[14].compute_node[3][2].in[0], compute_graph.flexml_layers[15].compute_node[3][2].in[0], compute_graph.flexml_layers[15].compute_node[3][2].in[2], compute_graph.flexml_layers[16].compute_node[3][2].in[0], compute_graph.flexml_layers[17].compute_node[3][2].in[0], compute_graph.flexml_layers[18].compute_node[3][2].in[0], compute_graph.flexml_layers[19].compute_node[3][2].in[0], compute_graph.flexml_layers[20].compute_node[3][2].in[0], compute_graph.flexml_layers[21].compute_node[3][2].in[0], compute_graph.flexml_layers[22].compute_node[3][2].in[0], compute_graph.flexml_layers[23].compute_node[3][2].in[0], compute_graph.flexml_layers[24].compute_node[3][2].in[0], compute_graph.flexml_layers[24].compute_node[3][2].in[2], compute_graph.flexml_layers[25].compute_node[3][2].in[0], compute_graph.flexml_layers[26].compute_node[3][2].in[0], compute_graph.flexml_layers[27].compute_node[3][2].in[0], compute_graph.flexml_layers[28].compute_node[3][2].in[0], compute_graph.flexml_layers[29].compute_node[3][2].in[0], compute_graph.flexml_layers[30].compute_node[3][2].in[0], compute_graph.flexml_layers[31].compute_node[3][2].in[0], compute_graph.flexml_layers[32].compute_node[3][2].in[0], compute_graph.flexml_layers[33].compute_node[3][2].in[0], compute_graph.flexml_layers[34].compute_node[3][2].in[0], compute_graph.flexml_layers[34].compute_node[3][2].in[2], compute_graph.flexml_layers[143].compute_node[3][2].in[0], compute_graph.flexml_layers[144].compute_node[3][2].in[0], compute_graph.flexml_layers[36].compute_node[3][2].in[0], compute_graph.flexml_layers[37].compute_node[3][2].in[0], compute_graph.flexml_layers[38].compute_node[3][2].in[0], compute_graph.flexml_layers[39].compute_node[3][2].in[0], compute_graph.flexml_layers[40].compute_node[3][2].in[0], compute_graph.flexml_layers[41].compute_node[3][2].in[0], compute_graph.flexml_layers[42].compute_node[3][2].in[0], compute_graph.flexml_layers[43].compute_node[3][2].in[0], compute_graph.flexml_layers[44].compute_node[3][2].in[0], compute_graph.flexml_layers[44].compute_node[3][2].in[2], compute_graph.flexml_layers[45].compute_node[3][2].in[0], compute_graph.flexml_layers[45].compute_node[3][2].in[2], compute_graph.flexml_layers[46].compute_node[3][2].in[0], compute_graph.flexml_layers[47].compute_node[3][2].in[0], compute_graph.flexml_layers[48].compute_node[3][2].in[0], compute_graph.flexml_layers[49].compute_node[3][2].in[0], compute_graph.flexml_layers[50].compute_node[3][2].in[0], compute_graph.flexml_layers[50].compute_node[3][2].in[2], compute_graph.flexml_layers[51].compute_node[3][2].in[0], compute_graph.flexml_layers[52].compute_node[3][2].in[0], compute_graph.flexml_layers[53].compute_node[3][2].in[0], compute_graph.flexml_layers[54].compute_node[3][2].in[0], compute_graph.flexml_layers[55].compute_node[3][2].in[0], compute_graph.flexml_layers[55].compute_node[3][2].in[2], compute_graph.flexml_layers[56].compute_node[3][2].in[0], compute_graph.flexml_layers[57].compute_node[3][2].in[0], compute_graph.flexml_layers[58].compute_node[3][2].in[0], compute_graph.flexml_layers[59].compute_node[3][2].in[0], compute_graph.flexml_layers[60].compute_node[3][2].in[0], compute_graph.flexml_layers[61].compute_node[3][2].in[0], compute_graph.flexml_layers[61].compute_node[3][2].in[2], compute_graph.flexml_layers[200].compute_node[3][2].in[0], compute_graph.flexml_layers[201].compute_node[3][2].in[0], compute_graph.flexml_layers[64].compute_node[3][2].in[0], compute_graph.flexml_layers[65].compute_node[3][2].in[0], compute_graph.flexml_layers[66].compute_node[3][2].in[0], compute_graph.flexml_layers[66].compute_node[3][2].in[2], compute_graph.flexml_layers[67].compute_node[3][2].in[0], compute_graph.flexml_layers[67].compute_node[3][2].in[2], compute_graph.flexml_layers[68].compute_node[3][2].in[0], compute_graph.flexml_layers[69].compute_node[3][2].in[0], compute_graph.flexml_layers[70].compute_node[3][2].in[0], compute_graph.flexml_layers[71].compute_node[3][2].in[0], compute_graph.flexml_layers[72].compute_node[3][2].in[0], compute_graph.flexml_layers[72].compute_node[3][2].in[2], compute_graph.flexml_layers[73].compute_node[3][2].in[0], compute_graph.flexml_layers[74].compute_node[3][2].in[0], compute_graph.flexml_layers[75].compute_node[3][2].in[0], compute_graph.flexml_layers[76].compute_node[3][2].in[0], compute_graph.flexml_layers[77].compute_node[3][2].in[0], compute_graph.flexml_layers[77].compute_node[3][2].in[2], compute_graph.flexml_layers[78].compute_node[3][2].in[0], compute_graph.flexml_layers[78].compute_node[3][2].in[2], compute_graph.flexml_layers[79].compute_node[3][2].in[0], compute_graph.flexml_layers[80].compute_node[3][2].in[0], compute_graph.flexml_layers[81].compute_node[3][2].in[0], compute_graph.flexml_layers[82].compute_node[3][2].in[0], compute_graph.flexml_layers[83].compute_node[3][2].in[0], compute_graph.flexml_layers[83].compute_node[3][2].in[2], compute_graph.flexml_layers[84].compute_node[3][2].in[0], compute_graph.flexml_layers[85].compute_node[3][2].in[0], compute_graph.flexml_layers[86].compute_node[3][2].in[0], compute_graph.flexml_layers[87].compute_node[3][2].in[0], compute_graph.flexml_layers[88].compute_node[3][2].in[0], compute_graph.flexml_layers[88].compute_node[3][2].in[2], compute_graph.flexml_layers[89].compute_node[3][2].in[0], compute_graph.flexml_layers[89].compute_node[3][2].in[2], compute_graph.flexml_layers[90].compute_node[3][2].in[0], compute_graph.flexml_layers[91].compute_node[3][2].in[0], compute_graph.flexml_layers[92].compute_node[3][2].in[0], compute_graph.flexml_layers[93].compute_node[3][2].in[0], compute_graph.flexml_layers[94].compute_node[3][2].in[0], compute_graph.flexml_layers[94].compute_node[3][2].in[2], compute_graph.flexml_layers[95].compute_node[3][2].in[0], compute_graph.flexml_layers[96].compute_node[3][2].in[0], compute_graph.flexml_layers[97].compute_node[3][2].in[0], compute_graph.flexml_layers[98].compute_node[3][2].in[0], compute_graph.flexml_layers[99].compute_node[3][2].in[0], compute_graph.flexml_layers[99].compute_node[3][2].in[2], compute_graph.flexml_layers[100].compute_node[3][2].in[0], compute_graph.flexml_layers[101].compute_node[3][2].in[0], compute_graph.flexml_layers[102].compute_node[3][2].in[0], compute_graph.flexml_layers[103].compute_node[3][2].in[0], compute_graph.flexml_layers[104].compute_node[3][2].in[0], compute_graph.flexml_layers[105].compute_node[3][2].in[0], compute_graph.flexml_layers[106].compute_node[3][2].in[0], compute_graph.flexml_layers[106].compute_node[3][2].in[2], compute_graph.flexml_layers[107].compute_node[3][2].in[0], compute_graph.flexml_layers[108].compute_node[3][2].in[0], compute_graph.flexml_layers[109].compute_node[3][2].in[0], compute_graph.flexml_layers[110].compute_node[3][2].in[0], compute_graph.flexml_layers[111].compute_node[3][2].in[0], compute_graph.flexml_layers[112].compute_node[3][2].in[0], compute_graph.flexml_layers[112].compute_node[3][2].in[2], compute_graph.flexml_layers[113].compute_node[3][2].in[0], compute_graph.flexml_layers[114].compute_node[3][2].in[0], compute_graph.flexml_layers[115].compute_node[3][2].in[0], compute_graph.flexml_layers[116].compute_node[3][2].in[0], compute_graph.flexml_layers[117].compute_node[3][2].in[0], compute_graph.flexml_layers[117].compute_node[3][2].in[2], compute_graph.flexml_layers[118].compute_node[3][2].in[0], compute_graph.flexml_layers[119].compute_node[3][2].in[0], compute_graph.flexml_layers[120].compute_node[3][2].in[0], compute_graph.flexml_layers[121].compute_node[3][2].in[0], compute_graph.flexml_layers[122].compute_node[3][2].in[0], compute_graph.flexml_layers[123].compute_node[3][2].in[0], compute_graph.flexml_layers[124].compute_node[3][2].in[0], compute_graph.flexml_layers[124].compute_node[3][2].in[2], compute_graph.flexml_layers[125].compute_node[3][2].in[0], compute_graph.flexml_layers[125].compute_node[3][2].in[2], compute_graph.flexml_layers[126].compute_node[3][2].in[0], compute_graph.flexml_layers[127].compute_node[3][2].in[0], compute_graph.flexml_layers[128].compute_node[3][2].in[0], compute_graph.flexml_layers[129].compute_node[3][2].in[0], compute_graph.flexml_layers[130].compute_node[3][2].in[0], compute_graph.flexml_layers[130].compute_node[3][2].in[2], compute_graph.flexml_layers[131].compute_node[3][2].in[0], compute_graph.flexml_layers[132].compute_node[3][2].in[0], compute_graph.flexml_layers[133].compute_node[3][2].in[0], compute_graph.flexml_layers[134].compute_node[3][2].in[0], compute_graph.flexml_layers[135].compute_node[3][2].in[0], compute_graph.flexml_layers[135].compute_node[3][2].in[2], compute_graph.flexml_layers[136].compute_node[3][2].in[0], compute_graph.flexml_layers[137].compute_node[3][2].in[0], compute_graph.flexml_layers[138].compute_node[3][2].in[0], compute_graph.flexml_layers[139].compute_node[3][2].in[0], compute_graph.flexml_layers[140].compute_node[3][2].in[0], compute_graph.flexml_layers[141].compute_node[3][2].in[0], compute_graph.flexml_layers[142].compute_node[3][2].in[0], compute_graph.flexml_layers[142].compute_node[3][2].in[2], compute_graph.flexml_layers[145].compute_node[3][2].in[0], compute_graph.flexml_layers[146].compute_node[3][2].in[0], compute_graph.flexml_layers[147].compute_node[3][2].in[0], compute_graph.flexml_layers[148].compute_node[3][2].in[0], compute_graph.flexml_layers[148].compute_node[3][2].in[2], compute_graph.flexml_layers[149].compute_node[3][2].in[0], compute_graph.flexml_layers[150].compute_node[3][2].in[0], compute_graph.flexml_layers[151].compute_node[3][2].in[0], compute_graph.flexml_layers[152].compute_node[3][2].in[0], compute_graph.flexml_layers[153].compute_node[3][2].in[0], compute_graph.flexml_layers[153].compute_node[3][2].in[2], compute_graph.flexml_layers[154].compute_node[3][2].in[0], compute_graph.flexml_layers[155].compute_node[3][2].in[0], compute_graph.flexml_layers[156].compute_node[3][2].in[0], compute_graph.flexml_layers[157].compute_node[3][2].in[0], compute_graph.flexml_layers[158].compute_node[3][2].in[0], compute_graph.flexml_layers[159].compute_node[3][2].in[0], compute_graph.flexml_layers[160].compute_node[3][2].in[0], compute_graph.flexml_layers[160].compute_node[3][2].in[2], compute_graph.flexml_layers[161].compute_node[3][2].in[0], compute_graph.flexml_layers[161].compute_node[3][2].in[2], compute_graph.flexml_layers[162].compute_node[3][2].in[0], compute_graph.flexml_layers[163].compute_node[3][2].in[0], compute_graph.flexml_layers[164].compute_node[3][2].in[0], compute_graph.flexml_layers[165].compute_node[3][2].in[0], compute_graph.flexml_layers[166].compute_node[3][2].in[0], compute_graph.flexml_layers[166].compute_node[3][2].in[2], compute_graph.flexml_layers[167].compute_node[3][2].in[0], compute_graph.flexml_layers[168].compute_node[3][2].in[0], compute_graph.flexml_layers[169].compute_node[3][2].in[0], compute_graph.flexml_layers[170].compute_node[3][2].in[0], compute_graph.flexml_layers[171].compute_node[3][2].in[0], compute_graph.flexml_layers[171].compute_node[3][2].in[2], compute_graph.flexml_layers[172].compute_node[3][2].in[0], compute_graph.flexml_layers[173].compute_node[3][2].in[0], compute_graph.flexml_layers[174].compute_node[3][2].in[0], compute_graph.flexml_layers[175].compute_node[3][2].in[0], compute_graph.flexml_layers[176].compute_node[3][2].in[0], compute_graph.flexml_layers[177].compute_node[3][2].in[0], compute_graph.flexml_layers[178].compute_node[3][2].in[0], compute_graph.flexml_layers[178].compute_node[3][2].in[2], compute_graph.flexml_layers[179].compute_node[3][2].in[0], compute_graph.flexml_layers[179].compute_node[3][2].in[2], compute_graph.flexml_layers[180].compute_node[3][2].in[0], compute_graph.flexml_layers[181].compute_node[3][2].in[0], compute_graph.flexml_layers[182].compute_node[3][2].in[0], compute_graph.flexml_layers[183].compute_node[3][2].in[0], compute_graph.flexml_layers[184].compute_node[3][2].in[0], compute_graph.flexml_layers[184].compute_node[3][2].in[2], compute_graph.flexml_layers[185].compute_node[3][2].in[0], compute_graph.flexml_layers[186].compute_node[3][2].in[0], compute_graph.flexml_layers[187].compute_node[3][2].in[0], compute_graph.flexml_layers[188].compute_node[3][2].in[0], compute_graph.flexml_layers[188].compute_node[3][2].in[2], compute_graph.flexml_layers[189].compute_node[3][2].in[0], compute_graph.flexml_layers[190].compute_node[3][2].in[0], compute_graph.flexml_layers[191].compute_node[3][2].in[0], compute_graph.flexml_layers[191].compute_node[3][2].in[2], compute_graph.flexml_layers[192].compute_node[3][2].in[0], compute_graph.flexml_layers[192].compute_node[3][2].in[2], compute_graph.flexml_layers[193].compute_node[3][2].in[0], compute_graph.flexml_layers[193].compute_node[3][2].in[2], compute_graph.flexml_layers[194].compute_node[3][2].in[0], compute_graph.flexml_layers[195].compute_node[3][2].in[0], compute_graph.flexml_layers[196].compute_node[3][2].in[0], compute_graph.flexml_layers[196].compute_node[3][2].in[2], compute_graph.flexml_layers[197].compute_node[3][2].in[0], compute_graph.flexml_layers[197].compute_node[3][2].in[2], compute_graph.flexml_layers[198].compute_node[3][2].in[0], compute_graph.flexml_layers[199].compute_node[3][2].in[0], compute_graph.flexml_layers[202].compute_node[3][2].in[0], compute_graph.flexml_layers[203].compute_node[3][2].in[0], compute_graph.flexml_layers[204].compute_node[3][2].in[0], compute_graph.flexml_layers[204].compute_node[3][2].in[2], compute_graph.flexml_layers[205].compute_node[3][2].in[0], compute_graph.flexml_layers[205].compute_node[3][2].in[2], compute_graph.flexml_layers[206].compute_node[3][2].in[0], compute_graph.flexml_layers[206].compute_node[3][2].in[2], compute_graph.flexml_layers[207].compute_node[3][2].in[0], compute_graph.flexml_layers[208].compute_node[3][2].in[0], compute_graph.flexml_layers[209].compute_node[3][2].in[0], compute_graph.flexml_layers[209].compute_node[3][2].in[2], compute_graph.flexml_layers[210].compute_node[3][2].in[0], compute_graph.flexml_layers[210].compute_node[3][2].in[2], compute_graph.flexml_layers[211].compute_node[3][2].in[0], compute_graph.flexml_layers[212].compute_node[3][2].in[0], compute_graph.flexml_layers[213].compute_node[3][2].in[0], compute_graph.flexml_layers[214].compute_node[3][2].in[0], compute_graph.flexml_layers[214].compute_node[3][2].in[2], compute_graph.flexml_layers[215].compute_node[3][2].in[0], compute_graph.flexml_layers[215].compute_node[3][2].in[2], compute_graph.flexml_layers[216].compute_node[3][2].in[0], compute_graph.flexml_layers[216].compute_node[3][2].in[2], compute_graph.flexml_layers[217].compute_node[3][2].in[0], compute_graph.flexml_layers[218].compute_node[3][2].in[0], compute_graph.flexml_layers[219].compute_node[3][2].in[0], compute_graph.flexml_layers[219].compute_node[3][2].in[2], compute_graph.flexml_layers[220].compute_node[3][2].in[0], compute_graph.flexml_layers[220].compute_node[3][2].in[2], compute_graph.flexml_layers[221].compute_node[3][2].in[0], compute_graph.flexml_layers[222].compute_node[3][2].in[0], compute_graph.flexml_layers[223].compute_node[3][2].in[0], compute_graph.flexml_layers[224].compute_node[3][2].in[0], compute_graph.flexml_layers[224].compute_node[3][2].in[2], compute_graph.flexml_layers[225].compute_node[3][2].in[0], compute_graph.flexml_layers[225].compute_node[3][2].in[2], compute_graph.flexml_layers[226].compute_node[3][2].in[0], compute_graph.flexml_layers[226].compute_node[3][2].in[2], compute_graph.flexml_layers[227].compute_node[3][2].in[0], compute_graph.flexml_layers[228].compute_node[3][2].in[0], compute_graph.flexml_layers[229].compute_node[3][2].in[0], compute_graph.flexml_layers[229].compute_node[3][2].in[2], compute_graph.flexml_layers[230].compute_node[3][2].in[0], compute_graph.flexml_layers[230].compute_node[3][2].in[2], compute_graph.flexml_layers[231].compute_node[3][2].in[0], compute_graph.flexml_layers[232].compute_node[3][2].in[0], compute_graph.flexml_layers[233].compute_node[3][2].in[0], compute_graph.flexml_layers[234].compute_node[3][2].in[0], compute_graph.flexml_layers[235].compute_node[3][2].in[0], compute_graph.flexml_layers[235].compute_node[3][2].in[2], compute_graph.flexml_layers[236].compute_node[3][2].in[0]) + +Column 3Row 2Channel 0 + +i18_po0(compute_graph.templated_graph_4.trans_comp_nd[3][2].out[0], compute_graph.flexml_layers[35].compute_node[3][2].out[0], compute_graph.templated_graph_231.mk[3][2].out[0], compute_graph.templated_graph_254.mk[3][2].out[0], compute_graph.templated_graph_258.compute_node[3][2].out[0], compute_graph.templated_graph_259.compute_node[3][2].out[0], compute_graph.templated_graph_260.mk[3][2].out[0], compute_graph.templated_graph_263.compute_node[3][2].out[0], compute_graph.templated_graph_266.compute_node[3][2].out[0], compute_graph.templated_graph_268.mk[3][2].out[0], compute_graph.templated_graph_273.mk[3][2].out[0], compute_graph.templated_graph_274.mk[3][2].out[0], compute_graph.flexml_layers[62].compute_node[3][2].out[0], compute_graph.flexml_layers[63].compute_node[3][2].out[0], compute_graph.templated_graph_293.mk[3][2].out[0], compute_graph.templated_graph_298.compute_node[3][2].out[0], compute_graph.templated_graph_300.compute_node[3][2].out[0], compute_graph.flexml_layers[0].compute_node[3][2].out[0], compute_graph.flexml_layers[1].compute_node[3][2].out[0], compute_graph.flexml_layers[2].compute_node[3][2].out[0], compute_graph.flexml_layers[3].compute_node[3][2].out[0], compute_graph.flexml_layers[4].compute_node[3][2].out[0], compute_graph.flexml_layers[5].compute_node[3][2].out[0], compute_graph.flexml_layers[6].compute_node[3][2].out[0], compute_graph.flexml_layers[7].compute_node[3][2].out[0], compute_graph.flexml_layers[8].compute_node[3][2].out[0], compute_graph.flexml_layers[9].compute_node[3][2].out[0], compute_graph.flexml_layers[10].compute_node[3][2].out[0], compute_graph.flexml_layers[11].compute_node[3][2].out[0], compute_graph.flexml_layers[12].compute_node[3][2].out[0], compute_graph.flexml_layers[13].compute_node[3][2].out[0], compute_graph.flexml_layers[14].compute_node[3][2].out[0], compute_graph.flexml_layers[15].compute_node[3][2].out[0], compute_graph.flexml_layers[16].compute_node[3][2].out[0], compute_graph.flexml_layers[17].compute_node[3][2].out[0], compute_graph.flexml_layers[18].compute_node[3][2].out[0], compute_graph.flexml_layers[19].compute_node[3][2].out[0], compute_graph.flexml_layers[20].compute_node[3][2].out[0], compute_graph.flexml_layers[21].compute_node[3][2].out[0], compute_graph.flexml_layers[22].compute_node[3][2].out[0], compute_graph.flexml_layers[23].compute_node[3][2].out[0], compute_graph.flexml_layers[24].compute_node[3][2].out[0], compute_graph.flexml_layers[25].compute_node[3][2].out[0], compute_graph.flexml_layers[26].compute_node[3][2].out[0], compute_graph.flexml_layers[27].compute_node[3][2].out[0], compute_graph.flexml_layers[28].compute_node[3][2].out[0], compute_graph.flexml_layers[29].compute_node[3][2].out[0], compute_graph.flexml_layers[30].compute_node[3][2].out[0], compute_graph.flexml_layers[31].compute_node[3][2].out[0], compute_graph.flexml_layers[32].compute_node[3][2].out[0], compute_graph.flexml_layers[33].compute_node[3][2].out[0], compute_graph.flexml_layers[34].compute_node[3][2].out[0], compute_graph.flexml_layers[143].compute_node[3][2].out[0], compute_graph.flexml_layers[144].compute_node[3][2].out[0], compute_graph.flexml_layers[36].compute_node[3][2].out[0], compute_graph.flexml_layers[37].compute_node[3][2].out[0], compute_graph.flexml_layers[38].compute_node[3][2].out[0], compute_graph.flexml_layers[39].compute_node[3][2].out[0], compute_graph.flexml_layers[40].compute_node[3][2].out[0], compute_graph.flexml_layers[41].compute_node[3][2].out[0], compute_graph.flexml_layers[42].compute_node[3][2].out[0], compute_graph.flexml_layers[43].compute_node[3][2].out[0], compute_graph.flexml_layers[44].compute_node[3][2].out[0], compute_graph.flexml_layers[45].compute_node[3][2].out[0], compute_graph.flexml_layers[46].compute_node[3][2].out[0], compute_graph.flexml_layers[47].compute_node[3][2].out[0], compute_graph.flexml_layers[48].compute_node[3][2].out[0], compute_graph.flexml_layers[49].compute_node[3][2].out[0], compute_graph.flexml_layers[50].compute_node[3][2].out[0], compute_graph.flexml_layers[51].compute_node[3][2].out[0], compute_graph.flexml_layers[52].compute_node[3][2].out[0], compute_graph.flexml_layers[53].compute_node[3][2].out[0], compute_graph.flexml_layers[54].compute_node[3][2].out[0], compute_graph.flexml_layers[55].compute_node[3][2].out[0], compute_graph.flexml_layers[56].compute_node[3][2].out[0], compute_graph.flexml_layers[57].compute_node[3][2].out[0], compute_graph.flexml_layers[58].compute_node[3][2].out[0], compute_graph.flexml_layers[59].compute_node[3][2].out[0], compute_graph.flexml_layers[60].compute_node[3][2].out[0], compute_graph.flexml_layers[61].compute_node[3][2].out[0], compute_graph.flexml_layers[200].compute_node[3][2].out[0], compute_graph.flexml_layers[201].compute_node[3][2].out[0], compute_graph.flexml_layers[64].compute_node[3][2].out[0], compute_graph.flexml_layers[65].compute_node[3][2].out[0], compute_graph.flexml_layers[66].compute_node[3][2].out[0], compute_graph.flexml_layers[67].compute_node[3][2].out[0], compute_graph.flexml_layers[68].compute_node[3][2].out[0], compute_graph.flexml_layers[69].compute_node[3][2].out[0], compute_graph.flexml_layers[70].compute_node[3][2].out[0], compute_graph.flexml_layers[71].compute_node[3][2].out[0], compute_graph.flexml_layers[72].compute_node[3][2].out[0], compute_graph.flexml_layers[73].compute_node[3][2].out[0], compute_graph.flexml_layers[74].compute_node[3][2].out[0], compute_graph.flexml_layers[75].compute_node[3][2].out[0], compute_graph.flexml_layers[76].compute_node[3][2].out[0], compute_graph.flexml_layers[77].compute_node[3][2].out[0], compute_graph.flexml_layers[78].compute_node[3][2].out[0], compute_graph.flexml_layers[79].compute_node[3][2].out[0], compute_graph.flexml_layers[80].compute_node[3][2].out[0], compute_graph.flexml_layers[81].compute_node[3][2].out[0], compute_graph.flexml_layers[82].compute_node[3][2].out[0], compute_graph.flexml_layers[83].compute_node[3][2].out[0], compute_graph.flexml_layers[84].compute_node[3][2].out[0], compute_graph.flexml_layers[85].compute_node[3][2].out[0], compute_graph.flexml_layers[86].compute_node[3][2].out[0], compute_graph.flexml_layers[87].compute_node[3][2].out[0], compute_graph.flexml_layers[88].compute_node[3][2].out[0], compute_graph.flexml_layers[89].compute_node[3][2].out[0], compute_graph.flexml_layers[90].compute_node[3][2].out[0], compute_graph.flexml_layers[91].compute_node[3][2].out[0], compute_graph.flexml_layers[92].compute_node[3][2].out[0], compute_graph.flexml_layers[93].compute_node[3][2].out[0], compute_graph.flexml_layers[94].compute_node[3][2].out[0], compute_graph.flexml_layers[95].compute_node[3][2].out[0], compute_graph.flexml_layers[96].compute_node[3][2].out[0], compute_graph.flexml_layers[97].compute_node[3][2].out[0], compute_graph.flexml_layers[98].compute_node[3][2].out[0], compute_graph.flexml_layers[99].compute_node[3][2].out[0], compute_graph.flexml_layers[100].compute_node[3][2].out[0], compute_graph.flexml_layers[101].compute_node[3][2].out[0], compute_graph.flexml_layers[102].compute_node[3][2].out[0], compute_graph.flexml_layers[103].compute_node[3][2].out[0], compute_graph.flexml_layers[104].compute_node[3][2].out[0], compute_graph.flexml_layers[105].compute_node[3][2].out[0], compute_graph.flexml_layers[106].compute_node[3][2].out[0], compute_graph.flexml_layers[107].compute_node[3][2].out[0], compute_graph.flexml_layers[108].compute_node[3][2].out[0], compute_graph.flexml_layers[109].compute_node[3][2].out[0], compute_graph.flexml_layers[110].compute_node[3][2].out[0], compute_graph.flexml_layers[111].compute_node[3][2].out[0], compute_graph.flexml_layers[112].compute_node[3][2].out[0], compute_graph.flexml_layers[113].compute_node[3][2].out[0], compute_graph.flexml_layers[114].compute_node[3][2].out[0], compute_graph.flexml_layers[115].compute_node[3][2].out[0], compute_graph.flexml_layers[116].compute_node[3][2].out[0], compute_graph.flexml_layers[117].compute_node[3][2].out[0], compute_graph.flexml_layers[118].compute_node[3][2].out[0], compute_graph.flexml_layers[119].compute_node[3][2].out[0], compute_graph.flexml_layers[120].compute_node[3][2].out[0], compute_graph.flexml_layers[121].compute_node[3][2].out[0], compute_graph.flexml_layers[122].compute_node[3][2].out[0], compute_graph.flexml_layers[123].compute_node[3][2].out[0], compute_graph.flexml_layers[124].compute_node[3][2].out[0], compute_graph.flexml_layers[125].compute_node[3][2].out[0], compute_graph.flexml_layers[126].compute_node[3][2].out[0], compute_graph.flexml_layers[127].compute_node[3][2].out[0], compute_graph.flexml_layers[128].compute_node[3][2].out[0], compute_graph.flexml_layers[129].compute_node[3][2].out[0], compute_graph.flexml_layers[130].compute_node[3][2].out[0], compute_graph.flexml_layers[131].compute_node[3][2].out[0], compute_graph.flexml_layers[132].compute_node[3][2].out[0], compute_graph.flexml_layers[133].compute_node[3][2].out[0], compute_graph.flexml_layers[134].compute_node[3][2].out[0], compute_graph.flexml_layers[135].compute_node[3][2].out[0], compute_graph.flexml_layers[136].compute_node[3][2].out[0], compute_graph.flexml_layers[137].compute_node[3][2].out[0], compute_graph.flexml_layers[138].compute_node[3][2].out[0], compute_graph.flexml_layers[139].compute_node[3][2].out[0], compute_graph.flexml_layers[140].compute_node[3][2].out[0], compute_graph.flexml_layers[141].compute_node[3][2].out[0], compute_graph.flexml_layers[142].compute_node[3][2].out[0], compute_graph.flexml_layers[145].compute_node[3][2].out[0], compute_graph.flexml_layers[146].compute_node[3][2].out[0], compute_graph.flexml_layers[147].compute_node[3][2].out[0], compute_graph.flexml_layers[148].compute_node[3][2].out[0], compute_graph.flexml_layers[149].compute_node[3][2].out[0], compute_graph.flexml_layers[150].compute_node[3][2].out[0], compute_graph.flexml_layers[151].compute_node[3][2].out[0], compute_graph.flexml_layers[152].compute_node[3][2].out[0], compute_graph.flexml_layers[153].compute_node[3][2].out[0], compute_graph.flexml_layers[154].compute_node[3][2].out[0], compute_graph.flexml_layers[155].compute_node[3][2].out[0], compute_graph.flexml_layers[156].compute_node[3][2].out[0], compute_graph.flexml_layers[157].compute_node[3][2].out[0], compute_graph.flexml_layers[158].compute_node[3][2].out[0], compute_graph.flexml_layers[159].compute_node[3][2].out[0], compute_graph.flexml_layers[160].compute_node[3][2].out[0], compute_graph.flexml_layers[161].compute_node[3][2].out[0], compute_graph.flexml_layers[162].compute_node[3][2].out[0], compute_graph.flexml_layers[163].compute_node[3][2].out[0], compute_graph.flexml_layers[164].compute_node[3][2].out[0], compute_graph.flexml_layers[165].compute_node[3][2].out[0], compute_graph.flexml_layers[166].compute_node[3][2].out[0], compute_graph.flexml_layers[167].compute_node[3][2].out[0], compute_graph.flexml_layers[168].compute_node[3][2].out[0], compute_graph.flexml_layers[169].compute_node[3][2].out[0], compute_graph.flexml_layers[170].compute_node[3][2].out[0], compute_graph.flexml_layers[171].compute_node[3][2].out[0], compute_graph.flexml_layers[172].compute_node[3][2].out[0], compute_graph.flexml_layers[173].compute_node[3][2].out[0], compute_graph.flexml_layers[174].compute_node[3][2].out[0], compute_graph.flexml_layers[175].compute_node[3][2].out[0], compute_graph.flexml_layers[176].compute_node[3][2].out[0], compute_graph.flexml_layers[177].compute_node[3][2].out[0], compute_graph.flexml_layers[178].compute_node[3][2].out[0], compute_graph.flexml_layers[179].compute_node[3][2].out[0], compute_graph.flexml_layers[180].compute_node[3][2].out[0], compute_graph.flexml_layers[181].compute_node[3][2].out[0], compute_graph.flexml_layers[182].compute_node[3][2].out[0], compute_graph.flexml_layers[183].compute_node[3][2].out[0], compute_graph.flexml_layers[184].compute_node[3][2].out[0], compute_graph.flexml_layers[185].compute_node[3][2].out[0], compute_graph.flexml_layers[186].compute_node[3][2].out[0], compute_graph.flexml_layers[187].compute_node[3][2].out[0], compute_graph.flexml_layers[188].compute_node[3][2].out[0], compute_graph.flexml_layers[189].compute_node[3][2].out[0], compute_graph.flexml_layers[190].compute_node[3][2].out[0], compute_graph.flexml_layers[191].compute_node[3][2].out[0], compute_graph.flexml_layers[192].compute_node[3][2].out[0], compute_graph.flexml_layers[193].compute_node[3][2].out[0], compute_graph.flexml_layers[194].compute_node[3][2].out[0], compute_graph.flexml_layers[195].compute_node[3][2].out[0], compute_graph.flexml_layers[196].compute_node[3][2].out[0], compute_graph.flexml_layers[197].compute_node[3][2].out[0], compute_graph.flexml_layers[198].compute_node[3][2].out[0], compute_graph.flexml_layers[199].compute_node[3][2].out[0], compute_graph.flexml_layers[202].compute_node[3][2].out[0], compute_graph.flexml_layers[203].compute_node[3][2].out[0], compute_graph.flexml_layers[204].compute_node[3][2].out[0], compute_graph.flexml_layers[205].compute_node[3][2].out[0], compute_graph.flexml_layers[206].compute_node[3][2].out[0], compute_graph.flexml_layers[207].compute_node[3][2].out[0], compute_graph.flexml_layers[208].compute_node[3][2].out[0], compute_graph.flexml_layers[209].compute_node[3][2].out[0], compute_graph.flexml_layers[210].compute_node[3][2].out[0], compute_graph.flexml_layers[211].compute_node[3][2].out[0], compute_graph.flexml_layers[212].compute_node[3][2].out[0], compute_graph.flexml_layers[213].compute_node[3][2].out[0], compute_graph.flexml_layers[214].compute_node[3][2].out[0], compute_graph.flexml_layers[215].compute_node[3][2].out[0], compute_graph.flexml_layers[216].compute_node[3][2].out[0], compute_graph.flexml_layers[217].compute_node[3][2].out[0], compute_graph.flexml_layers[218].compute_node[3][2].out[0], compute_graph.flexml_layers[219].compute_node[3][2].out[0], compute_graph.flexml_layers[220].compute_node[3][2].out[0], compute_graph.flexml_layers[221].compute_node[3][2].out[0], compute_graph.flexml_layers[222].compute_node[3][2].out[0], compute_graph.flexml_layers[223].compute_node[3][2].out[0], compute_graph.flexml_layers[224].compute_node[3][2].out[0], compute_graph.flexml_layers[225].compute_node[3][2].out[0], compute_graph.flexml_layers[226].compute_node[3][2].out[0], compute_graph.flexml_layers[227].compute_node[3][2].out[0], compute_graph.flexml_layers[228].compute_node[3][2].out[0], compute_graph.flexml_layers[229].compute_node[3][2].out[0], compute_graph.flexml_layers[230].compute_node[3][2].out[0], compute_graph.flexml_layers[231].compute_node[3][2].out[0], compute_graph.flexml_layers[232].compute_node[3][2].out[0], compute_graph.flexml_layers[233].compute_node[3][2].out[0], compute_graph.flexml_layers[234].compute_node[3][2].out[0], compute_graph.flexml_layers[235].compute_node[3][2].out[0], compute_graph.flexml_layers[236].compute_node[3][2].out[0]) + +Column 3Row 2Channel 1 + +i18_pi1(compute_graph.flexml_layers[35].compute_node[3][2].in[1], compute_graph.templated_graph_260.mk[3][2].in[1], compute_graph.templated_graph_268.mk[3][2].in[1], compute_graph.templated_graph_273.mk[3][2].in[1], compute_graph.flexml_layers[62].compute_node[3][2].in[1], compute_graph.flexml_layers[3].compute_node[3][2].in[1], compute_graph.flexml_layers[8].compute_node[3][2].in[1], compute_graph.flexml_layers[9].compute_node[3][2].in[1], compute_graph.flexml_layers[10].compute_node[3][2].in[1], compute_graph.flexml_layers[11].compute_node[3][2].in[1], compute_graph.flexml_layers[12].compute_node[3][2].in[1], compute_graph.flexml_layers[13].compute_node[3][2].in[1], compute_graph.flexml_layers[14].compute_node[3][2].in[1], compute_graph.flexml_layers[15].compute_node[3][2].in[1], compute_graph.flexml_layers[16].compute_node[3][2].in[1], compute_graph.flexml_layers[17].compute_node[3][2].in[1], compute_graph.flexml_layers[19].compute_node[3][2].in[1], compute_graph.flexml_layers[20].compute_node[3][2].in[1], compute_graph.flexml_layers[25].compute_node[3][2].in[1], compute_graph.flexml_layers[26].compute_node[3][2].in[1], compute_graph.flexml_layers[27].compute_node[3][2].in[1], compute_graph.flexml_layers[29].compute_node[3][2].in[1], compute_graph.flexml_layers[30].compute_node[3][2].in[1], compute_graph.flexml_layers[143].compute_node[3][2].in[1], compute_graph.flexml_layers[144].compute_node[3][2].in[1], compute_graph.flexml_layers[36].compute_node[3][2].in[1], compute_graph.flexml_layers[37].compute_node[3][2].in[1], compute_graph.flexml_layers[39].compute_node[3][2].in[1], compute_graph.flexml_layers[40].compute_node[3][2].in[1], compute_graph.flexml_layers[45].compute_node[3][2].in[1], compute_graph.flexml_layers[46].compute_node[3][2].in[1], compute_graph.flexml_layers[51].compute_node[3][2].in[1], compute_graph.flexml_layers[56].compute_node[3][2].in[1], compute_graph.flexml_layers[57].compute_node[3][2].in[1], compute_graph.flexml_layers[201].compute_node[3][2].in[1], compute_graph.flexml_layers[67].compute_node[3][2].in[1], compute_graph.flexml_layers[68].compute_node[3][2].in[1], compute_graph.flexml_layers[73].compute_node[3][2].in[1], compute_graph.flexml_layers[78].compute_node[3][2].in[1], compute_graph.flexml_layers[79].compute_node[3][2].in[1], compute_graph.flexml_layers[84].compute_node[3][2].in[1], compute_graph.flexml_layers[89].compute_node[3][2].in[1], compute_graph.flexml_layers[90].compute_node[3][2].in[1], compute_graph.flexml_layers[95].compute_node[3][2].in[1], compute_graph.flexml_layers[101].compute_node[3][2].in[1], compute_graph.flexml_layers[102].compute_node[3][2].in[1], compute_graph.flexml_layers[107].compute_node[3][2].in[1], compute_graph.flexml_layers[108].compute_node[3][2].in[1], compute_graph.flexml_layers[113].compute_node[3][2].in[1], compute_graph.flexml_layers[119].compute_node[3][2].in[1], compute_graph.flexml_layers[120].compute_node[3][2].in[1], compute_graph.flexml_layers[125].compute_node[3][2].in[1], compute_graph.flexml_layers[126].compute_node[3][2].in[1], compute_graph.flexml_layers[131].compute_node[3][2].in[1], compute_graph.flexml_layers[137].compute_node[3][2].in[1], compute_graph.flexml_layers[138].compute_node[3][2].in[1], compute_graph.flexml_layers[149].compute_node[3][2].in[1], compute_graph.flexml_layers[155].compute_node[3][2].in[1], compute_graph.flexml_layers[156].compute_node[3][2].in[1], compute_graph.flexml_layers[161].compute_node[3][2].in[1], compute_graph.flexml_layers[162].compute_node[3][2].in[1], compute_graph.flexml_layers[167].compute_node[3][2].in[1], compute_graph.flexml_layers[173].compute_node[3][2].in[1], compute_graph.flexml_layers[174].compute_node[3][2].in[1], compute_graph.flexml_layers[179].compute_node[3][2].in[1], compute_graph.flexml_layers[180].compute_node[3][2].in[1], compute_graph.flexml_layers[186].compute_node[3][2].in[1], compute_graph.flexml_layers[188].compute_node[3][2].in[1], compute_graph.flexml_layers[189].compute_node[3][2].in[1], compute_graph.flexml_layers[194].compute_node[3][2].in[1], compute_graph.flexml_layers[202].compute_node[3][2].in[1], compute_graph.flexml_layers[207].compute_node[3][2].in[1], compute_graph.flexml_layers[211].compute_node[3][2].in[1], compute_graph.flexml_layers[212].compute_node[3][2].in[1], compute_graph.flexml_layers[217].compute_node[3][2].in[1], compute_graph.flexml_layers[221].compute_node[3][2].in[1], compute_graph.flexml_layers[222].compute_node[3][2].in[1], compute_graph.flexml_layers[227].compute_node[3][2].in[1], compute_graph.flexml_layers[231].compute_node[3][2].in[1], compute_graph.flexml_layers[232].compute_node[3][2].in[1], compute_graph.flexml_layers[233].compute_node[3][2].in[1]) + +Column 3Row 3Channel 0 + +i19_pi0(compute_graph.templated_graph_4.trans_comp_nd[3][3].in[0], compute_graph.flexml_layers[35].compute_node[3][3].in[0], compute_graph.flexml_layers[35].compute_node[3][3].in[2], compute_graph.templated_graph_231.mk[3][3].in[0], compute_graph.templated_graph_254.mk[3][3].in[0], compute_graph.templated_graph_258.compute_node[3][3].in[0], compute_graph.templated_graph_259.compute_node[3][3].in[0], compute_graph.templated_graph_260.mk[3][3].in[0], compute_graph.templated_graph_263.compute_node[3][3].in[0], compute_graph.templated_graph_266.compute_node[3][3].in[0], compute_graph.templated_graph_268.mk[3][3].in[0], compute_graph.templated_graph_273.mk[3][3].in[0], compute_graph.templated_graph_274.mk[3][3].in[0], compute_graph.flexml_layers[62].compute_node[3][3].in[0], compute_graph.flexml_layers[63].compute_node[3][3].in[0], compute_graph.templated_graph_293.mk[3][3].in[0], compute_graph.templated_graph_298.compute_node[3][3].in[0], compute_graph.templated_graph_300.compute_node[3][3].in[0], compute_graph.flexml_layers[0].compute_node[3][3].in[0], compute_graph.flexml_layers[1].compute_node[3][3].in[0], compute_graph.flexml_layers[1].compute_node[3][3].in[2], compute_graph.flexml_layers[2].compute_node[3][3].in[0], compute_graph.flexml_layers[2].compute_node[3][3].in[2], compute_graph.flexml_layers[3].compute_node[3][3].in[0], compute_graph.flexml_layers[4].compute_node[3][3].in[0], compute_graph.flexml_layers[5].compute_node[3][3].in[0], compute_graph.flexml_layers[6].compute_node[3][3].in[0], compute_graph.flexml_layers[7].compute_node[3][3].in[0], compute_graph.flexml_layers[7].compute_node[3][3].in[2], compute_graph.flexml_layers[8].compute_node[3][3].in[0], compute_graph.flexml_layers[9].compute_node[3][3].in[0], compute_graph.flexml_layers[9].compute_node[3][3].in[2], compute_graph.flexml_layers[10].compute_node[3][3].in[0], compute_graph.flexml_layers[11].compute_node[3][3].in[0], compute_graph.flexml_layers[12].compute_node[3][3].in[0], compute_graph.flexml_layers[13].compute_node[3][3].in[0], compute_graph.flexml_layers[14].compute_node[3][3].in[0], compute_graph.flexml_layers[15].compute_node[3][3].in[0], compute_graph.flexml_layers[15].compute_node[3][3].in[2], compute_graph.flexml_layers[16].compute_node[3][3].in[0], compute_graph.flexml_layers[17].compute_node[3][3].in[0], compute_graph.flexml_layers[18].compute_node[3][3].in[0], compute_graph.flexml_layers[19].compute_node[3][3].in[0], compute_graph.flexml_layers[20].compute_node[3][3].in[0], compute_graph.flexml_layers[21].compute_node[3][3].in[0], compute_graph.flexml_layers[22].compute_node[3][3].in[0], compute_graph.flexml_layers[23].compute_node[3][3].in[0], compute_graph.flexml_layers[24].compute_node[3][3].in[0], compute_graph.flexml_layers[24].compute_node[3][3].in[2], compute_graph.flexml_layers[25].compute_node[3][3].in[0], compute_graph.flexml_layers[26].compute_node[3][3].in[0], compute_graph.flexml_layers[27].compute_node[3][3].in[0], compute_graph.flexml_layers[28].compute_node[3][3].in[0], compute_graph.flexml_layers[29].compute_node[3][3].in[0], compute_graph.flexml_layers[30].compute_node[3][3].in[0], compute_graph.flexml_layers[31].compute_node[3][3].in[0], compute_graph.flexml_layers[32].compute_node[3][3].in[0], compute_graph.flexml_layers[33].compute_node[3][3].in[0], compute_graph.flexml_layers[34].compute_node[3][3].in[0], compute_graph.flexml_layers[34].compute_node[3][3].in[2], compute_graph.flexml_layers[143].compute_node[3][3].in[0], compute_graph.flexml_layers[144].compute_node[3][3].in[0], compute_graph.flexml_layers[36].compute_node[3][3].in[0], compute_graph.flexml_layers[37].compute_node[3][3].in[0], compute_graph.flexml_layers[38].compute_node[3][3].in[0], compute_graph.flexml_layers[39].compute_node[3][3].in[0], compute_graph.flexml_layers[40].compute_node[3][3].in[0], compute_graph.flexml_layers[41].compute_node[3][3].in[0], compute_graph.flexml_layers[42].compute_node[3][3].in[0], compute_graph.flexml_layers[43].compute_node[3][3].in[0], compute_graph.flexml_layers[44].compute_node[3][3].in[0], compute_graph.flexml_layers[44].compute_node[3][3].in[2], compute_graph.flexml_layers[45].compute_node[3][3].in[0], compute_graph.flexml_layers[45].compute_node[3][3].in[2], compute_graph.flexml_layers[46].compute_node[3][3].in[0], compute_graph.flexml_layers[47].compute_node[3][3].in[0], compute_graph.flexml_layers[48].compute_node[3][3].in[0], compute_graph.flexml_layers[49].compute_node[3][3].in[0], compute_graph.flexml_layers[50].compute_node[3][3].in[0], compute_graph.flexml_layers[50].compute_node[3][3].in[2], compute_graph.flexml_layers[51].compute_node[3][3].in[0], compute_graph.flexml_layers[52].compute_node[3][3].in[0], compute_graph.flexml_layers[53].compute_node[3][3].in[0], compute_graph.flexml_layers[54].compute_node[3][3].in[0], compute_graph.flexml_layers[55].compute_node[3][3].in[0], compute_graph.flexml_layers[55].compute_node[3][3].in[2], compute_graph.flexml_layers[56].compute_node[3][3].in[0], compute_graph.flexml_layers[57].compute_node[3][3].in[0], compute_graph.flexml_layers[58].compute_node[3][3].in[0], compute_graph.flexml_layers[59].compute_node[3][3].in[0], compute_graph.flexml_layers[60].compute_node[3][3].in[0], compute_graph.flexml_layers[61].compute_node[3][3].in[0], compute_graph.flexml_layers[61].compute_node[3][3].in[2], compute_graph.flexml_layers[200].compute_node[3][3].in[0], compute_graph.flexml_layers[201].compute_node[3][3].in[0], compute_graph.flexml_layers[64].compute_node[3][3].in[0], compute_graph.flexml_layers[65].compute_node[3][3].in[0], compute_graph.flexml_layers[66].compute_node[3][3].in[0], compute_graph.flexml_layers[66].compute_node[3][3].in[2], compute_graph.flexml_layers[67].compute_node[3][3].in[0], compute_graph.flexml_layers[67].compute_node[3][3].in[2], compute_graph.flexml_layers[68].compute_node[3][3].in[0], compute_graph.flexml_layers[69].compute_node[3][3].in[0], compute_graph.flexml_layers[70].compute_node[3][3].in[0], compute_graph.flexml_layers[71].compute_node[3][3].in[0], compute_graph.flexml_layers[72].compute_node[3][3].in[0], compute_graph.flexml_layers[72].compute_node[3][3].in[2], compute_graph.flexml_layers[73].compute_node[3][3].in[0], compute_graph.flexml_layers[74].compute_node[3][3].in[0], compute_graph.flexml_layers[75].compute_node[3][3].in[0], compute_graph.flexml_layers[76].compute_node[3][3].in[0], compute_graph.flexml_layers[77].compute_node[3][3].in[0], compute_graph.flexml_layers[77].compute_node[3][3].in[2], compute_graph.flexml_layers[78].compute_node[3][3].in[0], compute_graph.flexml_layers[78].compute_node[3][3].in[2], compute_graph.flexml_layers[79].compute_node[3][3].in[0], compute_graph.flexml_layers[80].compute_node[3][3].in[0], compute_graph.flexml_layers[81].compute_node[3][3].in[0], compute_graph.flexml_layers[82].compute_node[3][3].in[0], compute_graph.flexml_layers[83].compute_node[3][3].in[0], compute_graph.flexml_layers[83].compute_node[3][3].in[2], compute_graph.flexml_layers[84].compute_node[3][3].in[0], compute_graph.flexml_layers[85].compute_node[3][3].in[0], compute_graph.flexml_layers[86].compute_node[3][3].in[0], compute_graph.flexml_layers[87].compute_node[3][3].in[0], compute_graph.flexml_layers[88].compute_node[3][3].in[0], compute_graph.flexml_layers[88].compute_node[3][3].in[2], compute_graph.flexml_layers[89].compute_node[3][3].in[0], compute_graph.flexml_layers[89].compute_node[3][3].in[2], compute_graph.flexml_layers[90].compute_node[3][3].in[0], compute_graph.flexml_layers[91].compute_node[3][3].in[0], compute_graph.flexml_layers[92].compute_node[3][3].in[0], compute_graph.flexml_layers[93].compute_node[3][3].in[0], compute_graph.flexml_layers[94].compute_node[3][3].in[0], compute_graph.flexml_layers[94].compute_node[3][3].in[2], compute_graph.flexml_layers[95].compute_node[3][3].in[0], compute_graph.flexml_layers[96].compute_node[3][3].in[0], compute_graph.flexml_layers[97].compute_node[3][3].in[0], compute_graph.flexml_layers[98].compute_node[3][3].in[0], compute_graph.flexml_layers[99].compute_node[3][3].in[0], compute_graph.flexml_layers[99].compute_node[3][3].in[2], compute_graph.flexml_layers[100].compute_node[3][3].in[0], compute_graph.flexml_layers[101].compute_node[3][3].in[0], compute_graph.flexml_layers[102].compute_node[3][3].in[0], compute_graph.flexml_layers[103].compute_node[3][3].in[0], compute_graph.flexml_layers[104].compute_node[3][3].in[0], compute_graph.flexml_layers[105].compute_node[3][3].in[0], compute_graph.flexml_layers[106].compute_node[3][3].in[0], compute_graph.flexml_layers[106].compute_node[3][3].in[2], compute_graph.flexml_layers[107].compute_node[3][3].in[0], compute_graph.flexml_layers[108].compute_node[3][3].in[0], compute_graph.flexml_layers[109].compute_node[3][3].in[0], compute_graph.flexml_layers[110].compute_node[3][3].in[0], compute_graph.flexml_layers[111].compute_node[3][3].in[0], compute_graph.flexml_layers[112].compute_node[3][3].in[0], compute_graph.flexml_layers[112].compute_node[3][3].in[2], compute_graph.flexml_layers[113].compute_node[3][3].in[0], compute_graph.flexml_layers[114].compute_node[3][3].in[0], compute_graph.flexml_layers[115].compute_node[3][3].in[0], compute_graph.flexml_layers[116].compute_node[3][3].in[0], compute_graph.flexml_layers[117].compute_node[3][3].in[0], compute_graph.flexml_layers[117].compute_node[3][3].in[2], compute_graph.flexml_layers[118].compute_node[3][3].in[0], compute_graph.flexml_layers[119].compute_node[3][3].in[0], compute_graph.flexml_layers[120].compute_node[3][3].in[0], compute_graph.flexml_layers[121].compute_node[3][3].in[0], compute_graph.flexml_layers[122].compute_node[3][3].in[0], compute_graph.flexml_layers[123].compute_node[3][3].in[0], compute_graph.flexml_layers[124].compute_node[3][3].in[0], compute_graph.flexml_layers[124].compute_node[3][3].in[2], compute_graph.flexml_layers[125].compute_node[3][3].in[0], compute_graph.flexml_layers[125].compute_node[3][3].in[2], compute_graph.flexml_layers[126].compute_node[3][3].in[0], compute_graph.flexml_layers[127].compute_node[3][3].in[0], compute_graph.flexml_layers[128].compute_node[3][3].in[0], compute_graph.flexml_layers[129].compute_node[3][3].in[0], compute_graph.flexml_layers[130].compute_node[3][3].in[0], compute_graph.flexml_layers[130].compute_node[3][3].in[2], compute_graph.flexml_layers[131].compute_node[3][3].in[0], compute_graph.flexml_layers[132].compute_node[3][3].in[0], compute_graph.flexml_layers[133].compute_node[3][3].in[0], compute_graph.flexml_layers[134].compute_node[3][3].in[0], compute_graph.flexml_layers[135].compute_node[3][3].in[0], compute_graph.flexml_layers[135].compute_node[3][3].in[2], compute_graph.flexml_layers[136].compute_node[3][3].in[0], compute_graph.flexml_layers[137].compute_node[3][3].in[0], compute_graph.flexml_layers[138].compute_node[3][3].in[0], compute_graph.flexml_layers[139].compute_node[3][3].in[0], compute_graph.flexml_layers[140].compute_node[3][3].in[0], compute_graph.flexml_layers[141].compute_node[3][3].in[0], compute_graph.flexml_layers[142].compute_node[3][3].in[0], compute_graph.flexml_layers[142].compute_node[3][3].in[2], compute_graph.flexml_layers[145].compute_node[3][3].in[0], compute_graph.flexml_layers[146].compute_node[3][3].in[0], compute_graph.flexml_layers[147].compute_node[3][3].in[0], compute_graph.flexml_layers[148].compute_node[3][3].in[0], compute_graph.flexml_layers[148].compute_node[3][3].in[2], compute_graph.flexml_layers[149].compute_node[3][3].in[0], compute_graph.flexml_layers[150].compute_node[3][3].in[0], compute_graph.flexml_layers[151].compute_node[3][3].in[0], compute_graph.flexml_layers[152].compute_node[3][3].in[0], compute_graph.flexml_layers[153].compute_node[3][3].in[0], compute_graph.flexml_layers[153].compute_node[3][3].in[2], compute_graph.flexml_layers[154].compute_node[3][3].in[0], compute_graph.flexml_layers[155].compute_node[3][3].in[0], compute_graph.flexml_layers[156].compute_node[3][3].in[0], compute_graph.flexml_layers[157].compute_node[3][3].in[0], compute_graph.flexml_layers[158].compute_node[3][3].in[0], compute_graph.flexml_layers[159].compute_node[3][3].in[0], compute_graph.flexml_layers[160].compute_node[3][3].in[0], compute_graph.flexml_layers[160].compute_node[3][3].in[2], compute_graph.flexml_layers[161].compute_node[3][3].in[0], compute_graph.flexml_layers[161].compute_node[3][3].in[2], compute_graph.flexml_layers[162].compute_node[3][3].in[0], compute_graph.flexml_layers[163].compute_node[3][3].in[0], compute_graph.flexml_layers[164].compute_node[3][3].in[0], compute_graph.flexml_layers[165].compute_node[3][3].in[0], compute_graph.flexml_layers[166].compute_node[3][3].in[0], compute_graph.flexml_layers[166].compute_node[3][3].in[2], compute_graph.flexml_layers[167].compute_node[3][3].in[0], compute_graph.flexml_layers[168].compute_node[3][3].in[0], compute_graph.flexml_layers[169].compute_node[3][3].in[0], compute_graph.flexml_layers[170].compute_node[3][3].in[0], compute_graph.flexml_layers[171].compute_node[3][3].in[0], compute_graph.flexml_layers[171].compute_node[3][3].in[2], compute_graph.flexml_layers[172].compute_node[3][3].in[0], compute_graph.flexml_layers[173].compute_node[3][3].in[0], compute_graph.flexml_layers[174].compute_node[3][3].in[0], compute_graph.flexml_layers[175].compute_node[3][3].in[0], compute_graph.flexml_layers[176].compute_node[3][3].in[0], compute_graph.flexml_layers[177].compute_node[3][3].in[0], compute_graph.flexml_layers[178].compute_node[3][3].in[0], compute_graph.flexml_layers[178].compute_node[3][3].in[2], compute_graph.flexml_layers[179].compute_node[3][3].in[0], compute_graph.flexml_layers[179].compute_node[3][3].in[2], compute_graph.flexml_layers[180].compute_node[3][3].in[0], compute_graph.flexml_layers[181].compute_node[3][3].in[0], compute_graph.flexml_layers[182].compute_node[3][3].in[0], compute_graph.flexml_layers[183].compute_node[3][3].in[0], compute_graph.flexml_layers[184].compute_node[3][3].in[0], compute_graph.flexml_layers[184].compute_node[3][3].in[2], compute_graph.flexml_layers[185].compute_node[3][3].in[0], compute_graph.flexml_layers[186].compute_node[3][3].in[0], compute_graph.flexml_layers[187].compute_node[3][3].in[0], compute_graph.flexml_layers[188].compute_node[3][3].in[0], compute_graph.flexml_layers[188].compute_node[3][3].in[2], compute_graph.flexml_layers[189].compute_node[3][3].in[0], compute_graph.flexml_layers[190].compute_node[3][3].in[0], compute_graph.flexml_layers[191].compute_node[3][3].in[0], compute_graph.flexml_layers[191].compute_node[3][3].in[2], compute_graph.flexml_layers[192].compute_node[3][3].in[0], compute_graph.flexml_layers[192].compute_node[3][3].in[2], compute_graph.flexml_layers[193].compute_node[3][3].in[0], compute_graph.flexml_layers[193].compute_node[3][3].in[2], compute_graph.flexml_layers[194].compute_node[3][3].in[0], compute_graph.flexml_layers[195].compute_node[3][3].in[0], compute_graph.flexml_layers[196].compute_node[3][3].in[0], compute_graph.flexml_layers[196].compute_node[3][3].in[2], compute_graph.flexml_layers[197].compute_node[3][3].in[0], compute_graph.flexml_layers[197].compute_node[3][3].in[2], compute_graph.flexml_layers[198].compute_node[3][3].in[0], compute_graph.flexml_layers[199].compute_node[3][3].in[0], compute_graph.flexml_layers[202].compute_node[3][3].in[0], compute_graph.flexml_layers[203].compute_node[3][3].in[0], compute_graph.flexml_layers[204].compute_node[3][3].in[0], compute_graph.flexml_layers[204].compute_node[3][3].in[2], compute_graph.flexml_layers[205].compute_node[3][3].in[0], compute_graph.flexml_layers[205].compute_node[3][3].in[2], compute_graph.flexml_layers[206].compute_node[3][3].in[0], compute_graph.flexml_layers[206].compute_node[3][3].in[2], compute_graph.flexml_layers[207].compute_node[3][3].in[0], compute_graph.flexml_layers[208].compute_node[3][3].in[0], compute_graph.flexml_layers[209].compute_node[3][3].in[0], compute_graph.flexml_layers[209].compute_node[3][3].in[2], compute_graph.flexml_layers[210].compute_node[3][3].in[0], compute_graph.flexml_layers[210].compute_node[3][3].in[2], compute_graph.flexml_layers[211].compute_node[3][3].in[0], compute_graph.flexml_layers[212].compute_node[3][3].in[0], compute_graph.flexml_layers[213].compute_node[3][3].in[0], compute_graph.flexml_layers[214].compute_node[3][3].in[0], compute_graph.flexml_layers[214].compute_node[3][3].in[2], compute_graph.flexml_layers[215].compute_node[3][3].in[0], compute_graph.flexml_layers[215].compute_node[3][3].in[2], compute_graph.flexml_layers[216].compute_node[3][3].in[0], compute_graph.flexml_layers[216].compute_node[3][3].in[2], compute_graph.flexml_layers[217].compute_node[3][3].in[0], compute_graph.flexml_layers[218].compute_node[3][3].in[0], compute_graph.flexml_layers[219].compute_node[3][3].in[0], compute_graph.flexml_layers[219].compute_node[3][3].in[2], compute_graph.flexml_layers[220].compute_node[3][3].in[0], compute_graph.flexml_layers[220].compute_node[3][3].in[2], compute_graph.flexml_layers[221].compute_node[3][3].in[0], compute_graph.flexml_layers[222].compute_node[3][3].in[0], compute_graph.flexml_layers[223].compute_node[3][3].in[0], compute_graph.flexml_layers[224].compute_node[3][3].in[0], compute_graph.flexml_layers[224].compute_node[3][3].in[2], compute_graph.flexml_layers[225].compute_node[3][3].in[0], compute_graph.flexml_layers[225].compute_node[3][3].in[2], compute_graph.flexml_layers[226].compute_node[3][3].in[0], compute_graph.flexml_layers[226].compute_node[3][3].in[2], compute_graph.flexml_layers[227].compute_node[3][3].in[0], compute_graph.flexml_layers[228].compute_node[3][3].in[0], compute_graph.flexml_layers[229].compute_node[3][3].in[0], compute_graph.flexml_layers[229].compute_node[3][3].in[2], compute_graph.flexml_layers[230].compute_node[3][3].in[0], compute_graph.flexml_layers[230].compute_node[3][3].in[2], compute_graph.flexml_layers[231].compute_node[3][3].in[0], compute_graph.flexml_layers[232].compute_node[3][3].in[0], compute_graph.flexml_layers[233].compute_node[3][3].in[0], compute_graph.flexml_layers[234].compute_node[3][3].in[0], compute_graph.flexml_layers[235].compute_node[3][3].in[0], compute_graph.flexml_layers[235].compute_node[3][3].in[2], compute_graph.flexml_layers[236].compute_node[3][3].in[0]) + +Column 3Row 3Channel 0 + +i19_po0(compute_graph.templated_graph_4.trans_comp_nd[3][3].out[0], compute_graph.flexml_layers[35].compute_node[3][3].out[0], compute_graph.templated_graph_231.mk[3][3].out[0], compute_graph.templated_graph_254.mk[3][3].out[0], compute_graph.templated_graph_258.compute_node[3][3].out[0], compute_graph.templated_graph_259.compute_node[3][3].out[0], compute_graph.templated_graph_260.mk[3][3].out[0], compute_graph.templated_graph_263.compute_node[3][3].out[0], compute_graph.templated_graph_266.compute_node[3][3].out[0], compute_graph.templated_graph_268.mk[3][3].out[0], compute_graph.templated_graph_273.mk[3][3].out[0], compute_graph.templated_graph_274.mk[3][3].out[0], compute_graph.flexml_layers[62].compute_node[3][3].out[0], compute_graph.flexml_layers[63].compute_node[3][3].out[0], compute_graph.templated_graph_293.mk[3][3].out[0], compute_graph.templated_graph_298.compute_node[3][3].out[0], compute_graph.templated_graph_300.compute_node[3][3].out[0], compute_graph.flexml_layers[0].compute_node[3][3].out[0], compute_graph.flexml_layers[1].compute_node[3][3].out[0], compute_graph.flexml_layers[2].compute_node[3][3].out[0], compute_graph.flexml_layers[3].compute_node[3][3].out[0], compute_graph.flexml_layers[4].compute_node[3][3].out[0], compute_graph.flexml_layers[5].compute_node[3][3].out[0], compute_graph.flexml_layers[6].compute_node[3][3].out[0], compute_graph.flexml_layers[7].compute_node[3][3].out[0], compute_graph.flexml_layers[8].compute_node[3][3].out[0], compute_graph.flexml_layers[9].compute_node[3][3].out[0], compute_graph.flexml_layers[10].compute_node[3][3].out[0], compute_graph.flexml_layers[11].compute_node[3][3].out[0], compute_graph.flexml_layers[12].compute_node[3][3].out[0], compute_graph.flexml_layers[13].compute_node[3][3].out[0], compute_graph.flexml_layers[14].compute_node[3][3].out[0], compute_graph.flexml_layers[15].compute_node[3][3].out[0], compute_graph.flexml_layers[16].compute_node[3][3].out[0], compute_graph.flexml_layers[17].compute_node[3][3].out[0], compute_graph.flexml_layers[18].compute_node[3][3].out[0], compute_graph.flexml_layers[19].compute_node[3][3].out[0], compute_graph.flexml_layers[20].compute_node[3][3].out[0], compute_graph.flexml_layers[21].compute_node[3][3].out[0], compute_graph.flexml_layers[22].compute_node[3][3].out[0], compute_graph.flexml_layers[23].compute_node[3][3].out[0], compute_graph.flexml_layers[24].compute_node[3][3].out[0], compute_graph.flexml_layers[25].compute_node[3][3].out[0], compute_graph.flexml_layers[26].compute_node[3][3].out[0], compute_graph.flexml_layers[27].compute_node[3][3].out[0], compute_graph.flexml_layers[28].compute_node[3][3].out[0], compute_graph.flexml_layers[29].compute_node[3][3].out[0], compute_graph.flexml_layers[30].compute_node[3][3].out[0], compute_graph.flexml_layers[31].compute_node[3][3].out[0], compute_graph.flexml_layers[32].compute_node[3][3].out[0], compute_graph.flexml_layers[33].compute_node[3][3].out[0], compute_graph.flexml_layers[34].compute_node[3][3].out[0], compute_graph.flexml_layers[143].compute_node[3][3].out[0], compute_graph.flexml_layers[144].compute_node[3][3].out[0], compute_graph.flexml_layers[36].compute_node[3][3].out[0], compute_graph.flexml_layers[37].compute_node[3][3].out[0], compute_graph.flexml_layers[38].compute_node[3][3].out[0], compute_graph.flexml_layers[39].compute_node[3][3].out[0], compute_graph.flexml_layers[40].compute_node[3][3].out[0], compute_graph.flexml_layers[41].compute_node[3][3].out[0], compute_graph.flexml_layers[42].compute_node[3][3].out[0], compute_graph.flexml_layers[43].compute_node[3][3].out[0], compute_graph.flexml_layers[44].compute_node[3][3].out[0], compute_graph.flexml_layers[45].compute_node[3][3].out[0], compute_graph.flexml_layers[46].compute_node[3][3].out[0], compute_graph.flexml_layers[47].compute_node[3][3].out[0], compute_graph.flexml_layers[48].compute_node[3][3].out[0], compute_graph.flexml_layers[49].compute_node[3][3].out[0], compute_graph.flexml_layers[50].compute_node[3][3].out[0], compute_graph.flexml_layers[51].compute_node[3][3].out[0], compute_graph.flexml_layers[52].compute_node[3][3].out[0], compute_graph.flexml_layers[53].compute_node[3][3].out[0], compute_graph.flexml_layers[54].compute_node[3][3].out[0], compute_graph.flexml_layers[55].compute_node[3][3].out[0], compute_graph.flexml_layers[56].compute_node[3][3].out[0], compute_graph.flexml_layers[57].compute_node[3][3].out[0], compute_graph.flexml_layers[58].compute_node[3][3].out[0], compute_graph.flexml_layers[59].compute_node[3][3].out[0], compute_graph.flexml_layers[60].compute_node[3][3].out[0], compute_graph.flexml_layers[61].compute_node[3][3].out[0], compute_graph.flexml_layers[200].compute_node[3][3].out[0], compute_graph.flexml_layers[201].compute_node[3][3].out[0], compute_graph.flexml_layers[64].compute_node[3][3].out[0], compute_graph.flexml_layers[65].compute_node[3][3].out[0], compute_graph.flexml_layers[66].compute_node[3][3].out[0], compute_graph.flexml_layers[67].compute_node[3][3].out[0], compute_graph.flexml_layers[68].compute_node[3][3].out[0], compute_graph.flexml_layers[69].compute_node[3][3].out[0], compute_graph.flexml_layers[70].compute_node[3][3].out[0], compute_graph.flexml_layers[71].compute_node[3][3].out[0], compute_graph.flexml_layers[72].compute_node[3][3].out[0], compute_graph.flexml_layers[73].compute_node[3][3].out[0], compute_graph.flexml_layers[74].compute_node[3][3].out[0], compute_graph.flexml_layers[75].compute_node[3][3].out[0], compute_graph.flexml_layers[76].compute_node[3][3].out[0], compute_graph.flexml_layers[77].compute_node[3][3].out[0], compute_graph.flexml_layers[78].compute_node[3][3].out[0], compute_graph.flexml_layers[79].compute_node[3][3].out[0], compute_graph.flexml_layers[80].compute_node[3][3].out[0], compute_graph.flexml_layers[81].compute_node[3][3].out[0], compute_graph.flexml_layers[82].compute_node[3][3].out[0], compute_graph.flexml_layers[83].compute_node[3][3].out[0], compute_graph.flexml_layers[84].compute_node[3][3].out[0], compute_graph.flexml_layers[85].compute_node[3][3].out[0], compute_graph.flexml_layers[86].compute_node[3][3].out[0], compute_graph.flexml_layers[87].compute_node[3][3].out[0], compute_graph.flexml_layers[88].compute_node[3][3].out[0], compute_graph.flexml_layers[89].compute_node[3][3].out[0], compute_graph.flexml_layers[90].compute_node[3][3].out[0], compute_graph.flexml_layers[91].compute_node[3][3].out[0], compute_graph.flexml_layers[92].compute_node[3][3].out[0], compute_graph.flexml_layers[93].compute_node[3][3].out[0], compute_graph.flexml_layers[94].compute_node[3][3].out[0], compute_graph.flexml_layers[95].compute_node[3][3].out[0], compute_graph.flexml_layers[96].compute_node[3][3].out[0], compute_graph.flexml_layers[97].compute_node[3][3].out[0], compute_graph.flexml_layers[98].compute_node[3][3].out[0], compute_graph.flexml_layers[99].compute_node[3][3].out[0], compute_graph.flexml_layers[100].compute_node[3][3].out[0], compute_graph.flexml_layers[101].compute_node[3][3].out[0], compute_graph.flexml_layers[102].compute_node[3][3].out[0], compute_graph.flexml_layers[103].compute_node[3][3].out[0], compute_graph.flexml_layers[104].compute_node[3][3].out[0], compute_graph.flexml_layers[105].compute_node[3][3].out[0], compute_graph.flexml_layers[106].compute_node[3][3].out[0], compute_graph.flexml_layers[107].compute_node[3][3].out[0], compute_graph.flexml_layers[108].compute_node[3][3].out[0], compute_graph.flexml_layers[109].compute_node[3][3].out[0], compute_graph.flexml_layers[110].compute_node[3][3].out[0], compute_graph.flexml_layers[111].compute_node[3][3].out[0], compute_graph.flexml_layers[112].compute_node[3][3].out[0], compute_graph.flexml_layers[113].compute_node[3][3].out[0], compute_graph.flexml_layers[114].compute_node[3][3].out[0], compute_graph.flexml_layers[115].compute_node[3][3].out[0], compute_graph.flexml_layers[116].compute_node[3][3].out[0], compute_graph.flexml_layers[117].compute_node[3][3].out[0], compute_graph.flexml_layers[118].compute_node[3][3].out[0], compute_graph.flexml_layers[119].compute_node[3][3].out[0], compute_graph.flexml_layers[120].compute_node[3][3].out[0], compute_graph.flexml_layers[121].compute_node[3][3].out[0], compute_graph.flexml_layers[122].compute_node[3][3].out[0], compute_graph.flexml_layers[123].compute_node[3][3].out[0], compute_graph.flexml_layers[124].compute_node[3][3].out[0], compute_graph.flexml_layers[125].compute_node[3][3].out[0], compute_graph.flexml_layers[126].compute_node[3][3].out[0], compute_graph.flexml_layers[127].compute_node[3][3].out[0], compute_graph.flexml_layers[128].compute_node[3][3].out[0], compute_graph.flexml_layers[129].compute_node[3][3].out[0], compute_graph.flexml_layers[130].compute_node[3][3].out[0], compute_graph.flexml_layers[131].compute_node[3][3].out[0], compute_graph.flexml_layers[132].compute_node[3][3].out[0], compute_graph.flexml_layers[133].compute_node[3][3].out[0], compute_graph.flexml_layers[134].compute_node[3][3].out[0], compute_graph.flexml_layers[135].compute_node[3][3].out[0], compute_graph.flexml_layers[136].compute_node[3][3].out[0], compute_graph.flexml_layers[137].compute_node[3][3].out[0], compute_graph.flexml_layers[138].compute_node[3][3].out[0], compute_graph.flexml_layers[139].compute_node[3][3].out[0], compute_graph.flexml_layers[140].compute_node[3][3].out[0], compute_graph.flexml_layers[141].compute_node[3][3].out[0], compute_graph.flexml_layers[142].compute_node[3][3].out[0], compute_graph.flexml_layers[145].compute_node[3][3].out[0], compute_graph.flexml_layers[146].compute_node[3][3].out[0], compute_graph.flexml_layers[147].compute_node[3][3].out[0], compute_graph.flexml_layers[148].compute_node[3][3].out[0], compute_graph.flexml_layers[149].compute_node[3][3].out[0], compute_graph.flexml_layers[150].compute_node[3][3].out[0], compute_graph.flexml_layers[151].compute_node[3][3].out[0], compute_graph.flexml_layers[152].compute_node[3][3].out[0], compute_graph.flexml_layers[153].compute_node[3][3].out[0], compute_graph.flexml_layers[154].compute_node[3][3].out[0], compute_graph.flexml_layers[155].compute_node[3][3].out[0], compute_graph.flexml_layers[156].compute_node[3][3].out[0], compute_graph.flexml_layers[157].compute_node[3][3].out[0], compute_graph.flexml_layers[158].compute_node[3][3].out[0], compute_graph.flexml_layers[159].compute_node[3][3].out[0], compute_graph.flexml_layers[160].compute_node[3][3].out[0], compute_graph.flexml_layers[161].compute_node[3][3].out[0], compute_graph.flexml_layers[162].compute_node[3][3].out[0], compute_graph.flexml_layers[163].compute_node[3][3].out[0], compute_graph.flexml_layers[164].compute_node[3][3].out[0], compute_graph.flexml_layers[165].compute_node[3][3].out[0], compute_graph.flexml_layers[166].compute_node[3][3].out[0], compute_graph.flexml_layers[167].compute_node[3][3].out[0], compute_graph.flexml_layers[168].compute_node[3][3].out[0], compute_graph.flexml_layers[169].compute_node[3][3].out[0], compute_graph.flexml_layers[170].compute_node[3][3].out[0], compute_graph.flexml_layers[171].compute_node[3][3].out[0], compute_graph.flexml_layers[172].compute_node[3][3].out[0], compute_graph.flexml_layers[173].compute_node[3][3].out[0], compute_graph.flexml_layers[174].compute_node[3][3].out[0], compute_graph.flexml_layers[175].compute_node[3][3].out[0], compute_graph.flexml_layers[176].compute_node[3][3].out[0], compute_graph.flexml_layers[177].compute_node[3][3].out[0], compute_graph.flexml_layers[178].compute_node[3][3].out[0], compute_graph.flexml_layers[179].compute_node[3][3].out[0], compute_graph.flexml_layers[180].compute_node[3][3].out[0], compute_graph.flexml_layers[181].compute_node[3][3].out[0], compute_graph.flexml_layers[182].compute_node[3][3].out[0], compute_graph.flexml_layers[183].compute_node[3][3].out[0], compute_graph.flexml_layers[184].compute_node[3][3].out[0], compute_graph.flexml_layers[185].compute_node[3][3].out[0], compute_graph.flexml_layers[186].compute_node[3][3].out[0], compute_graph.flexml_layers[187].compute_node[3][3].out[0], compute_graph.flexml_layers[188].compute_node[3][3].out[0], compute_graph.flexml_layers[189].compute_node[3][3].out[0], compute_graph.flexml_layers[190].compute_node[3][3].out[0], compute_graph.flexml_layers[191].compute_node[3][3].out[0], compute_graph.flexml_layers[192].compute_node[3][3].out[0], compute_graph.flexml_layers[193].compute_node[3][3].out[0], compute_graph.flexml_layers[194].compute_node[3][3].out[0], compute_graph.flexml_layers[195].compute_node[3][3].out[0], compute_graph.flexml_layers[196].compute_node[3][3].out[0], compute_graph.flexml_layers[197].compute_node[3][3].out[0], compute_graph.flexml_layers[198].compute_node[3][3].out[0], compute_graph.flexml_layers[199].compute_node[3][3].out[0], compute_graph.flexml_layers[202].compute_node[3][3].out[0], compute_graph.flexml_layers[203].compute_node[3][3].out[0], compute_graph.flexml_layers[204].compute_node[3][3].out[0], compute_graph.flexml_layers[205].compute_node[3][3].out[0], compute_graph.flexml_layers[206].compute_node[3][3].out[0], compute_graph.flexml_layers[207].compute_node[3][3].out[0], compute_graph.flexml_layers[208].compute_node[3][3].out[0], compute_graph.flexml_layers[209].compute_node[3][3].out[0], compute_graph.flexml_layers[210].compute_node[3][3].out[0], compute_graph.flexml_layers[211].compute_node[3][3].out[0], compute_graph.flexml_layers[212].compute_node[3][3].out[0], compute_graph.flexml_layers[213].compute_node[3][3].out[0], compute_graph.flexml_layers[214].compute_node[3][3].out[0], compute_graph.flexml_layers[215].compute_node[3][3].out[0], compute_graph.flexml_layers[216].compute_node[3][3].out[0], compute_graph.flexml_layers[217].compute_node[3][3].out[0], compute_graph.flexml_layers[218].compute_node[3][3].out[0], compute_graph.flexml_layers[219].compute_node[3][3].out[0], compute_graph.flexml_layers[220].compute_node[3][3].out[0], compute_graph.flexml_layers[221].compute_node[3][3].out[0], compute_graph.flexml_layers[222].compute_node[3][3].out[0], compute_graph.flexml_layers[223].compute_node[3][3].out[0], compute_graph.flexml_layers[224].compute_node[3][3].out[0], compute_graph.flexml_layers[225].compute_node[3][3].out[0], compute_graph.flexml_layers[226].compute_node[3][3].out[0], compute_graph.flexml_layers[227].compute_node[3][3].out[0], compute_graph.flexml_layers[228].compute_node[3][3].out[0], compute_graph.flexml_layers[229].compute_node[3][3].out[0], compute_graph.flexml_layers[230].compute_node[3][3].out[0], compute_graph.flexml_layers[231].compute_node[3][3].out[0], compute_graph.flexml_layers[232].compute_node[3][3].out[0], compute_graph.flexml_layers[233].compute_node[3][3].out[0], compute_graph.flexml_layers[234].compute_node[3][3].out[0], compute_graph.flexml_layers[235].compute_node[3][3].out[0], compute_graph.flexml_layers[236].compute_node[3][3].out[0]) + +Column 3Row 3Channel 1 + +i19_pi1(compute_graph.flexml_layers[35].compute_node[3][3].in[1], compute_graph.templated_graph_260.mk[3][3].in[1], compute_graph.templated_graph_268.mk[3][3].in[1], compute_graph.templated_graph_273.mk[3][3].in[1], compute_graph.flexml_layers[62].compute_node[3][3].in[1], compute_graph.flexml_layers[3].compute_node[3][3].in[1], compute_graph.flexml_layers[8].compute_node[3][3].in[1], compute_graph.flexml_layers[9].compute_node[3][3].in[1], compute_graph.flexml_layers[10].compute_node[3][3].in[1], compute_graph.flexml_layers[11].compute_node[3][3].in[1], compute_graph.flexml_layers[12].compute_node[3][3].in[1], compute_graph.flexml_layers[13].compute_node[3][3].in[1], compute_graph.flexml_layers[14].compute_node[3][3].in[1], compute_graph.flexml_layers[15].compute_node[3][3].in[1], compute_graph.flexml_layers[16].compute_node[3][3].in[1], compute_graph.flexml_layers[17].compute_node[3][3].in[1], compute_graph.flexml_layers[19].compute_node[3][3].in[1], compute_graph.flexml_layers[20].compute_node[3][3].in[1], compute_graph.flexml_layers[25].compute_node[3][3].in[1], compute_graph.flexml_layers[26].compute_node[3][3].in[1], compute_graph.flexml_layers[27].compute_node[3][3].in[1], compute_graph.flexml_layers[29].compute_node[3][3].in[1], compute_graph.flexml_layers[30].compute_node[3][3].in[1], compute_graph.flexml_layers[143].compute_node[3][3].in[1], compute_graph.flexml_layers[144].compute_node[3][3].in[1], compute_graph.flexml_layers[36].compute_node[3][3].in[1], compute_graph.flexml_layers[37].compute_node[3][3].in[1], compute_graph.flexml_layers[39].compute_node[3][3].in[1], compute_graph.flexml_layers[40].compute_node[3][3].in[1], compute_graph.flexml_layers[45].compute_node[3][3].in[1], compute_graph.flexml_layers[46].compute_node[3][3].in[1], compute_graph.flexml_layers[51].compute_node[3][3].in[1], compute_graph.flexml_layers[56].compute_node[3][3].in[1], compute_graph.flexml_layers[57].compute_node[3][3].in[1], compute_graph.flexml_layers[201].compute_node[3][3].in[1], compute_graph.flexml_layers[67].compute_node[3][3].in[1], compute_graph.flexml_layers[68].compute_node[3][3].in[1], compute_graph.flexml_layers[73].compute_node[3][3].in[1], compute_graph.flexml_layers[78].compute_node[3][3].in[1], compute_graph.flexml_layers[79].compute_node[3][3].in[1], compute_graph.flexml_layers[84].compute_node[3][3].in[1], compute_graph.flexml_layers[89].compute_node[3][3].in[1], compute_graph.flexml_layers[90].compute_node[3][3].in[1], compute_graph.flexml_layers[95].compute_node[3][3].in[1], compute_graph.flexml_layers[101].compute_node[3][3].in[1], compute_graph.flexml_layers[102].compute_node[3][3].in[1], compute_graph.flexml_layers[107].compute_node[3][3].in[1], compute_graph.flexml_layers[108].compute_node[3][3].in[1], compute_graph.flexml_layers[113].compute_node[3][3].in[1], compute_graph.flexml_layers[119].compute_node[3][3].in[1], compute_graph.flexml_layers[120].compute_node[3][3].in[1], compute_graph.flexml_layers[125].compute_node[3][3].in[1], compute_graph.flexml_layers[126].compute_node[3][3].in[1], compute_graph.flexml_layers[131].compute_node[3][3].in[1], compute_graph.flexml_layers[137].compute_node[3][3].in[1], compute_graph.flexml_layers[138].compute_node[3][3].in[1], compute_graph.flexml_layers[149].compute_node[3][3].in[1], compute_graph.flexml_layers[155].compute_node[3][3].in[1], compute_graph.flexml_layers[156].compute_node[3][3].in[1], compute_graph.flexml_layers[161].compute_node[3][3].in[1], compute_graph.flexml_layers[162].compute_node[3][3].in[1], compute_graph.flexml_layers[167].compute_node[3][3].in[1], compute_graph.flexml_layers[173].compute_node[3][3].in[1], compute_graph.flexml_layers[174].compute_node[3][3].in[1], compute_graph.flexml_layers[179].compute_node[3][3].in[1], compute_graph.flexml_layers[180].compute_node[3][3].in[1], compute_graph.flexml_layers[186].compute_node[3][3].in[1], compute_graph.flexml_layers[188].compute_node[3][3].in[1], compute_graph.flexml_layers[189].compute_node[3][3].in[1], compute_graph.flexml_layers[194].compute_node[3][3].in[1], compute_graph.flexml_layers[202].compute_node[3][3].in[1], compute_graph.flexml_layers[207].compute_node[3][3].in[1], compute_graph.flexml_layers[211].compute_node[3][3].in[1], compute_graph.flexml_layers[212].compute_node[3][3].in[1], compute_graph.flexml_layers[217].compute_node[3][3].in[1], compute_graph.flexml_layers[221].compute_node[3][3].in[1], compute_graph.flexml_layers[222].compute_node[3][3].in[1], compute_graph.flexml_layers[227].compute_node[3][3].in[1], compute_graph.flexml_layers[231].compute_node[3][3].in[1], compute_graph.flexml_layers[232].compute_node[3][3].in[1], compute_graph.flexml_layers[233].compute_node[3][3].in[1]) + +Column 3Row 0Channel 0 + +i20_po0(compute_graph.Layer_8_wts_ddr.out[0], compute_graph.Layer_13_wts_ddr.out[0], compute_graph.Layer_14_wts_ddr.out[0], compute_graph.Layer_15_wts_ddr.out[0], compute_graph.Layer_16_wts_ddr.out[0], compute_graph.Layer_18_wts_ddr.out[0], compute_graph.Layer_19_wts_ddr.out[0], compute_graph.Layer_20_wts_ddr.out[0], compute_graph.Layer_22_wts_ddr.out[0], compute_graph.Layer_23_wts_ddr.out[0], compute_graph.Layer_24_wts_ddr.out[0], compute_graph.Layer_27_wts_ddr.out[0], compute_graph.Layer_28_wts_ddr.out[0], compute_graph.Layer_34_wts_ddr.out[0], compute_graph.Layer_35_wts_ddr.out[0], compute_graph.Layer_36_wts_ddr.out[0], compute_graph.Layer_39_wts_ddr.out[0], compute_graph.Layer_40_wts_ddr.out[0], compute_graph.Layer_46_wts_ddr.out[0], compute_graph.Layer_47_wts_ddr.out[0], compute_graph.Layer_48_wts_ddr.out[0], compute_graph.Layer_51_wts_ddr.out[0], compute_graph.Layer_52_wts_ddr.out[0], compute_graph.Layer_58_wts_ddr.out[0], compute_graph.Layer_59_wts_ddr.out[0], compute_graph.Layer_65_wts_ddr.out[0], compute_graph.Layer_70_wts_ddr.out[0], compute_graph.Layer_71_wts_ddr.out[0], compute_graph.Layer_76_wts_ddr.out[0], compute_graph.Layer_81_wts_ddr.out[0], compute_graph.Layer_82_wts_ddr.out[0], compute_graph.Layer_87_wts_ddr.out[0], compute_graph.Layer_241_wts_ddr.out[0], compute_graph.Layer_92_wts_ddr.out[0], compute_graph.Layer_93_wts_ddr.out[0], compute_graph.Layer_98_wts_ddr.out[0], compute_graph.Layer_103_wts_ddr.out[0], compute_graph.Layer_104_wts_ddr.out[0], compute_graph.Layer_109_wts_ddr.out[0], compute_graph.Layer_116_wts_ddr.out[0], compute_graph.Layer_117_wts_ddr.out[0], compute_graph.Layer_123_wts_ddr.out[0], compute_graph.Layer_124_wts_ddr.out[0], compute_graph.Layer_129_wts_ddr.out[0], compute_graph.Layer_136_wts_ddr.out[0], compute_graph.Layer_137_wts_ddr.out[0], compute_graph.Layer_143_wts_ddr.out[0], compute_graph.Layer_144_wts_ddr.out[0], compute_graph.Layer_149_wts_ddr.out[0], compute_graph.Layer_156_wts_ddr.out[0], compute_graph.Layer_157_wts_ddr.out[0], compute_graph.Layer_163_wts_ddr.out[0], compute_graph.Layer_164_wts_ddr.out[0], compute_graph.Layer_169_wts_ddr.out[0], compute_graph.Layer_176_wts_ddr.out[0], compute_graph.Layer_177_wts_ddr.out[0], compute_graph.Layer_183_wts_ddr.out[0], compute_graph.Layer_184_wts_ddr.out[0], compute_graph.Layer_189_wts_ddr.out[0], compute_graph.Layer_196_wts_ddr.out[0], compute_graph.Layer_197_wts_ddr.out[0], compute_graph.Layer_203_wts_ddr.out[0], compute_graph.Layer_204_wts_ddr.out[0], compute_graph.Layer_211_wts_ddr.out[0], compute_graph.Layer_214_wts_ddr.out[0], compute_graph.Layer_218_wts_ddr.out[0], compute_graph.Layer_226_wts_ddr.out[0], compute_graph.Layer_237_wts_ddr.out[0], compute_graph.Layer_249_wts_ddr.out[0], compute_graph.Layer_257_wts_ddr.out[0], compute_graph.Layer_261_wts_ddr.out[0], compute_graph.Layer_269_wts_ddr.out[0], compute_graph.Layer_276_wts_ddr.out[0], compute_graph.Layer_280_wts_ddr.out[0], compute_graph.Layer_288_wts_ddr.out[0], compute_graph.Layer_295_wts_ddr.out[0], compute_graph.Layer_296_wts_ddr.out[0], compute_graph.Layer_297_wts_ddr.out[0], compute_graph.ifm_ddr_2.out[0], compute_graph.l2l3_267_spill.out[0], compute_graph.ofm_ddr_2_l2l3_272_spill.out[0]) + +Column 3Row 0Channel 1 + +i21_po0(compute_graph.Layer_8_wts_ddr.out[1], compute_graph.Layer_13_wts_ddr.out[1], compute_graph.Layer_14_wts_ddr.out[1], compute_graph.Layer_15_wts_ddr.out[1], compute_graph.Layer_16_wts_ddr.out[1], compute_graph.Layer_18_wts_ddr.out[1], compute_graph.Layer_19_wts_ddr.out[1], compute_graph.Layer_20_wts_ddr.out[1], compute_graph.Layer_22_wts_ddr.out[1], compute_graph.Layer_23_wts_ddr.out[1], compute_graph.Layer_24_wts_ddr.out[1], compute_graph.Layer_27_wts_ddr.out[1], compute_graph.Layer_28_wts_ddr.out[1], compute_graph.Layer_34_wts_ddr.out[1], compute_graph.Layer_35_wts_ddr.out[1], compute_graph.Layer_36_wts_ddr.out[1], compute_graph.Layer_39_wts_ddr.out[1], compute_graph.Layer_40_wts_ddr.out[1], compute_graph.Layer_46_wts_ddr.out[1], compute_graph.Layer_47_wts_ddr.out[1], compute_graph.Layer_48_wts_ddr.out[1], compute_graph.Layer_51_wts_ddr.out[1], compute_graph.Layer_52_wts_ddr.out[1], compute_graph.Layer_58_wts_ddr.out[1], compute_graph.Layer_59_wts_ddr.out[1], compute_graph.Layer_65_wts_ddr.out[1], compute_graph.Layer_70_wts_ddr.out[1], compute_graph.Layer_71_wts_ddr.out[1], compute_graph.Layer_76_wts_ddr.out[1], compute_graph.Layer_81_wts_ddr.out[1], compute_graph.Layer_82_wts_ddr.out[1], compute_graph.Layer_87_wts_ddr.out[1], compute_graph.Layer_241_wts_ddr.out[1], compute_graph.Layer_92_wts_ddr.out[1], compute_graph.Layer_93_wts_ddr.out[1], compute_graph.Layer_98_wts_ddr.out[1], compute_graph.Layer_103_wts_ddr.out[1], compute_graph.Layer_104_wts_ddr.out[1], compute_graph.Layer_109_wts_ddr.out[1], compute_graph.Layer_116_wts_ddr.out[1], compute_graph.Layer_117_wts_ddr.out[1], compute_graph.Layer_123_wts_ddr.out[1], compute_graph.Layer_124_wts_ddr.out[1], compute_graph.Layer_129_wts_ddr.out[1], compute_graph.Layer_136_wts_ddr.out[1], compute_graph.Layer_137_wts_ddr.out[1], compute_graph.Layer_143_wts_ddr.out[1], compute_graph.Layer_144_wts_ddr.out[1], compute_graph.Layer_149_wts_ddr.out[1], compute_graph.Layer_156_wts_ddr.out[1], compute_graph.Layer_157_wts_ddr.out[1], compute_graph.Layer_163_wts_ddr.out[1], compute_graph.Layer_164_wts_ddr.out[1], compute_graph.Layer_169_wts_ddr.out[1], compute_graph.Layer_176_wts_ddr.out[1], compute_graph.Layer_177_wts_ddr.out[1], compute_graph.Layer_183_wts_ddr.out[1], compute_graph.Layer_184_wts_ddr.out[1], compute_graph.Layer_189_wts_ddr.out[1], compute_graph.Layer_196_wts_ddr.out[1], compute_graph.Layer_197_wts_ddr.out[1], compute_graph.Layer_203_wts_ddr.out[1], compute_graph.Layer_204_wts_ddr.out[1], compute_graph.Layer_211_wts_ddr.out[1], compute_graph.Layer_214_wts_ddr.out[1], compute_graph.Layer_218_wts_ddr.out[1], compute_graph.Layer_226_wts_ddr.out[1], compute_graph.Layer_237_wts_ddr.out[1], compute_graph.Layer_249_wts_ddr.out[1], compute_graph.Layer_257_wts_ddr.out[1], compute_graph.Layer_261_wts_ddr.out[1], compute_graph.Layer_269_wts_ddr.out[1], compute_graph.Layer_276_wts_ddr.out[1], compute_graph.Layer_280_wts_ddr.out[1], compute_graph.Layer_288_wts_ddr.out[1], compute_graph.Layer_295_wts_ddr.out[1], compute_graph.Layer_296_wts_ddr.out[1], compute_graph.Layer_297_wts_ddr.out[1]) + +Column 0Row 0Channel 0 + +i22_po0(compute_graph.ifm_ddr.out[0], compute_graph.l2l3_1_spill.out[0], compute_graph.l2l3_2_spill.out[0], compute_graph.l2l3_scratch_0_3_spill.out[0], compute_graph.l2l3_3_spill.out[0], compute_graph.l2l3_4_spill.out[0], compute_graph.l2l3_scratch_0_5_spill.out[0], compute_graph.l2l3_8_spill.out[0], compute_graph.l2l3_10_spill.out[0], compute_graph.l2l3_12_spill.out[0], compute_graph.l2l3_12_spill.out[2], compute_graph.l2l3_13_spill.out[0], compute_graph.l2l3_15_spill.out[0], compute_graph.l2l3_16_spill.out[0], compute_graph.l2l3_scratch_0_17_spill.out[0], compute_graph.l2l3_17_spill.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.out[0], compute_graph.l2l3_20_spill.out[0], compute_graph.l2l3_scratch_0_21_spill.out[0], compute_graph.l2l3_21_spill.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.out[0], compute_graph.l2l3_24_spill.out[0], compute_graph.l2l3_25_spill.out[0], compute_graph.l2l3_31_spill.out[0], compute_graph.l2l3_scratch_0_32_spill.out[0], compute_graph.l2l3_scratch_1_32_spill.out[0], compute_graph.l2l3_32_spill.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.out[0], compute_graph.l2l3_36_spill.out[0], compute_graph.l2l3_37_spill.out[0], compute_graph.l2l3_43_spill.out[0], compute_graph.l2l3_scratch_0_44_spill.out[0], compute_graph.l2l3_scratch_1_44_spill.out[0], compute_graph.l2l3_44_spill.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.out[0], compute_graph.l2l3_48_spill.out[0], compute_graph.l2l3_49_spill.out[0], compute_graph.l2l3_55_spill.out[0], compute_graph.l2l3_scratch_0_56_spill.out[0], compute_graph.l2l3_scratch_1_56_spill.out[0], compute_graph.l2l3_56_spill.out[0], compute_graph.l2l3_63_spill.out[0], compute_graph.l2l3_scratch_0_64_spill.out[0], compute_graph.l2l3_64_spill.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.out[0], compute_graph.l2l3_113_spill.out[0], compute_graph.l2l3_114_spill.out[0], compute_graph.l2l3_120_spill.out[0], compute_graph.l2l3_scratch_0_121_spill.out[0], compute_graph.l2l3_scratch_1_121_spill.out[0], compute_graph.l2l3_121_spill.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.out[0], compute_graph.l2l3_133_spill.out[0], compute_graph.l2l3_134_spill.out[0], compute_graph.l2l3_140_spill.out[0], compute_graph.l2l3_scratch_0_141_spill.out[0], compute_graph.l2l3_scratch_1_141_spill.out[0], compute_graph.l2l3_141_spill.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.out[0], compute_graph.l2l3_153_spill.out[0], compute_graph.l2l3_154_spill.out[0], compute_graph.l2l3_160_spill.out[0], compute_graph.l2l3_scratch_1_161_spill.out[0], compute_graph.l2l3_scratch_0_161_spill.out[0], compute_graph.l2l3_161_spill.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.out[0], compute_graph.l2l3_168_spill.out[0], compute_graph.l2l3_173_spill.out[0], compute_graph.l2l3_173_spill.out[1], compute_graph.l2l3_174_spill.out[0], compute_graph.l2l3_180_spill.out[0], compute_graph.l2l3_scratch_0_181_spill.out[0], compute_graph.l2l3_scratch_1_181_spill.out[0], compute_graph.l2l3_181_spill.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.out[0], compute_graph.l2l3_188_spill.out[0], compute_graph.l2l3_193_spill.out[0], compute_graph.l2l3_193_spill.out[1], compute_graph.l2l3_194_spill.out[0], compute_graph.l2l3_200_spill.out[0], compute_graph.l2l3_scratch_0_201_spill.out[0], compute_graph.l2l3_scratch_1_201_spill.out[0], compute_graph.l2l3_201_spill.out[0], compute_graph.l2l3_208_spill.out[0], compute_graph.l2l3_208_spill.out[1], compute_graph.l2l3_209_spill.out[0], compute_graph.l2l3_212_spill.out[0], compute_graph.l2l3_scratch_0_213_spill.out[0], compute_graph.l2l3_scratch_1_213_spill.out[0], compute_graph.l2l3_213_spill.out[0], compute_graph.l2l3_214_spill.out[0], compute_graph.l2l3_214_spill.out[2], compute_graph.l2l3_215_spill.out[0], compute_graph.l2l3_216_spill.out[0], compute_graph.l2l3_216_spill.out[2], compute_graph.ifm_ddr_4.out[0], compute_graph.ifm_ddr_4.out[2], compute_graph.ifm_ddr_4.out[4], compute_graph.spill_L3_Concat_Buffer_layer_217.out[0], compute_graph.l2l3_219_spill.out[0], compute_graph.l2l3_219_spill.out[2], compute_graph.l2l3_220_spill.out[0], compute_graph.l2l3_220_spill.out[2], compute_graph.const_ifm_ddr_3.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.out[0], compute_graph.l2l3_223_spill.out[0], compute_graph.spill_L3_Concat_Buffer_layer_225.out[0], compute_graph.spill_L3_Concat_Buffer_layer_230.out[0], compute_graph.l2l3_231_spill.out[0], compute_graph.l2l3_232_spill.out[0], compute_graph.l2l3_233_spill.out[0], compute_graph.spill_L3_Concat_Buffer_layer_236.out[0], compute_graph.l2l3_237_spill.out[0], compute_graph.l2l3_237_spill.out[2], compute_graph.l2l3_238_spill.out[0], compute_graph.l2l3_239_spill.out[0], compute_graph.l2l3_239_spill.out[2], compute_graph.ifm_ddr_3.out[0], compute_graph.ifm_ddr_3.out[2], compute_graph.ifm_ddr_3.out[4], compute_graph.spill_L3_Concat_Buffer_layer_240.out[0], compute_graph.l2l3_242_spill.out[0], compute_graph.l2l3_242_spill.out[2], compute_graph.l2l3_243_spill.out[0], compute_graph.l2l3_243_spill.out[2], compute_graph.const_ifm_ddr_2.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.out[0], compute_graph.l2l3_246_spill.out[0], compute_graph.spill_L3_Concat_Buffer_layer_248.out[0], compute_graph.spill_L3_Concat_Buffer_layer_253.out[0], compute_graph.l2l3_254_spill.out[0], compute_graph.spill_L3_Concat_Buffer_layer_256.out[0], compute_graph.l2l3_257_spill.out[0], compute_graph.l2l3_257_spill.out[2], compute_graph.l2l3_258_spill.out[0], compute_graph.l2l3_259_spill.out[0], compute_graph.l2l3_259_spill.out[1], compute_graph.ifm_ddr_2.out[1], compute_graph.ifm_ddr_2.out[3], compute_graph.l2l3_260_spill.out[0], compute_graph.l2l3_262_spill.out[0], compute_graph.l2l3_262_spill.out[2], compute_graph.l2l3_263_spill.out[0], compute_graph.l2l3_263_spill.out[2], compute_graph.const_ifm_ddr_1.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.out[0], compute_graph.l2l3_266_spill.out[0], compute_graph.l2l3_268_spill.out[0], compute_graph.l2l3_273_spill.out[0], compute_graph.spill_L3_Concat_Buffer_layer_275.out[0], compute_graph.l2l3_276_spill.out[0], compute_graph.l2l3_276_spill.out[2], compute_graph.l2l3_277_spill.out[0], compute_graph.l2l3_278_spill.out[0], compute_graph.l2l3_278_spill.out[2], compute_graph.ifm_ddr_1.out[0], compute_graph.spill_L3_Concat_Buffer_layer_279.out[0], compute_graph.l2l3_281_spill.out[0], compute_graph.l2l3_281_spill.out[2], compute_graph.spill_L3_Concat_Buffer_layer_287.out[0], compute_graph.l2l3_288_spill.out[0], compute_graph.spill_L3_Concat_Buffer_layer_292.out[0], compute_graph.spill_L3_Concat_Buffer_layer_294.out[0], compute_graph.l2l3_297_spill.out[0], compute_graph.l2l3_297_spill.out[2]) + +Column 0Row 0Channel 1 + +i23_po0(compute_graph.ifm_ddr.out[1], compute_graph.l2l3_1_spill.out[1], compute_graph.l2l3_8_spill.out[1], compute_graph.l2l3_10_spill.out[1], compute_graph.l2l3_12_spill.out[1], compute_graph.l2l3_12_spill.out[3], compute_graph.l2l3_13_spill.out[1], compute_graph.l2l3_15_spill.out[1], compute_graph.l2l3_17_spill.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.out[1], compute_graph.l2l3_21_spill.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.out[1], compute_graph.l2l3_25_spill.out[1], compute_graph.l2l3_32_spill.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.out[1], compute_graph.l2l3_37_spill.out[1], compute_graph.l2l3_44_spill.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.out[1], compute_graph.l2l3_49_spill.out[1], compute_graph.l2l3_56_spill.out[1], compute_graph.l2l3_64_spill.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.out[1], compute_graph.l2l3_114_spill.out[1], compute_graph.l2l3_121_spill.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.out[1], compute_graph.l2l3_134_spill.out[1], compute_graph.l2l3_141_spill.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.out[1], compute_graph.l2l3_154_spill.out[1], compute_graph.l2l3_161_spill.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.out[1], compute_graph.l2l3_168_spill.out[1], compute_graph.l2l3_173_spill.out[2], compute_graph.l2l3_174_spill.out[1], compute_graph.l2l3_181_spill.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.out[1], compute_graph.l2l3_188_spill.out[1], compute_graph.l2l3_193_spill.out[2], compute_graph.l2l3_194_spill.out[1], compute_graph.l2l3_201_spill.out[1], compute_graph.l2l3_208_spill.out[2], compute_graph.l2l3_209_spill.out[1], compute_graph.l2l3_213_spill.out[1], compute_graph.l2l3_214_spill.out[1], compute_graph.l2l3_214_spill.out[3], compute_graph.l2l3_215_spill.out[1], compute_graph.l2l3_216_spill.out[1], compute_graph.l2l3_216_spill.out[3], compute_graph.ifm_ddr_4.out[1], compute_graph.ifm_ddr_4.out[3], compute_graph.ifm_ddr_4.out[5], compute_graph.spill_L3_Concat_Buffer_layer_217.out[1], compute_graph.l2l3_219_spill.out[1], compute_graph.l2l3_219_spill.out[3], compute_graph.l2l3_220_spill.out[1], compute_graph.l2l3_220_spill.out[3], compute_graph.const_ifm_ddr_3.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.out[1], compute_graph.l2l3_223_spill.out[1], compute_graph.spill_L3_Concat_Buffer_layer_225.out[1], compute_graph.spill_L3_Concat_Buffer_layer_230.out[1], compute_graph.l2l3_231_spill.out[1], compute_graph.l2l3_232_spill.out[1], compute_graph.l2l3_233_spill.out[1], compute_graph.spill_L3_Concat_Buffer_layer_236.out[1], compute_graph.l2l3_237_spill.out[1], compute_graph.l2l3_237_spill.out[3], compute_graph.l2l3_238_spill.out[1], compute_graph.l2l3_239_spill.out[1], compute_graph.l2l3_239_spill.out[3], compute_graph.ifm_ddr_3.out[1], compute_graph.ifm_ddr_3.out[3], compute_graph.ifm_ddr_3.out[5], compute_graph.spill_L3_Concat_Buffer_layer_240.out[1], compute_graph.l2l3_242_spill.out[1], compute_graph.l2l3_242_spill.out[3], compute_graph.l2l3_243_spill.out[1], compute_graph.l2l3_243_spill.out[3], compute_graph.const_ifm_ddr_2.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.out[1], compute_graph.l2l3_246_spill.out[1], compute_graph.spill_L3_Concat_Buffer_layer_248.out[1], compute_graph.spill_L3_Concat_Buffer_layer_253.out[1], compute_graph.l2l3_254_spill.out[1], compute_graph.spill_L3_Concat_Buffer_layer_256.out[1], compute_graph.l2l3_257_spill.out[1], compute_graph.l2l3_257_spill.out[3], compute_graph.ifm_ddr_2.out[2], compute_graph.ifm_ddr_2.out[4], compute_graph.l2l3_260_spill.out[1], compute_graph.l2l3_262_spill.out[1], compute_graph.l2l3_262_spill.out[3], compute_graph.l2l3_263_spill.out[1], compute_graph.l2l3_263_spill.out[3], compute_graph.const_ifm_ddr_1.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.out[1], compute_graph.l2l3_266_spill.out[1], compute_graph.l2l3_268_spill.out[1], compute_graph.l2l3_273_spill.out[1], compute_graph.spill_L3_Concat_Buffer_layer_275.out[1], compute_graph.l2l3_276_spill.out[1], compute_graph.l2l3_276_spill.out[3], compute_graph.l2l3_277_spill.out[1], compute_graph.l2l3_278_spill.out[1], compute_graph.l2l3_278_spill.out[3], compute_graph.ifm_ddr_1.out[1], compute_graph.spill_L3_Concat_Buffer_layer_279.out[1], compute_graph.l2l3_281_spill.out[1], compute_graph.l2l3_281_spill.out[3], compute_graph.spill_L3_Concat_Buffer_layer_287.out[1], compute_graph.l2l3_288_spill.out[1], compute_graph.spill_L3_Concat_Buffer_layer_292.out[1], compute_graph.spill_L3_Concat_Buffer_layer_294.out[1], compute_graph.l2l3_297_spill.out[1], compute_graph.l2l3_297_spill.out[3]) + +Column 0Row 0Channel 0 + +i24_pi0(compute_graph.l2l3_1_spill.in[0], compute_graph.l2l3_6_spill.in[0], compute_graph.l2l3_7_spill.in[0], compute_graph.l2l3_8_spill.in[0], compute_graph.l2l3_9_spill.in[0], compute_graph.l2l3_10_spill.in[0], compute_graph.l2l3_11_spill.in[0], compute_graph.l2l3_12_spill.in[0], compute_graph.l2l3_13_spill.in[0], compute_graph.l2l3_15_spill.in[0], compute_graph.l2l3_19_spill.in[0], compute_graph.l2l3_20_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.in[0], compute_graph.l2l3_24_spill.in[0], compute_graph.l2l3_31_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.in[0], compute_graph.l2l3_36_spill.in[0], compute_graph.l2l3_43_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.in[0], compute_graph.l2l3_48_spill.in[0], compute_graph.l2l3_55_spill.in[0], compute_graph.l2l3_62_spill.in[0], compute_graph.l2l3_63_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.in[0], compute_graph.l2l3_113_spill.in[0], compute_graph.l2l3_120_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.in[0], compute_graph.l2l3_133_spill.in[0], compute_graph.l2l3_140_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.in[0], compute_graph.l2l3_153_spill.in[0], compute_graph.l2l3_160_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.in[0], compute_graph.l2l3_167_spill.in[0], compute_graph.l2l3_168_spill.in[0], compute_graph.l2l3_172_spill.in[0], compute_graph.l2l3_173_spill.in[0], compute_graph.l2l3_180_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.in[0], compute_graph.l2l3_187_spill.in[0], compute_graph.l2l3_188_spill.in[0], compute_graph.l2l3_192_spill.in[0], compute_graph.l2l3_193_spill.in[0], compute_graph.l2l3_200_spill.in[0], compute_graph.l2l3_207_spill.in[0], compute_graph.l2l3_208_spill.in[0], compute_graph.l2l3_212_spill.in[0], compute_graph.l2l3_214_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_217.in[0], compute_graph.spill_L3_Concat_Buffer_layer_217.in[2], compute_graph.l2l3_219_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.in[0], compute_graph.spill_L3_Concat_Buffer_layer_225.in[0], compute_graph.spill_L3_Concat_Buffer_layer_225.in[2], compute_graph.ofm_ddr_0_l2l3_229_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_230.in[0], compute_graph.spill_L3_Concat_Buffer_layer_230.in[2], compute_graph.l2l3_233_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_236.in[0], compute_graph.spill_L3_Concat_Buffer_layer_236.in[2], compute_graph.spill_L3_Concat_Buffer_layer_236.in[4], compute_graph.l2l3_237_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_240.in[0], compute_graph.spill_L3_Concat_Buffer_layer_240.in[2], compute_graph.l2l3_242_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.in[0], compute_graph.spill_L3_Concat_Buffer_layer_248.in[0], compute_graph.spill_L3_Concat_Buffer_layer_248.in[2], compute_graph.ofm_ddr_1_l2l3_252_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_253.in[0], compute_graph.spill_L3_Concat_Buffer_layer_253.in[2], compute_graph.spill_L3_Concat_Buffer_layer_256.in[2], compute_graph.l2l3_262_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.in[0], compute_graph.l2l3_267_spill.in[0], compute_graph.ofm_ddr_2_l2l3_272_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[4], compute_graph.spill_L3_Concat_Buffer_layer_279.in[0], compute_graph.spill_L3_Concat_Buffer_layer_279.in[2], compute_graph.l2l3_281_spill.in[0], compute_graph.l2l3_283_spill.in[0], compute_graph.l2l3_284_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_287.in[0], compute_graph.spill_L3_Concat_Buffer_layer_287.in[2], compute_graph.l2l3_289_spill.in[0], compute_graph.l2l3_290_spill.in[0], compute_graph.ofm_ddr_3_l2l3_291_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_292.in[0], compute_graph.spill_L3_Concat_Buffer_layer_292.in[2], compute_graph.l2l3_296_spill.in[0], compute_graph.l2l3_297_spill.in[0], compute_graph.ofm_ddr_4.in[0], compute_graph.l2l3_301_spill.in[0], compute_graph.ofm_ddr_5.in[0]) + +Column 0Row 0Channel 1 + +i25_pi0(compute_graph.l2l3_1_spill.in[1], compute_graph.l2l3_6_spill.in[1], compute_graph.l2l3_7_spill.in[1], compute_graph.l2l3_8_spill.in[1], compute_graph.l2l3_10_spill.in[1], compute_graph.l2l3_12_spill.in[1], compute_graph.l2l3_15_spill.in[1], compute_graph.l2l3_19_spill.in[1], compute_graph.l2l3_20_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.in[1], compute_graph.l2l3_24_spill.in[1], compute_graph.l2l3_31_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.in[1], compute_graph.l2l3_36_spill.in[1], compute_graph.l2l3_43_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.in[1], compute_graph.l2l3_48_spill.in[1], compute_graph.l2l3_55_spill.in[1], compute_graph.l2l3_62_spill.in[1], compute_graph.l2l3_63_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.in[1], compute_graph.l2l3_113_spill.in[1], compute_graph.l2l3_120_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.in[1], compute_graph.l2l3_133_spill.in[1], compute_graph.l2l3_140_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.in[1], compute_graph.l2l3_153_spill.in[1], compute_graph.l2l3_160_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.in[1], compute_graph.l2l3_167_spill.in[1], compute_graph.l2l3_168_spill.in[1], compute_graph.l2l3_172_spill.in[1], compute_graph.l2l3_173_spill.in[1], compute_graph.l2l3_180_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.in[1], compute_graph.l2l3_187_spill.in[1], compute_graph.l2l3_188_spill.in[1], compute_graph.l2l3_192_spill.in[1], compute_graph.l2l3_193_spill.in[1], compute_graph.l2l3_200_spill.in[1], compute_graph.l2l3_207_spill.in[1], compute_graph.l2l3_208_spill.in[1], compute_graph.l2l3_212_spill.in[1], compute_graph.l2l3_214_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_217.in[1], compute_graph.spill_L3_Concat_Buffer_layer_217.in[3], compute_graph.l2l3_219_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.in[1], compute_graph.spill_L3_Concat_Buffer_layer_225.in[1], compute_graph.spill_L3_Concat_Buffer_layer_225.in[3], compute_graph.ofm_ddr_0_l2l3_229_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_230.in[1], compute_graph.spill_L3_Concat_Buffer_layer_230.in[3], compute_graph.l2l3_233_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_236.in[1], compute_graph.spill_L3_Concat_Buffer_layer_236.in[3], compute_graph.spill_L3_Concat_Buffer_layer_236.in[5], compute_graph.l2l3_237_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_240.in[1], compute_graph.spill_L3_Concat_Buffer_layer_240.in[3], compute_graph.l2l3_242_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.in[1], compute_graph.spill_L3_Concat_Buffer_layer_248.in[1], compute_graph.spill_L3_Concat_Buffer_layer_248.in[3], compute_graph.ofm_ddr_1_l2l3_252_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_253.in[1], compute_graph.spill_L3_Concat_Buffer_layer_253.in[3], compute_graph.spill_L3_Concat_Buffer_layer_256.in[3], compute_graph.l2l3_262_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.in[1], compute_graph.l2l3_267_spill.in[1], compute_graph.ofm_ddr_2_l2l3_272_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[5], compute_graph.spill_L3_Concat_Buffer_layer_279.in[1], compute_graph.spill_L3_Concat_Buffer_layer_279.in[3], compute_graph.l2l3_281_spill.in[1], compute_graph.l2l3_283_spill.in[1], compute_graph.l2l3_284_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_287.in[1], compute_graph.spill_L3_Concat_Buffer_layer_287.in[3], compute_graph.l2l3_290_spill.in[1], compute_graph.ofm_ddr_3_l2l3_291_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_292.in[1], compute_graph.spill_L3_Concat_Buffer_layer_292.in[3], compute_graph.l2l3_296_spill.in[1], compute_graph.l2l3_297_spill.in[1], compute_graph.ofm_ddr_4.in[1], compute_graph.l2l3_301_spill.in[1], compute_graph.ofm_ddr_5.in[1]) + +Column 2Row 0Channel 0 + +i26_pi0(compute_graph.l2l3_2_spill.in[0], compute_graph.l2l3_scratch_0_3_spill.in[0], compute_graph.l2l3_3_spill.in[0], compute_graph.l2l3_4_spill.in[0], compute_graph.l2l3_scratch_0_5_spill.in[0], compute_graph.l2l3_5_spill.in[0], compute_graph.l2l3_9_spill.in[1], compute_graph.l2l3_11_spill.in[1], compute_graph.l2l3_13_spill.in[1], compute_graph.l2l3_14_spill.in[0], compute_graph.l2l3_16_spill.in[0], compute_graph.l2l3_scratch_0_17_spill.in[0], compute_graph.l2l3_17_spill.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.in[0], compute_graph.l2l3_scratch_0_21_spill.in[0], compute_graph.l2l3_21_spill.in[0], compute_graph.l2l3_25_spill.in[0], compute_graph.l2l3_scratch_0_32_spill.in[0], compute_graph.l2l3_scratch_1_32_spill.in[0], compute_graph.l2l3_32_spill.in[0], compute_graph.l2l3_37_spill.in[0], compute_graph.l2l3_scratch_0_44_spill.in[0], compute_graph.l2l3_scratch_1_44_spill.in[0], compute_graph.l2l3_44_spill.in[0], compute_graph.l2l3_49_spill.in[0], compute_graph.l2l3_scratch_0_56_spill.in[0], compute_graph.l2l3_scratch_1_56_spill.in[0], compute_graph.l2l3_56_spill.in[0], compute_graph.l2l3_59_spill.in[0], compute_graph.l2l3_scratch_0_64_spill.in[0], compute_graph.l2l3_64_spill.in[0], compute_graph.l2l3_114_spill.in[0], compute_graph.l2l3_scratch_0_121_spill.in[0], compute_graph.l2l3_scratch_1_121_spill.in[0], compute_graph.l2l3_121_spill.in[0], compute_graph.l2l3_134_spill.in[0], compute_graph.l2l3_scratch_0_141_spill.in[0], compute_graph.l2l3_scratch_1_141_spill.in[0], compute_graph.l2l3_141_spill.in[0], compute_graph.l2l3_154_spill.in[0], compute_graph.l2l3_scratch_1_161_spill.in[0], compute_graph.l2l3_scratch_0_161_spill.in[0], compute_graph.l2l3_161_spill.in[0], compute_graph.l2l3_164_spill.in[0], compute_graph.l2l3_169_spill.in[0], compute_graph.l2l3_174_spill.in[0], compute_graph.l2l3_scratch_0_181_spill.in[0], compute_graph.l2l3_scratch_1_181_spill.in[0], compute_graph.l2l3_181_spill.in[0], compute_graph.l2l3_184_spill.in[0], compute_graph.l2l3_189_spill.in[0], compute_graph.l2l3_194_spill.in[0], compute_graph.l2l3_scratch_0_201_spill.in[0], compute_graph.l2l3_scratch_1_201_spill.in[0], compute_graph.l2l3_201_spill.in[0], compute_graph.l2l3_204_spill.in[0], compute_graph.l2l3_209_spill.in[0], compute_graph.l2l3_scratch_0_213_spill.in[0], compute_graph.l2l3_scratch_1_213_spill.in[0], compute_graph.l2l3_213_spill.in[0], compute_graph.l2l3_215_spill.in[0], compute_graph.l2l3_216_spill.in[0], compute_graph.l2l3_220_spill.in[0], compute_graph.l2l3_223_spill.in[0], compute_graph.l2l3_231_spill.in[0], compute_graph.l2l3_232_spill.in[0], compute_graph.l2l3_238_spill.in[0], compute_graph.l2l3_239_spill.in[0], compute_graph.l2l3_243_spill.in[0], compute_graph.l2l3_246_spill.in[0], compute_graph.l2l3_254_spill.in[0], compute_graph.l2l3_255_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_256.in[0], compute_graph.spill_L3_Concat_Buffer_layer_256.in[4], compute_graph.l2l3_257_spill.in[0], compute_graph.l2l3_258_spill.in[0], compute_graph.l2l3_259_spill.in[0], compute_graph.l2l3_260_spill.in[0], compute_graph.l2l3_263_spill.in[0], compute_graph.l2l3_266_spill.in[0], compute_graph.l2l3_268_spill.in[0], compute_graph.l2l3_273_spill.in[0], compute_graph.l2l3_276_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[2], compute_graph.l2l3_277_spill.in[0], compute_graph.l2l3_278_spill.in[0], compute_graph.l2l3_280_spill.in[0], compute_graph.l2l3_282_spill.in[0], compute_graph.l2l3_285_spill.in[0], compute_graph.l2l3_288_spill.in[0], compute_graph.l2l3_289_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_294.in[0], compute_graph.spill_L3_Concat_Buffer_layer_294.in[2], compute_graph.l2l3_295_spill.in[0], compute_graph.l2l3_298_spill.in[0], compute_graph.l2l3_300_spill.in[0]) + +Column 2Row 0Channel 1 + +i27_pi0(compute_graph.l2l3_2_spill.in[1], compute_graph.l2l3_14_spill.in[1], compute_graph.l2l3_16_spill.in[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.in[1], compute_graph.l2l3_59_spill.in[1], compute_graph.l2l3_164_spill.in[1], compute_graph.l2l3_169_spill.in[1], compute_graph.l2l3_184_spill.in[1], compute_graph.l2l3_189_spill.in[1], compute_graph.l2l3_204_spill.in[1], compute_graph.l2l3_215_spill.in[1], compute_graph.l2l3_216_spill.in[1], compute_graph.l2l3_220_spill.in[1], compute_graph.l2l3_223_spill.in[1], compute_graph.l2l3_231_spill.in[1], compute_graph.l2l3_232_spill.in[1], compute_graph.l2l3_238_spill.in[1], compute_graph.l2l3_239_spill.in[1], compute_graph.l2l3_243_spill.in[1], compute_graph.l2l3_246_spill.in[1], compute_graph.l2l3_254_spill.in[1], compute_graph.l2l3_255_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_256.in[1], compute_graph.spill_L3_Concat_Buffer_layer_256.in[5], compute_graph.l2l3_257_spill.in[1], compute_graph.l2l3_258_spill.in[1], compute_graph.l2l3_259_spill.in[1], compute_graph.l2l3_263_spill.in[1], compute_graph.l2l3_266_spill.in[1], compute_graph.l2l3_276_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[3], compute_graph.l2l3_277_spill.in[1], compute_graph.l2l3_278_spill.in[1], compute_graph.l2l3_280_spill.in[1], compute_graph.l2l3_282_spill.in[1], compute_graph.l2l3_285_spill.in[1], compute_graph.l2l3_288_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_294.in[3], compute_graph.spill_L3_Concat_Buffer_layer_294.in[1], compute_graph.l2l3_295_spill.in[1], compute_graph.l2l3_298_spill.in[1], compute_graph.l2l3_300_spill.in[1]) + +Column 2Row 0Channel 0 + +i28_po0(compute_graph.l2l3_5_spill.out[0], compute_graph.l2l3_5_spill.out[2], compute_graph.l2l3_5_spill.out[4], compute_graph.l2l3_5_spill.out[6], compute_graph.const_ifm_ddr_5.out[0], compute_graph.l2l3_6_spill.out[0], compute_graph.const_ifm_ddr_4.out[0], compute_graph.l2l3_7_spill.out[0], compute_graph.l2l3_8_spill.out[2], compute_graph.l2l3_9_spill.out[0], compute_graph.l2l3_11_spill.out[0], compute_graph.l2l3_14_spill.out[0], compute_graph.l2l3_19_spill.out[0], compute_graph.l2l3_59_spill.out[0], compute_graph.l2l3_62_spill.out[0], compute_graph.l2l3_164_spill.out[0], compute_graph.l2l3_167_spill.out[0], compute_graph.l2l3_169_spill.out[0], compute_graph.l2l3_172_spill.out[0], compute_graph.l2l3_184_spill.out[0], compute_graph.l2l3_187_spill.out[0], compute_graph.l2l3_189_spill.out[0], compute_graph.l2l3_192_spill.out[0], compute_graph.l2l3_204_spill.out[0], compute_graph.l2l3_207_spill.out[0], compute_graph.l2l3_255_spill.out[0], compute_graph.ifm_ddr_1.out[2], compute_graph.ifm_ddr_1.out[4], compute_graph.l2l3_280_spill.out[0], compute_graph.l2l3_282_spill.out[0], compute_graph.l2l3_282_spill.out[2], compute_graph.const_ifm_ddr.out[0], compute_graph.l2l3_283_spill.out[0], compute_graph.l2l3_284_spill.out[0], compute_graph.l2l3_285_spill.out[0], compute_graph.l2l3_289_spill.out[0], compute_graph.l2l3_290_spill.out[0], compute_graph.l2l3_295_spill.out[0], compute_graph.l2l3_296_spill.out[0], compute_graph.l2l3_298_spill.out[0], compute_graph.l2l3_300_spill.out[0], compute_graph.l2l3_301_spill.out[0]) + +Column 2Row 0Channel 1 + +i29_po0(compute_graph.l2l3_5_spill.out[1], compute_graph.l2l3_5_spill.out[3], compute_graph.l2l3_5_spill.out[5], compute_graph.l2l3_5_spill.out[7], compute_graph.const_ifm_ddr_5.out[1], compute_graph.l2l3_6_spill.out[1], compute_graph.const_ifm_ddr_4.out[1], compute_graph.l2l3_7_spill.out[1], compute_graph.l2l3_8_spill.out[3], compute_graph.l2l3_9_spill.out[1], compute_graph.l2l3_11_spill.out[1], compute_graph.l2l3_14_spill.out[1], compute_graph.l2l3_19_spill.out[1], compute_graph.l2l3_59_spill.out[1], compute_graph.l2l3_62_spill.out[1], compute_graph.l2l3_164_spill.out[1], compute_graph.l2l3_167_spill.out[1], compute_graph.l2l3_169_spill.out[1], compute_graph.l2l3_172_spill.out[1], compute_graph.l2l3_184_spill.out[1], compute_graph.l2l3_187_spill.out[1], compute_graph.l2l3_189_spill.out[1], compute_graph.l2l3_192_spill.out[1], compute_graph.l2l3_204_spill.out[1], compute_graph.l2l3_207_spill.out[1], compute_graph.l2l3_255_spill.out[1], compute_graph.ifm_ddr_1.out[3], compute_graph.ifm_ddr_1.out[5], compute_graph.l2l3_280_spill.out[1], compute_graph.l2l3_282_spill.out[1], compute_graph.l2l3_282_spill.out[3], compute_graph.const_ifm_ddr.out[1], compute_graph.l2l3_283_spill.out[1], compute_graph.l2l3_284_spill.out[1], compute_graph.l2l3_285_spill.out[1], compute_graph.l2l3_289_spill.out[1], compute_graph.l2l3_290_spill.out[1], compute_graph.l2l3_295_spill.out[1], compute_graph.l2l3_296_spill.out[1], compute_graph.l2l3_298_spill.out[1], compute_graph.l2l3_300_spill.out[1], compute_graph.l2l3_301_spill.out[1]) + +INFO: [aiecompiler 77-5545] Generating overlay graph information at 'Work/temp/top_folded.json' +INFO: [aiecompiler 77-22451] ### Entering External Buffer Coalesing Pass +INFO: [aiecompiler 77-22449] ### Applying external buffer coalesing file /app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/flexmlrt-hsi.json +INFO: [aiecompiler 77-23091] Parsing MeIR graph to create runtime_buffer_map and qualified_to_parent_name_map +INFO: [aiecompiler 77-23001] Buffer Coalescing Actions +runtime_buffers_map entries +Name: coalesed_spills + xrt_id: 2 + size_in_bytes: 69157168 + extra_padding_before: 0 + extra_padding_before: 0 +Name: coalesed_weights + xrt_id: 0 + size_in_bytes: 11809920 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ifm_ddr + xrt_id: 1 + size_in_bytes: 471040 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ifm_ddr_1 + xrt_id: 10 + size_in_bytes: 460800 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ifm_ddr_2 + xrt_id: 8 + size_in_bytes: 172800 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ifm_ddr_3 + xrt_id: 6 + size_in_bytes: 73600 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ifm_ddr_4 + xrt_id: 4 + size_in_bytes: 30720 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_0_l2l3_229_spill + xrt_id: 5 + size_in_bytes: 30720 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_1_l2l3_252_spill + xrt_id: 7 + size_in_bytes: 73600 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_2_l2l3_272_spill + xrt_id: 9 + size_in_bytes: 172800 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_3_l2l3_291_spill + xrt_id: 11 + size_in_bytes: 460800 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_4 + xrt_id: 12 + size_in_bytes: 921600 + extra_padding_before: 0 + extra_padding_before: 0 +Name: compute_graph.ofm_ddr_5 + xrt_id: 13 + size_in_bytes: 921600 + extra_padding_before: 0 + extra_padding_before: 0 +Name: runtime_control_packet + xrt_id: 3 + size_in_bytes: 0 + extra_padding_before: 0 + extra_padding_before: 0 + +INFO: [aiecompiler 77-22450] ### Done with External Buffer Coalesing Pass +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_8_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_8_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_13_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_13_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_14_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_14_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_15_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_15_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_16_l2_wts is 6. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_16_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_18_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_18_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_19_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_19_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_20_l2_wts is 2. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_20_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_22_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_22_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_23_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_23_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_24_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_24_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_27_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_27_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_28_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_28_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_34_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_34_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_35_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_35_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_36_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_36_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_39_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_39_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_40_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_40_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_46_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_46_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_47_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_47_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_48_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_48_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_51_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_51_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_52_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_52_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_58_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_58_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_59_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_59_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_65_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_65_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_70_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_70_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_71_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_71_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_76_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_76_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_81_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_81_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_82_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_82_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_87_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_87_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_92_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_92_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_93_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_93_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_98_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_98_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_103_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_103_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_104_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_104_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_109_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_109_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_116_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_116_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_117_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_117_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_123_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_123_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_124_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_124_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_129_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_129_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_136_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_136_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_137_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_137_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_143_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_143_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_144_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_144_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_149_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_149_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_156_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_156_wts_ddr is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_157_l2_wts is 1. +INFO: [aiecompiler 77-5917] Repetition count for compute_graph.Layer_157_wts_ddr is 1. +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-5917' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-404] Executing Cmd: cd Work/aie/ir + /usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +Wllvm,-O2,-fno-jump-tables,-femit-all-decls,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include empty.cc -o _header.ll + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "empty.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -femit-all-decls -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g empty.cc -o_header.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i719_wrap_transpose4d_adf_wrapper.makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I ../../.. -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i719_wrap_transpose4d_adf_wrapper.cpp -o i719_wrap_transpose4d_adf_wrapper.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "i719_wrap_transpose4d_adf_wrapper.cpp" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I../../.. -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i719_wrap_transpose4d_adf_wrapper.cpp -oi719_wrap_transpose4d_adf_wrapper.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i719_wrap_transpose4d_adf_wrapper.cpp:1: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/transpose4d_adf_wrapper.cpp:68:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 68 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +1 warning generated. +Compilation finished successfully (1 errors, 1 warnings) +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i802_wrap_resize_adf_wrapper.makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I ../../.. -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i802_wrap_resize_adf_wrapper.cpp -o i802_wrap_resize_adf_wrapper.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "i802_wrap_resize_adf_wrapper.cpp" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I../../.. -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i802_wrap_resize_adf_wrapper.cpp -oi802_wrap_resize_adf_wrapper.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i802_wrap_resize_adf_wrapper.cpp:1: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/resize_adf_wrapper.cpp:70:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 70 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +1 warning generated. +Compilation finished successfully (1 errors, 1 warnings) +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i852_wrap_slice_adf_wrapper.makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I ../../.. -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i852_wrap_slice_adf_wrapper.cpp -o i852_wrap_slice_adf_wrapper.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "i852_wrap_slice_adf_wrapper.cpp" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I../../.. -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i852_wrap_slice_adf_wrapper.cpp -oi852_wrap_slice_adf_wrapper.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i897_wrap_concat_adf_wrapper.makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I ../../.. -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp -o i897_wrap_concat_adf_wrapper.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "i897_wrap_concat_adf_wrapper.cpp" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I../../.. -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp -oi897_wrap_concat_adf_wrapper.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:66: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d.h:66: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:111:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 111 | #define LAYER_PARAM_BYTES 0 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw_params.h:70:9: note: previous definition is here + 70 | #define LAYER_PARAM_BYTES 64 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:66: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d.h:66: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:1200:9: warning: 'WT_DM_BANK' macro redefined [-Wmacro-redefined] + 1200 | #define WT_DM_BANK __aie_dm_resource_b + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:76:9: note: previous definition is here + 76 | #define WT_DM_BANK __aie_dm_resource_a + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:69: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw.h:98:13: warning: 'LOAD_64' macro redefined [-Wmacro-redefined] + 98 | #define LOAD_64(ifm,ibuff,ifm_incr) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:98:13: note: previous definition is here + 98 | #define LOAD_64(ifm,ibuff) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:71: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16.h:227:13: warning: 'LOAD_32' macro redefined [-Wmacro-redefined] + 227 | #define LOAD_32(ifm,ibuff) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:365:13: note: previous definition is here + 365 | #define LOAD_32(ifm,ibuff) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:71: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16.h:231:13: warning: 'VMUL_W4C8' macro redefined [-Wmacro-redefined] + 231 | #define VMUL_W4C8(ifm,wts,acc) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:370:13: note: previous definition is here + 370 | #define VMUL_W4C8(ifm,wts,acc) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:73: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_transpose.h:67: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_transpose_params.h:76:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 76 | #define LAYER_PARAM_BYTES 64 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16_params.h:69:9: note: previous definition is here + 69 | #define LAYER_PARAM_BYTES 0 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:746:12: warning: 'clr64' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 746 | return clr64(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331425:7: note: 'clr64' has been explicitly marked deprecated here + 331425 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64acc32 clr64(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:749:12: warning: 'clr32' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 749 | return clr32(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331508:7: note: 'clr32' has been explicitly marked deprecated here + 331508 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v32acc64 clr32(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:752:40: warning: 'clr64f' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 752 | v64accfloat chess_storage(dm4) n = clr64f(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331757:7: note: 'clr64f' has been explicitly marked deprecated here + 331757 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64accfloat clr64f(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:757:12: warning: 'clr64f' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 757 | return clr64f(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331757:7: note: 'clr64f' has been explicitly marked deprecated here + 331757 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64accfloat clr64f(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:167:9: warning: 'CHESS_STORAGE' macro redefined [-Wmacro-redefined] + 167 | #define CHESS_STORAGE( T, accs ) { \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:1020:9: note: previous definition is here + 1020 | #define CHESS_STORAGE(T, accs) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:481:13: warning: 'LOAD_MAT_B_DATA_2_8x8' macro redefined [-Wmacro-redefined] + 481 | #define LOAD_MAT_B_DATA_2_8x8(p_mat_b, iterator_mat_b, param, Y0, Y1)\ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:414:13: note: previous definition is here + 414 | #define LOAD_MAT_B_DATA_2_8x8(p_mat_b, iterator_mat_b, param, Y0, Y1, Y2, Y3)\ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:600:13: warning: 'STORE_ACC_8_4x8_FULL_DATA' macro redefined [-Wmacro-redefined] + 600 | #define STORE_ACC_8_4x8_FULL_DATA(p_mat_c, iterator_mat_c, param, C0, C1, C2, C3, C4, C5, C6, C7, p_bias_in) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:474:13: note: previous definition is here + 474 | #define STORE_ACC_8_4x8_FULL_DATA(p_mat_c, C0, C1, C2, C3) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:637:13: warning: 'MMUL_MAC_8_4x8x8' macro redefined [-Wmacro-redefined] + 637 | #define MMUL_MAC_8_4x8x8(C0, C1, C2, C3, C4, C5, C6, C7, X0, X1, X2, X3, Y0, Y1)\ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:504:13: note: previous definition is here + 504 | #define MMUL_MAC_8_4x8x8(C0, C1, C2, C3, X0, Y0, Y1, Y2, Y3)\ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:110: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/mul2d.h:58: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/mul2d_params.h:59:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 59 | #define LAYER_PARAM_BYTES 64 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_params.h:79:9: note: previous definition is here + 79 | #define LAYER_PARAM_BYTES 0 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:148: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/gelu1d.h:56: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/gelu1d_impl.h:197:59: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 197 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:151: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/tanh.h:54: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/tanh_impl.h:280:59: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 280 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:218: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/floor.h:57: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/floor_impl.h:121:55: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 121 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:236: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops.h:51: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:155:17: warning: anonymous non-C-compatible type given name for linkage purposes by typedef declaration; add a tag name here [-Wnon-c-typedef-for-linkage] + 155 | typedef struct alignas(v32uint16) + | ^ + | params_t +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:161:5: note: type is not C-compatible due to this default member initializer + 161 | int8_t ifm1_shift=0;// not support any shift + | ^~~~~~~~~~~~~~~~~ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:190:5: note: type is given name 'params_t' for linkage purposes by this typedef declaration + 190 | } params_t; + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:238: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/ceil.h:54: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/ceil_impl.h:118:55: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 118 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:332: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_asym_qdq/conv2d_asym_qdq.h:3: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq.h:7: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq_helpers.hpp:171: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq_kernel_helpers.h:782:1: warning: '/*' within block comment [-Wcomment] + 782 | /* + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:348: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm_bfp16.h:87:21: warning: extra tokens at end of #ifdef directive [-Wextra-tokens] + 87 | #ifdef __AIE_ARCH__ == 20 + | ^ + | // +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i897_wrap_concat_adf_wrapper.cpp:1: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:70:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 70 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +23 warnings generated. +Compilation finished successfully (1 errors, 23 warnings) +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i1009_wrap_resize_adf_wrapper.makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I ../../.. -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i1009_wrap_resize_adf_wrapper.cpp -o i1009_wrap_resize_adf_wrapper.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "i1009_wrap_resize_adf_wrapper.cpp" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I../../.. -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i1009_wrap_resize_adf_wrapper.cpp -oi1009_wrap_resize_adf_wrapper.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir/i1009_wrap_resize_adf_wrapper.cpp:1: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/resize_adf_wrapper.cpp:70:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 70 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +1 warning generated. +Compilation finished successfully (1 errors, 1 warnings) +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie/ir -O -j1 -f i1100_superkernels.makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I ../../.. -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp -o i1100_superkernels.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "superkernels.cpp" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I../../.. -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp -oi1100_superkernels.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:66: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d.h:66: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:111:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 111 | #define LAYER_PARAM_BYTES 0 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw_params.h:70:9: note: previous definition is here + 70 | #define LAYER_PARAM_BYTES 64 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:66: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d.h:66: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:1200:9: warning: 'WT_DM_BANK' macro redefined [-Wmacro-redefined] + 1200 | #define WT_DM_BANK __aie_dm_resource_b + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:76:9: note: previous definition is here + 76 | #define WT_DM_BANK __aie_dm_resource_a + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:69: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw.h:98:13: warning: 'LOAD_64' macro redefined [-Wmacro-redefined] + 98 | #define LOAD_64(ifm,ibuff,ifm_incr) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:98:13: note: previous definition is here + 98 | #define LOAD_64(ifm,ibuff) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:71: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16.h:227:13: warning: 'LOAD_32' macro redefined [-Wmacro-redefined] + 227 | #define LOAD_32(ifm,ibuff) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:365:13: note: previous definition is here + 365 | #define LOAD_32(ifm,ibuff) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:71: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16.h:231:13: warning: 'VMUL_W4C8' macro redefined [-Wmacro-redefined] + 231 | #define VMUL_W4C8(ifm,wts,acc) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:370:13: note: previous definition is here + 370 | #define VMUL_W4C8(ifm,wts,acc) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:73: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_transpose.h:67: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_transpose_params.h:76:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 76 | #define LAYER_PARAM_BYTES 64 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16_params.h:69:9: note: previous definition is here + 69 | #define LAYER_PARAM_BYTES 0 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:746:12: warning: 'clr64' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 746 | return clr64(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331425:7: note: 'clr64' has been explicitly marked deprecated here + 331425 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64acc32 clr64(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:749:12: warning: 'clr32' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 749 | return clr32(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331508:7: note: 'clr32' has been explicitly marked deprecated here + 331508 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v32acc64 clr32(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:752:40: warning: 'clr64f' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 752 | v64accfloat chess_storage(dm4) n = clr64f(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331757:7: note: 'clr64f' has been explicitly marked deprecated here + 331757 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64accfloat clr64f(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:757:12: warning: 'clr64f' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 757 | return clr64f(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331757:7: note: 'clr64f' has been explicitly marked deprecated here + 331757 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64accfloat clr64f(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:167:9: warning: 'CHESS_STORAGE' macro redefined [-Wmacro-redefined] + 167 | #define CHESS_STORAGE( T, accs ) { \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:1020:9: note: previous definition is here + 1020 | #define CHESS_STORAGE(T, accs) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:481:13: warning: 'LOAD_MAT_B_DATA_2_8x8' macro redefined [-Wmacro-redefined] + 481 | #define LOAD_MAT_B_DATA_2_8x8(p_mat_b, iterator_mat_b, param, Y0, Y1)\ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:414:13: note: previous definition is here + 414 | #define LOAD_MAT_B_DATA_2_8x8(p_mat_b, iterator_mat_b, param, Y0, Y1, Y2, Y3)\ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:600:13: warning: 'STORE_ACC_8_4x8_FULL_DATA' macro redefined [-Wmacro-redefined] + 600 | #define STORE_ACC_8_4x8_FULL_DATA(p_mat_c, iterator_mat_c, param, C0, C1, C2, C3, C4, C5, C6, C7, p_bias_in) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:474:13: note: previous definition is here + 474 | #define STORE_ACC_8_4x8_FULL_DATA(p_mat_c, C0, C1, C2, C3) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:637:13: warning: 'MMUL_MAC_8_4x8x8' macro redefined [-Wmacro-redefined] + 637 | #define MMUL_MAC_8_4x8x8(C0, C1, C2, C3, C4, C5, C6, C7, X0, X1, X2, X3, Y0, Y1)\ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:504:13: note: previous definition is here + 504 | #define MMUL_MAC_8_4x8x8(C0, C1, C2, C3, X0, Y0, Y1, Y2, Y3)\ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:110: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/mul2d.h:58: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/mul2d_params.h:59:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 59 | #define LAYER_PARAM_BYTES 64 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_params.h:79:9: note: previous definition is here + 79 | #define LAYER_PARAM_BYTES 0 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:148: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/gelu1d.h:56: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/gelu1d_impl.h:197:59: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 197 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:151: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/tanh.h:54: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/tanh_impl.h:280:59: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 280 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:218: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/floor.h:57: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/floor_impl.h:121:55: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 121 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:236: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops.h:51: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:155:17: warning: anonymous non-C-compatible type given name for linkage purposes by typedef declaration; add a tag name here [-Wnon-c-typedef-for-linkage] + 155 | typedef struct alignas(v32uint16) + | ^ + | params_t +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:161:5: note: type is not C-compatible due to this default member initializer + 161 | int8_t ifm1_shift=0;// not support any shift + | ^~~~~~~~~~~~~~~~~ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:190:5: note: type is given name 'params_t' for linkage purposes by this typedef declaration + 190 | } params_t; + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:238: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/ceil.h:54: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/ceil_impl.h:118:55: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 118 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:332: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_asym_qdq/conv2d_asym_qdq.h:3: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq.h:7: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq_helpers.hpp:171: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq_kernel_helpers.h:782:1: warning: '/*' within block comment [-Wcomment] + 782 | /* + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:348: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm_bfp16.h:87:21: warning: extra tokens at end of #ifdef directive [-Wextra-tokens] + 87 | #ifdef __AIE_ARCH__ == 20 + | ^ + | // +/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:604:5: warning: label at end of compound statement is a C++23 extension [-Wc++23-extensions] + 604 | } + | ^ +/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:641:5: warning: label at end of compound statement is a C++23 extension [-Wc++23-extensions] + 641 | } + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:2: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:198: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/arg_max_c8.h:57: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/arg_max_c8_impl.h:53: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/reduce_base_c8.h:322:81: warning: expression result unused [-Wunused-value] + 322 | AIE_ASSERT((params.reduce_axis > 0) && (params.reduce_axis < 8) && "[ERROR]: reduce_axis must be in range [1, 7]"); + | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ +/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend/superkernels.cpp:480:40: note: in instantiation of member function 'reduce_skeleton_c8, reduce_mean_c8_params_t>::setup' requested here + 480 | reduce_mean_c8::setup(reduce_mean_c8_params, layer_params); + | ^ +25 warnings generated. +Compilation finished successfully (1 errors, 25 warnings) +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/ir' +INFO: [aiecompiler 77-404] Executing Cmd: rm -f Work/aie/ir/i719_wrap_transpose4d_adf_wrapper.makefile Work/aie/ir/i802_wrap_resize_adf_wrapper.makefile Work/aie/ir/i852_wrap_slice_adf_wrapper.makefile Work/aie/ir/i897_wrap_concat_adf_wrapper.makefile Work/aie/ir/i1009_wrap_resize_adf_wrapper.makefile Work/aie/ir/i1100_superkernels.makefile; + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +INFO: [aiecompiler 77-404] Executing Cmd: cd Work/aie/ir + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i719_wrap_transpose4d_adf_wrapper_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i719_wrap_transpose4d_adf_wrapper.ll -spec=i719_wrap_transpose4d_adf_wrapper_spec.json > /dev/null + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i802_wrap_resize_adf_wrapper_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i802_wrap_resize_adf_wrapper.ll -spec=i802_wrap_resize_adf_wrapper_spec.json > /dev/null + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i852_wrap_slice_adf_wrapper_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i852_wrap_slice_adf_wrapper.ll -spec=i852_wrap_slice_adf_wrapper_spec.json > /dev/null + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i897_wrap_concat_adf_wrapper_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i897_wrap_concat_adf_wrapper.ll -spec=i897_wrap_concat_adf_wrapper_spec.json > /dev/null + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i1009_wrap_resize_adf_wrapper_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i1009_wrap_resize_adf_wrapper.ll -spec=i1009_wrap_resize_adf_wrapper_spec.json > /dev/null + export XILINX_CARDANO_KERNEL_ANALYSIS_OPTIONS="-spec=i1100_superkernels_spec.json -max-required-vector-alignment=64 " + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin /usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMKernelAnalysis.so -passes=kernel_analysis i1100_superkernels.ll -spec=i1100_superkernels_spec.json > /dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +INFO: [aiecompiler 77-404] Executing Cmd: cd Work/aie/ir + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i719_wrap_transpose4d_adf_wrapper_spec.json i719_wrap_transpose4d_adf_wrapper.ll -o i719_wrap_transpose4d_adf_wrapper.ll 2>/dev/null + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i802_wrap_resize_adf_wrapper_spec.json i802_wrap_resize_adf_wrapper.ll -o i802_wrap_resize_adf_wrapper.ll 2>/dev/null + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i852_wrap_slice_adf_wrapper_spec.json i852_wrap_slice_adf_wrapper.ll -o i852_wrap_slice_adf_wrapper.ll 2>/dev/null + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i897_wrap_concat_adf_wrapper_spec.json i897_wrap_concat_adf_wrapper.ll -o i897_wrap_concat_adf_wrapper.ll 2>/dev/null + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i1009_wrap_resize_adf_wrapper_spec.json i1009_wrap_resize_adf_wrapper.ll -o i1009_wrap_resize_adf_wrapper.ll 2>/dev/null + /usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOptPreProc.so -passes=XLOptPreProc -spec=i1100_superkernels_spec.json i1100_superkernels.ll -o i1100_superkernels.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +WARNING: [aiecompiler 77-4232] All nodes in partition PT0[0.01]:[i4] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT0[0.01]:[i4] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT1[0.01]:[i5] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT1[0.01]:[i5] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT2[0.01]:[i6] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT2[0.01]:[i6] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT3[0.01]:[i7] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT3[0.01]:[i7] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT4[0.01]:[i8] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT4[0.01]:[i8] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT5[0.01]:[i9] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT5[0.01]:[i9] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT6[0.01]:[i10] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT6[0.01]:[i10] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT7[0.01]:[i11] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT7[0.01]:[i11] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT8[0.01]:[i12] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT8[0.01]:[i12] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT9[0.01]:[i13] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT9[0.01]:[i13] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT10[0.01]:[i14] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT10[0.01]:[i14] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT11[0.01]:[i15] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT11[0.01]:[i15] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT12[0.01]:[i16] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT12[0.01]:[i16] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT13[0.01]:[i17] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT13[0.01]:[i17] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT14[0.01]:[i18] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT14[0.01]:[i18] have heapsize constraint. Using the sum as the partition's heapsize +WARNING: [aiecompiler 77-4232] All nodes in partition PT15[0.01]:[i19] have stacksize constraint. Using the max among those as the partition's stacksize +WARNING: [aiecompiler 77-4231] All nodes in partition PT15[0.01]:[i19] have heapsize constraint. Using the sum as the partition's heapsize +INFO: [aiecompiler 77-281] ###Writing Partition Data To JSON File Work/temp/top_partition.json +INFO: [aiecompiler 77-6291] Entering MAPPING ANALYSIS pass +INFO: [aiecompiler 77-6284] Done with MAPPING ANALYSIS pass +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/pm_reload_analysis0/timestamped_log +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f pm_reload_analysis0.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc -o ir/pm_reload_analysis0_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "pm_reload_analysis0.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc -oir/pm_reload_analysis0_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:9: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/resize_adf_wrapper.cpp:70:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 70 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:24: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/transpose4d_adf_wrapper.cpp:68:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 68 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:66: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d.h:66: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:111:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 111 | #define LAYER_PARAM_BYTES 0 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw_params.h:70:9: note: previous definition is here + 70 | #define LAYER_PARAM_BYTES 64 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:66: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d.h:66: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:1200:9: warning: 'WT_DM_BANK' macro redefined [-Wmacro-redefined] + 1200 | #define WT_DM_BANK __aie_dm_resource_b + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:76:9: note: previous definition is here + 76 | #define WT_DM_BANK __aie_dm_resource_a + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:69: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw.h:98:13: warning: 'LOAD_64' macro redefined [-Wmacro-redefined] + 98 | #define LOAD_64(ifm,ibuff,ifm_incr) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:98:13: note: previous definition is here + 98 | #define LOAD_64(ifm,ibuff) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:71: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16.h:227:13: warning: 'LOAD_32' macro redefined [-Wmacro-redefined] + 227 | #define LOAD_32(ifm,ibuff) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:365:13: note: previous definition is here + 365 | #define LOAD_32(ifm,ibuff) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:71: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16.h:231:13: warning: 'VMUL_W4C8' macro redefined [-Wmacro-redefined] + 231 | #define VMUL_W4C8(ifm,wts,acc) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:370:13: note: previous definition is here + 370 | #define VMUL_W4C8(ifm,wts,acc) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:73: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_transpose.h:67: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_transpose_params.h:76:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 76 | #define LAYER_PARAM_BYTES 64 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16_params.h:69:9: note: previous definition is here + 69 | #define LAYER_PARAM_BYTES 0 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:746:12: warning: 'clr64' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 746 | return clr64(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331425:7: note: 'clr64' has been explicitly marked deprecated here + 331425 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64acc32 clr64(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:749:12: warning: 'clr32' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 749 | return clr32(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331508:7: note: 'clr32' has been explicitly marked deprecated here + 331508 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v32acc64 clr32(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:752:40: warning: 'clr64f' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 752 | v64accfloat chess_storage(dm4) n = clr64f(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331757:7: note: 'clr64f' has been explicitly marked deprecated here + 331757 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64accfloat clr64f(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:757:12: warning: 'clr64f' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 757 | return clr64f(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331757:7: note: 'clr64f' has been explicitly marked deprecated here + 331757 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64accfloat clr64f(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:167:9: warning: 'CHESS_STORAGE' macro redefined [-Wmacro-redefined] + 167 | #define CHESS_STORAGE( T, accs ) { \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:1020:9: note: previous definition is here + 1020 | #define CHESS_STORAGE(T, accs) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:481:13: warning: 'LOAD_MAT_B_DATA_2_8x8' macro redefined [-Wmacro-redefined] + 481 | #define LOAD_MAT_B_DATA_2_8x8(p_mat_b, iterator_mat_b, param, Y0, Y1)\ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:414:13: note: previous definition is here + 414 | #define LOAD_MAT_B_DATA_2_8x8(p_mat_b, iterator_mat_b, param, Y0, Y1, Y2, Y3)\ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:600:13: warning: 'STORE_ACC_8_4x8_FULL_DATA' macro redefined [-Wmacro-redefined] + 600 | #define STORE_ACC_8_4x8_FULL_DATA(p_mat_c, iterator_mat_c, param, C0, C1, C2, C3, C4, C5, C6, C7, p_bias_in) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:474:13: note: previous definition is here + 474 | #define STORE_ACC_8_4x8_FULL_DATA(p_mat_c, C0, C1, C2, C3) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:637:13: warning: 'MMUL_MAC_8_4x8x8' macro redefined [-Wmacro-redefined] + 637 | #define MMUL_MAC_8_4x8x8(C0, C1, C2, C3, C4, C5, C6, C7, X0, X1, X2, X3, Y0, Y1)\ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:504:13: note: previous definition is here + 504 | #define MMUL_MAC_8_4x8x8(C0, C1, C2, C3, X0, Y0, Y1, Y2, Y3)\ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:110: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/mul2d.h:58: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/mul2d_params.h:59:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 59 | #define LAYER_PARAM_BYTES 64 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_params.h:79:9: note: previous definition is here + 79 | #define LAYER_PARAM_BYTES 0 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:148: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/gelu1d.h:56: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/gelu1d_impl.h:197:59: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 197 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:151: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/tanh.h:54: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/tanh_impl.h:280:59: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 280 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:218: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/floor.h:57: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/floor_impl.h:121:55: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 121 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:236: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops.h:51: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:155:17: warning: anonymous non-C-compatible type given name for linkage purposes by typedef declaration; add a tag name here [-Wnon-c-typedef-for-linkage] + 155 | typedef struct alignas(v32uint16) + | ^ + | params_t +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:161:5: note: type is not C-compatible due to this default member initializer + 161 | int8_t ifm1_shift=0;// not support any shift + | ^~~~~~~~~~~~~~~~~ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:190:5: note: type is given name 'params_t' for linkage purposes by this typedef declaration + 190 | } params_t; + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:238: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/ceil.h:54: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/ceil_impl.h:118:55: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 118 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:332: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_asym_qdq/conv2d_asym_qdq.h:3: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq.h:7: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq_helpers.hpp:171: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq_kernel_helpers.h:782:1: warning: '/*' within block comment [-Wcomment] + 782 | /* + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:348: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm_bfp16.h:87:21: warning: extra tokens at end of #ifdef directive [-Wextra-tokens] + 87 | #ifdef __AIE_ARCH__ == 20 + | ^ + | // +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/pm_reload_analysis0/src/pm_reload_analysis0.cc:25: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:70:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 70 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +25 warnings generated. +Compilation finished successfully (1 errors, 25 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/pm_reload_analysis0_orig.ll ir/i1100_superkernels.ll ir/pm_reload_analysis0_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/pm_reload_analysis0_orig.ll -o ir/pm_reload_analysis0.ll 2> pm_reload_analysis0/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/pm_reload_analysis0.ll -o ir/pm_reload_analysis0.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f pm_reload_analysis0.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release pm_reload_analysis0/scripts/pm_reload_analysis0.prx) 2>&1 |& tee -a pm_reload_analysis0/pm_reload_analysis0.log pm_reload_analysis0/timestamped_log/pm_reload_analysis0.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "pm_reload_analysis0.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/pm_reload_analysis0.ll -o../Release/chesswork495/pm_reload_analysis0.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork495 ../Release/chesswork495/pm_reload_analysis0.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend pm_reload_analysis0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +chess-backend pm_reload_analysis0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +chess-backend pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +chess-backend pm_reload_analysis0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3ou_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +chess-backend pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_buffe_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +chess-backend pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferI_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferI_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferI_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferI_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferI_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferI_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferI_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferI_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferI_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferI_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend pm_reload_analysis0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common pm_reload_analysis0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend pm_reload_analysis0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_IS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_IS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_IS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_IS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_IS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_IS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_IS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp", line 112, column 4: (loop #961) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. +--showcolor -b -Obbl --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend pm_reload_analysis0-F_Z18pm_reload_analysisv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z18pm_reload_analysisv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z18pm_reload_analysisv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z18pm_reload_analysisv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z18pm_reload_analysisv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z18pm_reload_analysisv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z18pm_reload_analysisv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -Vpm_reload_analysis0 -L --common pm_reload_analysis0-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 pm_reload_analysis0.objlist -o../pm_reload_analysis0.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/pm_reload_analysis0.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/pm_reload_analysis0" +bridge -o../Release/pm_reload_analysis0 ../Release/pm_reload_analysis0.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -cpm_reload_analysis0.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork495 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/pm_reload_analysis0 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (200 errors, 5 warnings) +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-6292] Entering Scheduler pass +INFO: [aiecompiler 77-22516] Done with mergeKernelLayerNodes() for cc0 +INFO: [aiecompiler 77-22516] Done with SchedulerUtils::identifyOfmDDRSpillingScheduleWithKernelLayer() for cc0 +INFO: [aiecompiler 77-22516] Done with SchedulerUtils::identifyBufferToBufferScheduleWithKernelLayer() for cc0 +INFO: [#UNDEF] Mask used: 0x0000ffff, Adjusted PM Id: 0x714e0000 +tileType 0, col 0, row 0 +tileType 0, col 0, row 1 +tileType 0, col 0, row 2 +tileType 0, col 0, row 3 +tileType 0, col 1, row 0 +tileType 0, col 1, row 1 +tileType 0, col 1, row 2 +tileType 0, col 1, row 3 +tileType 0, col 2, row 0 +tileType 0, col 2, row 1 +tileType 0, col 2, row 2 +tileType 0, col 2, row 3 +tileType 0, col 3, row 0 +tileType 0, col 3, row 1 +tileType 0, col 3, row 2 +tileType 0, col 3, row 3 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1012): mode(0), layer(0): {compute_graph.ifm_ddr.out[0], compute_graph.ifm_ddr.out[1], compute_graph.L2_IFM_Buffer_for_input0_0_port0.in[0], compute_graph.L2_IFM_Buffer_for_input0_0_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1328), layer(0): compute_graph.flexml_layers[0].compute_node[0][0], compute_graph.flexml_layers[0].compute_node[0][1], compute_graph.flexml_layers[0].compute_node[0][2], compute_graph.flexml_layers[0].compute_node[0][3], compute_graph.flexml_layers[0].compute_node[1][0], compute_graph.flexml_layers[0].compute_node[1][1], compute_graph.flexml_layers[0].compute_node[1][2], compute_graph.flexml_layers[0].compute_node[1][3], compute_graph.flexml_layers[0].compute_node[2][0], compute_graph.flexml_layers[0].compute_node[2][1], compute_graph.flexml_layers[0].compute_node[2][2], compute_graph.flexml_layers[0].compute_node[2][3], compute_graph.flexml_layers[0].compute_node[3][0], compute_graph.flexml_layers[0].compute_node[3][1], compute_graph.flexml_layers[0].compute_node[3][2], compute_graph.flexml_layers[0].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[0], compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[1], compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[2], compute_graph.L2_IFM_Buffer_for_input0_0_port0.out[3], compute_graph.l2_1.in[0], compute_graph.l2_1.in[1], compute_graph.l2_1.in[2], compute_graph.l2_1.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(897): mode(0), layer(0): {compute_graph.l2_1.out[0], compute_graph.l2_1.out[1], compute_graph.l2l3_1_spill.in[0], compute_graph.l2l3_1_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1013): mode(0), layer(1): {compute_graph.l2l3_1_spill.out[0], compute_graph.l2l3_1_spill.out[1], compute_graph.templated_graph_2.ifm.in[0], compute_graph.templated_graph_2.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1225): mode(0), layer(1): {compute_graph.templated_graph_2.ifm.out[0], compute_graph.templated_graph_2.ifm.out[1], compute_graph.l2l3_2_spill.in[0], compute_graph.l2l3_2_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1014): mode(0), layer(2): {compute_graph.l2l3_2_spill.out[0], compute_graph.templated_graph_3.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1226): mode(0), layer(2): {compute_graph.templated_graph_3.ifm.out[0], compute_graph.templated_graph_3.ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1227): mode(0), layer(2): {compute_graph.templated_graph_3.ofm.out[0], compute_graph.l2l3_scratch_0_3_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1015): mode(0), layer(3): {compute_graph.l2l3_scratch_0_3_spill.out[0], compute_graph.templated_graph_3.ifm2.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1228): mode(0), layer(3): {compute_graph.templated_graph_3.ifm2.out[0], compute_graph.l2l3_3_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1016): mode(0), layer(4): {compute_graph.l2l3_3_spill.out[0], compute_graph.templated_graph_4.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1230), layer(4): compute_graph.templated_graph_4.trans_comp_nd[0][0], compute_graph.templated_graph_4.trans_comp_nd[0][1], compute_graph.templated_graph_4.trans_comp_nd[0][2], compute_graph.templated_graph_4.trans_comp_nd[0][3], compute_graph.templated_graph_4.trans_comp_nd[1][0], compute_graph.templated_graph_4.trans_comp_nd[1][1], compute_graph.templated_graph_4.trans_comp_nd[1][2], compute_graph.templated_graph_4.trans_comp_nd[1][3], compute_graph.templated_graph_4.trans_comp_nd[2][0], compute_graph.templated_graph_4.trans_comp_nd[2][1], compute_graph.templated_graph_4.trans_comp_nd[2][2], compute_graph.templated_graph_4.trans_comp_nd[2][3], compute_graph.templated_graph_4.trans_comp_nd[3][0], compute_graph.templated_graph_4.trans_comp_nd[3][1], compute_graph.templated_graph_4.trans_comp_nd[3][2], compute_graph.templated_graph_4.trans_comp_nd[3][3], {compute_graph.templated_graph_4.trans_mt_ifm.out[0], compute_graph.templated_graph_4.trans_mt_ifm.out[1], compute_graph.templated_graph_4.trans_mt_ifm.out[2], compute_graph.templated_graph_4.trans_mt_ifm.out[3], compute_graph.templated_graph_4.trans_mt_ofm.in[0], compute_graph.templated_graph_4.trans_mt_ofm.in[1], compute_graph.templated_graph_4.trans_mt_ofm.in[2], compute_graph.templated_graph_4.trans_mt_ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1229): mode(3), layer(4): {compute_graph.templated_graph_4.trans_mt_ofm.out[0], compute_graph.l2l3_4_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1229) is pipelined with KernelLayerNode(1230) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1017): mode(0), layer(5): {compute_graph.l2l3_4_spill.out[0], compute_graph.templated_graph_5.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1231): mode(0), layer(5): {compute_graph.templated_graph_5.ifm.out[0], compute_graph.templated_graph_5.ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1232): mode(0), layer(5): {compute_graph.templated_graph_5.ofm.out[0], compute_graph.l2l3_scratch_0_5_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1018): mode(0), layer(6): {compute_graph.l2l3_scratch_0_5_spill.out[0], compute_graph.templated_graph_5.ifm2.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1233): mode(0), layer(6): {compute_graph.templated_graph_5.ifm2.out[0], compute_graph.l2l3_5_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 898 will not be pipelined because it's in the same layer 7 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1565), layer(7): {compute_graph.l2l3_5_spill.out[0], compute_graph.const_ifm_ddr_5.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1566), layer(7): {compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1567), layer(7): {compute_graph.l2l3_5_spill.out[1], compute_graph.const_ifm_ddr_5.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1568), layer(7): {compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1329), layer(7): compute_graph.flexml_layers[1].compute_node[0][0], compute_graph.flexml_layers[1].compute_node[0][1], compute_graph.flexml_layers[1].compute_node[0][2], compute_graph.flexml_layers[1].compute_node[0][3], compute_graph.flexml_layers[1].compute_node[1][0], compute_graph.flexml_layers[1].compute_node[1][1], compute_graph.flexml_layers[1].compute_node[1][2], compute_graph.flexml_layers[1].compute_node[1][3], compute_graph.flexml_layers[1].compute_node[2][0], compute_graph.flexml_layers[1].compute_node[2][1], compute_graph.flexml_layers[1].compute_node[2][2], compute_graph.flexml_layers[1].compute_node[2][3], compute_graph.flexml_layers[1].compute_node[3][0], compute_graph.flexml_layers[1].compute_node[3][1], compute_graph.flexml_layers[1].compute_node[3][2], compute_graph.flexml_layers[1].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.out[3], compute_graph.l2_6.in[0], compute_graph.l2_6.in[1], compute_graph.l2_6.in[2], compute_graph.l2_6.in[3]}, Scheduler computes number of sub-layer phases to be 9. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(898): mode(3), layer(7): {compute_graph.l2_6.out[0], compute_graph.l2_6.out[1], compute_graph.l2l3_6_spill.in[0], compute_graph.l2l3_6_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 899 will not be pipelined because it's in the same layer 8 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1569), layer(8): {compute_graph.l2l3_6_spill.out[0], compute_graph.const_ifm_ddr_4.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1570), layer(8): {compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1571), layer(8): {compute_graph.l2l3_6_spill.out[1], compute_graph.const_ifm_ddr_4.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1572), layer(8): {compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1330), layer(8): compute_graph.flexml_layers[2].compute_node[0][0], compute_graph.flexml_layers[2].compute_node[0][1], compute_graph.flexml_layers[2].compute_node[0][2], compute_graph.flexml_layers[2].compute_node[0][3], compute_graph.flexml_layers[2].compute_node[1][0], compute_graph.flexml_layers[2].compute_node[1][1], compute_graph.flexml_layers[2].compute_node[1][2], compute_graph.flexml_layers[2].compute_node[1][3], compute_graph.flexml_layers[2].compute_node[2][0], compute_graph.flexml_layers[2].compute_node[2][1], compute_graph.flexml_layers[2].compute_node[2][2], compute_graph.flexml_layers[2].compute_node[2][3], compute_graph.flexml_layers[2].compute_node[3][0], compute_graph.flexml_layers[2].compute_node[3][1], compute_graph.flexml_layers[2].compute_node[3][2], compute_graph.flexml_layers[2].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.out[3], compute_graph.l2_7.in[0], compute_graph.l2_7.in[1], compute_graph.l2_7.in[2], compute_graph.l2_7.in[3]}, Scheduler computes number of sub-layer phases to be 9. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(819): mode(3), layer(9): {compute_graph.Layer_8_wts_ddr.out[0], compute_graph.Layer_8_wts_ddr.out[1], compute_graph.Layer_8_l2_wts.in[0], compute_graph.Layer_8_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(819) is pipelined with KernelLayerNode(1330) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(899): mode(3), layer(8): {compute_graph.l2_7.out[0], compute_graph.l2_7.out[1], compute_graph.l2l3_7_spill.in[0], compute_graph.l2l3_7_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1026): mode(0), layer(9): {compute_graph.l2l3_7_spill.out[0], compute_graph.l2l3_7_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[3], compute_graph.l2_8.in[0], compute_graph.l2_8.in[1], compute_graph.l2_8.in[2], compute_graph.l2_8.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1331), layer(9): compute_graph.flexml_layers[3].compute_node[0][0], compute_graph.flexml_layers[3].compute_node[0][1], compute_graph.flexml_layers[3].compute_node[0][2], compute_graph.flexml_layers[3].compute_node[0][3], compute_graph.flexml_layers[3].compute_node[1][0], compute_graph.flexml_layers[3].compute_node[1][1], compute_graph.flexml_layers[3].compute_node[1][2], compute_graph.flexml_layers[3].compute_node[1][3], compute_graph.flexml_layers[3].compute_node[2][0], compute_graph.flexml_layers[3].compute_node[2][1], compute_graph.flexml_layers[3].compute_node[2][2], compute_graph.flexml_layers[3].compute_node[2][3], compute_graph.flexml_layers[3].compute_node[3][0], compute_graph.flexml_layers[3].compute_node[3][1], compute_graph.flexml_layers[3].compute_node[3][2], compute_graph.flexml_layers[3].compute_node[3][3], {compute_graph.Layer_8_l2_wts.out[0], compute_graph.Layer_8_l2_wts.out[1], compute_graph.Layer_8_l2_wts.out[2], compute_graph.Layer_8_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.out[3], compute_graph.l2_8.in[0], compute_graph.l2_8.in[1], compute_graph.l2_8.in[2], compute_graph.l2_8.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(900): mode(3), layer(9): {compute_graph.l2_8.out[0], compute_graph.l2_8.out[1], compute_graph.l2l3_8_spill.in[0], compute_graph.l2l3_8_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(900) is pipelined with KernelLayerNode(1331) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1027): mode(0), layer(10): {compute_graph.l2l3_8_spill.out[0], compute_graph.l2l3_8_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1332), layer(10): compute_graph.flexml_layers[4].compute_node[0][0], compute_graph.flexml_layers[4].compute_node[0][1], compute_graph.flexml_layers[4].compute_node[0][2], compute_graph.flexml_layers[4].compute_node[0][3], compute_graph.flexml_layers[4].compute_node[1][0], compute_graph.flexml_layers[4].compute_node[1][1], compute_graph.flexml_layers[4].compute_node[1][2], compute_graph.flexml_layers[4].compute_node[1][3], compute_graph.flexml_layers[4].compute_node[2][0], compute_graph.flexml_layers[4].compute_node[2][1], compute_graph.flexml_layers[4].compute_node[2][2], compute_graph.flexml_layers[4].compute_node[2][3], compute_graph.flexml_layers[4].compute_node[3][0], compute_graph.flexml_layers[4].compute_node[3][1], compute_graph.flexml_layers[4].compute_node[3][2], compute_graph.flexml_layers[4].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.out[3], compute_graph.l2_9.in[0], compute_graph.l2_9.in[1], compute_graph.l2_9.in[2], compute_graph.l2_9.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(901): mode(0), layer(10): {compute_graph.l2_9.out[0], compute_graph.l2_9.out[1], compute_graph.l2l3_9_spill.in[0], compute_graph.l2l3_9_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1029): mode(0), layer(11): {compute_graph.l2l3_9_spill.out[0], compute_graph.l2l3_9_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1333), layer(11): compute_graph.flexml_layers[5].compute_node[0][0], compute_graph.flexml_layers[5].compute_node[0][1], compute_graph.flexml_layers[5].compute_node[0][2], compute_graph.flexml_layers[5].compute_node[0][3], compute_graph.flexml_layers[5].compute_node[1][0], compute_graph.flexml_layers[5].compute_node[1][1], compute_graph.flexml_layers[5].compute_node[1][2], compute_graph.flexml_layers[5].compute_node[1][3], compute_graph.flexml_layers[5].compute_node[2][0], compute_graph.flexml_layers[5].compute_node[2][1], compute_graph.flexml_layers[5].compute_node[2][2], compute_graph.flexml_layers[5].compute_node[2][3], compute_graph.flexml_layers[5].compute_node[3][0], compute_graph.flexml_layers[5].compute_node[3][1], compute_graph.flexml_layers[5].compute_node[3][2], compute_graph.flexml_layers[5].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.out[3], compute_graph.l2_10.in[0], compute_graph.l2_10.in[1], compute_graph.l2_10.in[2], compute_graph.l2_10.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(902): mode(3), layer(11): {compute_graph.l2_10.out[0], compute_graph.l2_10.out[1], compute_graph.l2l3_10_spill.in[0], compute_graph.l2l3_10_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(902) is pipelined with KernelLayerNode(1333) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1030): mode(0), layer(12): {compute_graph.l2l3_10_spill.out[0], compute_graph.l2l3_10_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1334), layer(12): compute_graph.flexml_layers[6].compute_node[0][0], compute_graph.flexml_layers[6].compute_node[0][1], compute_graph.flexml_layers[6].compute_node[0][2], compute_graph.flexml_layers[6].compute_node[0][3], compute_graph.flexml_layers[6].compute_node[1][0], compute_graph.flexml_layers[6].compute_node[1][1], compute_graph.flexml_layers[6].compute_node[1][2], compute_graph.flexml_layers[6].compute_node[1][3], compute_graph.flexml_layers[6].compute_node[2][0], compute_graph.flexml_layers[6].compute_node[2][1], compute_graph.flexml_layers[6].compute_node[2][2], compute_graph.flexml_layers[6].compute_node[2][3], compute_graph.flexml_layers[6].compute_node[3][0], compute_graph.flexml_layers[6].compute_node[3][1], compute_graph.flexml_layers[6].compute_node[3][2], compute_graph.flexml_layers[6].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.out[3], compute_graph.l2_11.in[0], compute_graph.l2_11.in[1], compute_graph.l2_11.in[2], compute_graph.l2_11.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(903): mode(0), layer(12): {compute_graph.l2_11.out[0], compute_graph.l2_11.out[1], compute_graph.l2l3_11_spill.in[0], compute_graph.l2l3_11_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 904 will not be pipelined because it's in the same layer 13 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1573), layer(13): {compute_graph.l2l3_8_spill.out[2], compute_graph.l2l3_11_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1574), layer(13): {compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1575), layer(13): {compute_graph.l2l3_8_spill.out[3], compute_graph.l2l3_11_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1576), layer(13): {compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1335), layer(13): compute_graph.flexml_layers[7].compute_node[0][0], compute_graph.flexml_layers[7].compute_node[0][1], compute_graph.flexml_layers[7].compute_node[0][2], compute_graph.flexml_layers[7].compute_node[0][3], compute_graph.flexml_layers[7].compute_node[1][0], compute_graph.flexml_layers[7].compute_node[1][1], compute_graph.flexml_layers[7].compute_node[1][2], compute_graph.flexml_layers[7].compute_node[1][3], compute_graph.flexml_layers[7].compute_node[2][0], compute_graph.flexml_layers[7].compute_node[2][1], compute_graph.flexml_layers[7].compute_node[2][2], compute_graph.flexml_layers[7].compute_node[2][3], compute_graph.flexml_layers[7].compute_node[3][0], compute_graph.flexml_layers[7].compute_node[3][1], compute_graph.flexml_layers[7].compute_node[3][2], compute_graph.flexml_layers[7].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.out[3], compute_graph.l2_12.in[0], compute_graph.l2_12.in[1], compute_graph.l2_12.in[2], compute_graph.l2_12.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(820): mode(3), layer(14): {compute_graph.Layer_13_wts_ddr.out[0], compute_graph.Layer_13_wts_ddr.out[1], compute_graph.Layer_13_l2_wts.in[0], compute_graph.Layer_13_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(820) is pipelined with KernelLayerNode(1335) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(904): mode(3), layer(13): {compute_graph.l2_12.out[0], compute_graph.l2_12.out[1], compute_graph.l2l3_12_spill.in[0], compute_graph.l2l3_12_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1032): mode(0), layer(14): {compute_graph.l2l3_12_spill.out[0], compute_graph.l2l3_12_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1336), layer(14): compute_graph.flexml_layers[8].compute_node[0][0], compute_graph.flexml_layers[8].compute_node[0][1], compute_graph.flexml_layers[8].compute_node[0][2], compute_graph.flexml_layers[8].compute_node[0][3], compute_graph.flexml_layers[8].compute_node[1][0], compute_graph.flexml_layers[8].compute_node[1][1], compute_graph.flexml_layers[8].compute_node[1][2], compute_graph.flexml_layers[8].compute_node[1][3], compute_graph.flexml_layers[8].compute_node[2][0], compute_graph.flexml_layers[8].compute_node[2][1], compute_graph.flexml_layers[8].compute_node[2][2], compute_graph.flexml_layers[8].compute_node[2][3], compute_graph.flexml_layers[8].compute_node[3][0], compute_graph.flexml_layers[8].compute_node[3][1], compute_graph.flexml_layers[8].compute_node[3][2], compute_graph.flexml_layers[8].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.out[3], compute_graph.Layer_13_l2_wts.out[0], compute_graph.Layer_13_l2_wts.out[1], compute_graph.Layer_13_l2_wts.out[2], compute_graph.Layer_13_l2_wts.out[3], compute_graph.l2_13.in[0], compute_graph.l2_13.in[1], compute_graph.l2_13.in[2], compute_graph.l2_13.in[3]}, Scheduler computes number of sub-layer phases to be 8. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(821): mode(3), layer(15): {compute_graph.Layer_14_wts_ddr.out[0], compute_graph.Layer_14_wts_ddr.out[1], compute_graph.Layer_14_l2_wts.in[0], compute_graph.Layer_14_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(821) is pipelined with KernelLayerNode(1336) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(905): mode(0), layer(14): {compute_graph.l2_13.out[0], compute_graph.l2_13.out[1], compute_graph.l2l3_13_spill.in[0], compute_graph.l2l3_13_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1577), layer(15): {compute_graph.l2l3_12_spill.out[2], compute_graph.l2l3_13_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1578), layer(15): {compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1579), layer(15): {compute_graph.l2l3_12_spill.out[3], compute_graph.l2l3_13_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1580), layer(15): {compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[3], compute_graph.l2_14.in[0], compute_graph.l2_14.in[1], compute_graph.l2_14.in[2], compute_graph.l2_14.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1337), layer(15): compute_graph.flexml_layers[9].compute_node[0][0], compute_graph.flexml_layers[9].compute_node[0][1], compute_graph.flexml_layers[9].compute_node[0][2], compute_graph.flexml_layers[9].compute_node[0][3], compute_graph.flexml_layers[9].compute_node[1][0], compute_graph.flexml_layers[9].compute_node[1][1], compute_graph.flexml_layers[9].compute_node[1][2], compute_graph.flexml_layers[9].compute_node[1][3], compute_graph.flexml_layers[9].compute_node[2][0], compute_graph.flexml_layers[9].compute_node[2][1], compute_graph.flexml_layers[9].compute_node[2][2], compute_graph.flexml_layers[9].compute_node[2][3], compute_graph.flexml_layers[9].compute_node[3][0], compute_graph.flexml_layers[9].compute_node[3][1], compute_graph.flexml_layers[9].compute_node[3][2], compute_graph.flexml_layers[9].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.out[3], compute_graph.Layer_14_l2_wts.out[0], compute_graph.Layer_14_l2_wts.out[1], compute_graph.Layer_14_l2_wts.out[2], compute_graph.Layer_14_l2_wts.out[3], compute_graph.l2_14.in[0], compute_graph.l2_14.in[1], compute_graph.l2_14.in[2], compute_graph.l2_14.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(822): mode(3), layer(16): {compute_graph.Layer_15_wts_ddr.out[0], compute_graph.Layer_15_wts_ddr.out[1], compute_graph.Layer_15_l2_wts.in[0], compute_graph.Layer_15_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(822) is pipelined with KernelLayerNode(1337) +INFO: [aiecompiler 77-6295] For BufferSenderNode(1581), layer(15): {compute_graph.l2_14.out[0], compute_graph.l2_14.out[2]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1582), layer(15): {compute_graph.l2l3_14_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[2]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1583), layer(15): {compute_graph.l2_14.out[1], compute_graph.l2_14.out[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1584), layer(15): {compute_graph.l2l3_14_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1035): mode(0), layer(16): {compute_graph.l2l3_14_spill.out[0], compute_graph.l2l3_14_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[3], compute_graph.l2_15.in[0], compute_graph.l2_15.in[1], compute_graph.l2_15.in[2], compute_graph.l2_15.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1338), layer(16): compute_graph.flexml_layers[10].compute_node[0][0], compute_graph.flexml_layers[10].compute_node[0][1], compute_graph.flexml_layers[10].compute_node[0][2], compute_graph.flexml_layers[10].compute_node[0][3], compute_graph.flexml_layers[10].compute_node[1][0], compute_graph.flexml_layers[10].compute_node[1][1], compute_graph.flexml_layers[10].compute_node[1][2], compute_graph.flexml_layers[10].compute_node[1][3], compute_graph.flexml_layers[10].compute_node[2][0], compute_graph.flexml_layers[10].compute_node[2][1], compute_graph.flexml_layers[10].compute_node[2][2], compute_graph.flexml_layers[10].compute_node[2][3], compute_graph.flexml_layers[10].compute_node[3][0], compute_graph.flexml_layers[10].compute_node[3][1], compute_graph.flexml_layers[10].compute_node[3][2], compute_graph.flexml_layers[10].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.out[3], compute_graph.Layer_15_l2_wts.out[0], compute_graph.Layer_15_l2_wts.out[1], compute_graph.Layer_15_l2_wts.out[2], compute_graph.Layer_15_l2_wts.out[3], compute_graph.l2_15.in[0], compute_graph.l2_15.in[1], compute_graph.l2_15.in[2], compute_graph.l2_15.in[3]}, Scheduler computes number of sub-layer phases to be 5. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(908): mode(3), layer(16): {compute_graph.l2_15.out[0], compute_graph.l2_15.out[1], compute_graph.l2l3_15_spill.in[0], compute_graph.l2l3_15_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(908) is pipelined with KernelLayerNode(1338) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(823): mode(3), layer(17): {compute_graph.Layer_16_wts_ddr.out[0], compute_graph.Layer_16_wts_ddr.out[1], compute_graph.Layer_16_l2_wts.in[0], compute_graph.Layer_16_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(823) is pipelined with KernelLayerNode(1338) +INFO: [aiecompiler 77-23129] BufferToBufferNode 909 will not be pipelined because it's in the same layer 17 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2l3_15_spill.out[0], compute_graph.l2l3_15_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1036): mode(0), layer(17): {compute_graph.l2l3_15_spill.out[0], compute_graph.l2l3_15_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[1]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.Layer_16_l2_wts.out[0], compute_graph.Layer_16_l2_wts.out[1], compute_graph.Layer_16_l2_wts.out[2], compute_graph.Layer_16_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1339), layer(17): compute_graph.flexml_layers[11].compute_node[0][0], compute_graph.flexml_layers[11].compute_node[0][1], compute_graph.flexml_layers[11].compute_node[0][2], compute_graph.flexml_layers[11].compute_node[0][3], compute_graph.flexml_layers[11].compute_node[1][0], compute_graph.flexml_layers[11].compute_node[1][1], compute_graph.flexml_layers[11].compute_node[1][2], compute_graph.flexml_layers[11].compute_node[1][3], compute_graph.flexml_layers[11].compute_node[2][0], compute_graph.flexml_layers[11].compute_node[2][1], compute_graph.flexml_layers[11].compute_node[2][2], compute_graph.flexml_layers[11].compute_node[2][3], compute_graph.flexml_layers[11].compute_node[3][0], compute_graph.flexml_layers[11].compute_node[3][1], compute_graph.flexml_layers[11].compute_node[3][2], compute_graph.flexml_layers[11].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.out[3], compute_graph.Layer_16_l2_wts.out[0], compute_graph.Layer_16_l2_wts.out[1], compute_graph.Layer_16_l2_wts.out[2], compute_graph.Layer_16_l2_wts.out[3], compute_graph.l2_16.in[0], compute_graph.l2_16.in[1], compute_graph.l2_16.in[2], compute_graph.l2_16.in[3]}, Scheduler computes number of sub-layer phases to be 6. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(909): mode(3), layer(17): {compute_graph.l2_16.out[0], compute_graph.l2_16.out[1], compute_graph.l2l3_16_spill.in[0], compute_graph.l2l3_16_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1037): mode(0), layer(18): {compute_graph.l2l3_16_spill.out[0], compute_graph.templated_graph_17.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1234): mode(0), layer(18): {compute_graph.templated_graph_17.ifm.out[0], compute_graph.templated_graph_17.ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1235): mode(0), layer(18): {compute_graph.templated_graph_17.ofm.out[0], compute_graph.l2l3_scratch_0_17_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1038): mode(0), layer(19): {compute_graph.l2l3_scratch_0_17_spill.out[0], compute_graph.templated_graph_17.ifm2.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1236): mode(0), layer(19): {compute_graph.templated_graph_17.ifm2.out[0], compute_graph.l2l3_17_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1039): mode(0), layer(20): {compute_graph.l2l3_17_spill.out[0], compute_graph.l2l3_17_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(824): mode(4), layer(20): {compute_graph.Layer_18_wts_ddr.out[0], compute_graph.Layer_18_wts_ddr.out[1], compute_graph.Layer_18_l2_wts.in[0], compute_graph.Layer_18_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(824) is pipelined with BufferToBufferNode(1039) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1340), layer(20): compute_graph.flexml_layers[12].compute_node[0][0], compute_graph.flexml_layers[12].compute_node[0][1], compute_graph.flexml_layers[12].compute_node[0][2], compute_graph.flexml_layers[12].compute_node[0][3], compute_graph.flexml_layers[12].compute_node[1][0], compute_graph.flexml_layers[12].compute_node[1][1], compute_graph.flexml_layers[12].compute_node[1][2], compute_graph.flexml_layers[12].compute_node[1][3], compute_graph.flexml_layers[12].compute_node[2][0], compute_graph.flexml_layers[12].compute_node[2][1], compute_graph.flexml_layers[12].compute_node[2][2], compute_graph.flexml_layers[12].compute_node[2][3], compute_graph.flexml_layers[12].compute_node[3][0], compute_graph.flexml_layers[12].compute_node[3][1], compute_graph.flexml_layers[12].compute_node[3][2], compute_graph.flexml_layers[12].compute_node[3][3], {compute_graph.Layer_18_l2_wts.out[0], compute_graph.Layer_18_l2_wts.out[1], compute_graph.Layer_18_l2_wts.out[2], compute_graph.Layer_18_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.out[3], compute_graph.l2_18.in[0], compute_graph.l2_18.in[1], compute_graph.l2_18.in[2], compute_graph.l2_18.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(825): mode(3), layer(21): {compute_graph.Layer_19_wts_ddr.out[0], compute_graph.Layer_19_wts_ddr.out[1], compute_graph.Layer_19_l2_wts.in[0], compute_graph.Layer_19_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(825) is pipelined with KernelLayerNode(1340) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2_18.out[0], compute_graph.l2_18.out[1], compute_graph.l2_18.out[2], compute_graph.l2_18.out[3], compute_graph.l2_19.in[0], compute_graph.l2_19.in[1], compute_graph.l2_19.in[2], compute_graph.l2_19.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1341), layer(21): compute_graph.flexml_layers[13].compute_node[0][0], compute_graph.flexml_layers[13].compute_node[0][1], compute_graph.flexml_layers[13].compute_node[0][2], compute_graph.flexml_layers[13].compute_node[0][3], compute_graph.flexml_layers[13].compute_node[1][0], compute_graph.flexml_layers[13].compute_node[1][1], compute_graph.flexml_layers[13].compute_node[1][2], compute_graph.flexml_layers[13].compute_node[1][3], compute_graph.flexml_layers[13].compute_node[2][0], compute_graph.flexml_layers[13].compute_node[2][1], compute_graph.flexml_layers[13].compute_node[2][2], compute_graph.flexml_layers[13].compute_node[2][3], compute_graph.flexml_layers[13].compute_node[3][0], compute_graph.flexml_layers[13].compute_node[3][1], compute_graph.flexml_layers[13].compute_node[3][2], compute_graph.flexml_layers[13].compute_node[3][3], {compute_graph.Layer_19_l2_wts.out[0], compute_graph.Layer_19_l2_wts.out[1], compute_graph.Layer_19_l2_wts.out[2], compute_graph.Layer_19_l2_wts.out[3], compute_graph.l2_18.out[0], compute_graph.l2_18.out[1], compute_graph.l2_18.out[2], compute_graph.l2_18.out[3], compute_graph.l2_19.in[0], compute_graph.l2_19.in[1], compute_graph.l2_19.in[2], compute_graph.l2_19.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(910): mode(3), layer(21): {compute_graph.l2_18.out[4], compute_graph.l2_18.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(910) is pipelined with KernelLayerNode(1341) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(826): mode(3), layer(22): {compute_graph.Layer_20_wts_ddr.out[0], compute_graph.Layer_20_wts_ddr.out[1], compute_graph.Layer_20_l2_wts.in[0], compute_graph.Layer_20_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(826) is pipelined with KernelLayerNode(1341) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(911): mode(0), layer(21): {compute_graph.l2_19.out[0], compute_graph.l2_19.out[1], compute_graph.l2l3_19_spill.in[0], compute_graph.l2l3_19_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1041): mode(0), layer(22): {compute_graph.l2l3_19_spill.out[0], compute_graph.l2l3_19_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.Layer_20_l2_wts.out[0], compute_graph.Layer_20_l2_wts.out[1], compute_graph.Layer_20_l2_wts.out[2], compute_graph.Layer_20_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1342), layer(22): compute_graph.flexml_layers[14].compute_node[0][0], compute_graph.flexml_layers[14].compute_node[0][1], compute_graph.flexml_layers[14].compute_node[0][2], compute_graph.flexml_layers[14].compute_node[0][3], compute_graph.flexml_layers[14].compute_node[1][0], compute_graph.flexml_layers[14].compute_node[1][1], compute_graph.flexml_layers[14].compute_node[1][2], compute_graph.flexml_layers[14].compute_node[1][3], compute_graph.flexml_layers[14].compute_node[2][0], compute_graph.flexml_layers[14].compute_node[2][1], compute_graph.flexml_layers[14].compute_node[2][2], compute_graph.flexml_layers[14].compute_node[2][3], compute_graph.flexml_layers[14].compute_node[3][0], compute_graph.flexml_layers[14].compute_node[3][1], compute_graph.flexml_layers[14].compute_node[3][2], compute_graph.flexml_layers[14].compute_node[3][3], {compute_graph.Layer_20_l2_wts.out[0], compute_graph.Layer_20_l2_wts.out[1], compute_graph.Layer_20_l2_wts.out[2], compute_graph.Layer_20_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.out[3], compute_graph.l2_20.in[0], compute_graph.l2_20.in[1], compute_graph.l2_20.in[2], compute_graph.l2_20.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(912): mode(3), layer(22): {compute_graph.l2_20.out[0], compute_graph.l2_20.out[1], compute_graph.l2l3_20_spill.in[0], compute_graph.l2l3_20_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(912) is pipelined with KernelLayerNode(1342) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1042): mode(0), layer(23): {compute_graph.l2l3_20_spill.out[0], compute_graph.templated_graph_21.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1237): mode(0), layer(23): {compute_graph.templated_graph_21.ifm.out[0], compute_graph.templated_graph_21.ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1238): mode(0), layer(23): {compute_graph.templated_graph_21.ofm.out[0], compute_graph.l2l3_scratch_0_21_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1043): mode(0), layer(24): {compute_graph.l2l3_scratch_0_21_spill.out[0], compute_graph.templated_graph_21.ifm2.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1239): mode(0), layer(24): {compute_graph.templated_graph_21.ifm2.out[0], compute_graph.l2l3_21_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818 overlaps with memory space used by compute_graph.l2_22. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1040): mode(0), layer(25): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(827): mode(4), layer(25): {compute_graph.Layer_22_wts_ddr.out[0], compute_graph.Layer_22_wts_ddr.out[1], compute_graph.Layer_22_l2_wts.in[0], compute_graph.Layer_22_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(827) is pipelined with BufferToBufferNode(1040) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1044): mode(0), layer(25): {compute_graph.l2l3_21_spill.out[0], compute_graph.l2l3_21_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[3], compute_graph.l2_22.in[0], compute_graph.l2_22.in[1], compute_graph.l2_22.in[2], compute_graph.l2_22.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1343), layer(25): compute_graph.flexml_layers[15].compute_node[0][0], compute_graph.flexml_layers[15].compute_node[0][1], compute_graph.flexml_layers[15].compute_node[0][2], compute_graph.flexml_layers[15].compute_node[0][3], compute_graph.flexml_layers[15].compute_node[1][0], compute_graph.flexml_layers[15].compute_node[1][1], compute_graph.flexml_layers[15].compute_node[1][2], compute_graph.flexml_layers[15].compute_node[1][3], compute_graph.flexml_layers[15].compute_node[2][0], compute_graph.flexml_layers[15].compute_node[2][1], compute_graph.flexml_layers[15].compute_node[2][2], compute_graph.flexml_layers[15].compute_node[2][3], compute_graph.flexml_layers[15].compute_node[3][0], compute_graph.flexml_layers[15].compute_node[3][1], compute_graph.flexml_layers[15].compute_node[3][2], compute_graph.flexml_layers[15].compute_node[3][3], {compute_graph.Layer_22_l2_wts.out[0], compute_graph.Layer_22_l2_wts.out[1], compute_graph.Layer_22_l2_wts.out[2], compute_graph.Layer_22_l2_wts.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.out[3], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.out[3], compute_graph.l2_22.in[0], compute_graph.l2_22.in[1], compute_graph.l2_22.in[2], compute_graph.l2_22.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(828): mode(3), layer(26): {compute_graph.Layer_23_wts_ddr.out[0], compute_graph.Layer_23_wts_ddr.out[1], compute_graph.Layer_23_l2_wts.in[0], compute_graph.Layer_23_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(828) is pipelined with KernelLayerNode(1343) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2_22.out[0], compute_graph.l2_22.out[1], compute_graph.l2_22.out[2], compute_graph.l2_22.out[3], compute_graph.l2_23.in[0], compute_graph.l2_23.in[1], compute_graph.l2_23.in[2], compute_graph.l2_23.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1344), layer(26): compute_graph.flexml_layers[16].compute_node[0][0], compute_graph.flexml_layers[16].compute_node[0][1], compute_graph.flexml_layers[16].compute_node[0][2], compute_graph.flexml_layers[16].compute_node[0][3], compute_graph.flexml_layers[16].compute_node[1][0], compute_graph.flexml_layers[16].compute_node[1][1], compute_graph.flexml_layers[16].compute_node[1][2], compute_graph.flexml_layers[16].compute_node[1][3], compute_graph.flexml_layers[16].compute_node[2][0], compute_graph.flexml_layers[16].compute_node[2][1], compute_graph.flexml_layers[16].compute_node[2][2], compute_graph.flexml_layers[16].compute_node[2][3], compute_graph.flexml_layers[16].compute_node[3][0], compute_graph.flexml_layers[16].compute_node[3][1], compute_graph.flexml_layers[16].compute_node[3][2], compute_graph.flexml_layers[16].compute_node[3][3], {compute_graph.Layer_23_l2_wts.out[0], compute_graph.Layer_23_l2_wts.out[1], compute_graph.Layer_23_l2_wts.out[2], compute_graph.Layer_23_l2_wts.out[3], compute_graph.l2_22.out[0], compute_graph.l2_22.out[1], compute_graph.l2_22.out[2], compute_graph.l2_22.out[3], compute_graph.l2_23.in[0], compute_graph.l2_23.in[1], compute_graph.l2_23.in[2], compute_graph.l2_23.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(913): mode(3), layer(26): {compute_graph.l2_22.out[4], compute_graph.l2_22.out[5], compute_graph.spill_L3_Concat_Buffer_layer_256.in[2], compute_graph.spill_L3_Concat_Buffer_layer_256.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(913) is pipelined with KernelLayerNode(1344) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(829): mode(3), layer(27): {compute_graph.Layer_24_wts_ddr.out[0], compute_graph.Layer_24_wts_ddr.out[1], compute_graph.Layer_24_l2_wts.in[0], compute_graph.Layer_24_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(829) is pipelined with KernelLayerNode(1344) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2_23.out[0], compute_graph.l2_23.out[1], compute_graph.l2_23.out[2], compute_graph.l2_23.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1345), layer(27): compute_graph.flexml_layers[17].compute_node[0][0], compute_graph.flexml_layers[17].compute_node[0][1], compute_graph.flexml_layers[17].compute_node[0][2], compute_graph.flexml_layers[17].compute_node[0][3], compute_graph.flexml_layers[17].compute_node[1][0], compute_graph.flexml_layers[17].compute_node[1][1], compute_graph.flexml_layers[17].compute_node[1][2], compute_graph.flexml_layers[17].compute_node[1][3], compute_graph.flexml_layers[17].compute_node[2][0], compute_graph.flexml_layers[17].compute_node[2][1], compute_graph.flexml_layers[17].compute_node[2][2], compute_graph.flexml_layers[17].compute_node[2][3], compute_graph.flexml_layers[17].compute_node[3][0], compute_graph.flexml_layers[17].compute_node[3][1], compute_graph.flexml_layers[17].compute_node[3][2], compute_graph.flexml_layers[17].compute_node[3][3], {compute_graph.l2_23.out[0], compute_graph.l2_23.out[1], compute_graph.l2_23.out[2], compute_graph.l2_23.out[3], compute_graph.Layer_24_l2_wts.out[0], compute_graph.Layer_24_l2_wts.out[1], compute_graph.Layer_24_l2_wts.out[2], compute_graph.Layer_24_l2_wts.out[3], compute_graph.l2_24.in[0], compute_graph.l2_24.in[1], compute_graph.l2_24.in[2], compute_graph.l2_24.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(915): mode(0), layer(27): {compute_graph.l2_24.out[2], compute_graph.l2_24.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(914): mode(0), layer(27): {compute_graph.l2_24.out[0], compute_graph.l2_24.out[1], compute_graph.l2l3_24_spill.in[0], compute_graph.l2l3_24_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1046): mode(0), layer(28): {compute_graph.l2l3_24_spill.out[0], compute_graph.templated_graph_25.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1240): mode(0), layer(28): {compute_graph.templated_graph_25.trans_mt_ifm.out[0], compute_graph.templated_graph_25.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1241): mode(0), layer(28): {compute_graph.templated_graph_25.trans_mt_ofm.out[0], compute_graph.l2l3_25_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1047): mode(0), layer(29): {compute_graph.l2l3_25_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1346), layer(29): compute_graph.flexml_layers[18].compute_node[0][0], compute_graph.flexml_layers[18].compute_node[0][1], compute_graph.flexml_layers[18].compute_node[0][2], compute_graph.flexml_layers[18].compute_node[0][3], compute_graph.flexml_layers[18].compute_node[1][0], compute_graph.flexml_layers[18].compute_node[1][1], compute_graph.flexml_layers[18].compute_node[1][2], compute_graph.flexml_layers[18].compute_node[1][3], compute_graph.flexml_layers[18].compute_node[2][0], compute_graph.flexml_layers[18].compute_node[2][1], compute_graph.flexml_layers[18].compute_node[2][2], compute_graph.flexml_layers[18].compute_node[2][3], compute_graph.flexml_layers[18].compute_node[3][0], compute_graph.flexml_layers[18].compute_node[3][1], compute_graph.flexml_layers[18].compute_node[3][2], compute_graph.flexml_layers[18].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.out[3], compute_graph.l2_26.in[0], compute_graph.l2_26.in[1], compute_graph.l2_26.in[2], compute_graph.l2_26.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(830): mode(3), layer(30): {compute_graph.Layer_27_wts_ddr.out[0], compute_graph.Layer_27_wts_ddr.out[1], compute_graph.Layer_27_l2_wts.in[0], compute_graph.Layer_27_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(830) is pipelined with KernelLayerNode(1346) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1347), layer(30): compute_graph.flexml_layers[19].compute_node[0][0], compute_graph.flexml_layers[19].compute_node[0][1], compute_graph.flexml_layers[19].compute_node[0][2], compute_graph.flexml_layers[19].compute_node[0][3], compute_graph.flexml_layers[19].compute_node[1][0], compute_graph.flexml_layers[19].compute_node[1][1], compute_graph.flexml_layers[19].compute_node[1][2], compute_graph.flexml_layers[19].compute_node[1][3], compute_graph.flexml_layers[19].compute_node[2][0], compute_graph.flexml_layers[19].compute_node[2][1], compute_graph.flexml_layers[19].compute_node[2][2], compute_graph.flexml_layers[19].compute_node[2][3], compute_graph.flexml_layers[19].compute_node[3][0], compute_graph.flexml_layers[19].compute_node[3][1], compute_graph.flexml_layers[19].compute_node[3][2], compute_graph.flexml_layers[19].compute_node[3][3], {compute_graph.l2_26.out[0], compute_graph.l2_26.out[1], compute_graph.l2_26.out[2], compute_graph.l2_26.out[3], compute_graph.Layer_27_l2_wts.out[0], compute_graph.Layer_27_l2_wts.out[1], compute_graph.Layer_27_l2_wts.out[2], compute_graph.Layer_27_l2_wts.out[3], compute_graph.l2_27.in[0], compute_graph.l2_27.in[1], compute_graph.l2_27.in[2], compute_graph.l2_27.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(831): mode(3), layer(31): {compute_graph.Layer_28_wts_ddr.out[0], compute_graph.Layer_28_wts_ddr.out[1], compute_graph.Layer_28_l2_wts.in[0], compute_graph.Layer_28_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(831) is pipelined with KernelLayerNode(1347) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1348), layer(31): compute_graph.flexml_layers[20].compute_node[0][0], compute_graph.flexml_layers[20].compute_node[0][1], compute_graph.flexml_layers[20].compute_node[0][2], compute_graph.flexml_layers[20].compute_node[0][3], compute_graph.flexml_layers[20].compute_node[1][0], compute_graph.flexml_layers[20].compute_node[1][1], compute_graph.flexml_layers[20].compute_node[1][2], compute_graph.flexml_layers[20].compute_node[1][3], compute_graph.flexml_layers[20].compute_node[2][0], compute_graph.flexml_layers[20].compute_node[2][1], compute_graph.flexml_layers[20].compute_node[2][2], compute_graph.flexml_layers[20].compute_node[2][3], compute_graph.flexml_layers[20].compute_node[3][0], compute_graph.flexml_layers[20].compute_node[3][1], compute_graph.flexml_layers[20].compute_node[3][2], compute_graph.flexml_layers[20].compute_node[3][3], {compute_graph.l2_27.out[0], compute_graph.l2_27.out[1], compute_graph.l2_27.out[2], compute_graph.l2_27.out[3], compute_graph.Layer_28_l2_wts.out[0], compute_graph.Layer_28_l2_wts.out[1], compute_graph.Layer_28_l2_wts.out[2], compute_graph.Layer_28_l2_wts.out[3], compute_graph.l2_28.in[0], compute_graph.l2_28.in[1], compute_graph.l2_28.in[2], compute_graph.l2_28.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1349), layer(32): compute_graph.flexml_layers[21].compute_node[0][0], compute_graph.flexml_layers[21].compute_node[0][1], compute_graph.flexml_layers[21].compute_node[0][2], compute_graph.flexml_layers[21].compute_node[0][3], compute_graph.flexml_layers[21].compute_node[1][0], compute_graph.flexml_layers[21].compute_node[1][1], compute_graph.flexml_layers[21].compute_node[1][2], compute_graph.flexml_layers[21].compute_node[1][3], compute_graph.flexml_layers[21].compute_node[2][0], compute_graph.flexml_layers[21].compute_node[2][1], compute_graph.flexml_layers[21].compute_node[2][2], compute_graph.flexml_layers[21].compute_node[2][3], compute_graph.flexml_layers[21].compute_node[3][0], compute_graph.flexml_layers[21].compute_node[3][1], compute_graph.flexml_layers[21].compute_node[3][2], compute_graph.flexml_layers[21].compute_node[3][3], {compute_graph.l2_28.out[0], compute_graph.l2_28.out[1], compute_graph.l2_28.out[2], compute_graph.l2_28.out[3], compute_graph.l2_29.in[0], compute_graph.l2_29.in[1], compute_graph.l2_29.in[2], compute_graph.l2_29.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1350), layer(33): compute_graph.flexml_layers[22].compute_node[0][0], compute_graph.flexml_layers[22].compute_node[0][1], compute_graph.flexml_layers[22].compute_node[0][2], compute_graph.flexml_layers[22].compute_node[0][3], compute_graph.flexml_layers[22].compute_node[1][0], compute_graph.flexml_layers[22].compute_node[1][1], compute_graph.flexml_layers[22].compute_node[1][2], compute_graph.flexml_layers[22].compute_node[1][3], compute_graph.flexml_layers[22].compute_node[2][0], compute_graph.flexml_layers[22].compute_node[2][1], compute_graph.flexml_layers[22].compute_node[2][2], compute_graph.flexml_layers[22].compute_node[2][3], compute_graph.flexml_layers[22].compute_node[3][0], compute_graph.flexml_layers[22].compute_node[3][1], compute_graph.flexml_layers[22].compute_node[3][2], compute_graph.flexml_layers[22].compute_node[3][3], {compute_graph.l2_29.out[0], compute_graph.l2_29.out[1], compute_graph.l2_29.out[2], compute_graph.l2_29.out[3], compute_graph.l2_30.in[0], compute_graph.l2_30.in[1], compute_graph.l2_30.in[2], compute_graph.l2_30.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1351), layer(34): compute_graph.flexml_layers[23].compute_node[0][0], compute_graph.flexml_layers[23].compute_node[0][1], compute_graph.flexml_layers[23].compute_node[0][2], compute_graph.flexml_layers[23].compute_node[0][3], compute_graph.flexml_layers[23].compute_node[1][0], compute_graph.flexml_layers[23].compute_node[1][1], compute_graph.flexml_layers[23].compute_node[1][2], compute_graph.flexml_layers[23].compute_node[1][3], compute_graph.flexml_layers[23].compute_node[2][0], compute_graph.flexml_layers[23].compute_node[2][1], compute_graph.flexml_layers[23].compute_node[2][2], compute_graph.flexml_layers[23].compute_node[2][3], compute_graph.flexml_layers[23].compute_node[3][0], compute_graph.flexml_layers[23].compute_node[3][1], compute_graph.flexml_layers[23].compute_node[3][2], compute_graph.flexml_layers[23].compute_node[3][3], {compute_graph.l2_30.out[0], compute_graph.l2_30.out[1], compute_graph.l2_30.out[2], compute_graph.l2_30.out[3], compute_graph.l2_31.in[0], compute_graph.l2_31.in[1], compute_graph.l2_31.in[2], compute_graph.l2_31.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(916): mode(0), layer(34): {compute_graph.l2_31.out[1], compute_graph.l2l3_31_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1048): mode(0), layer(35): {compute_graph.l2l3_31_spill.out[0], compute_graph.templated_graph_32.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1242): mode(0), layer(35): {compute_graph.templated_graph_32.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_32_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1049): mode(0), layer(36): {compute_graph.l2l3_scratch_0_32_spill.out[0], compute_graph.templated_graph_32.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1243): mode(0), layer(36): {compute_graph.templated_graph_32.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_32_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1050): mode(0), layer(37): {compute_graph.l2l3_scratch_1_32_spill.out[0], compute_graph.templated_graph_32.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1244): mode(0), layer(37): {compute_graph.templated_graph_32.g2.ifm_mem.out[0], compute_graph.l2l3_32_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1045): mode(0), layer(38): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1051): mode(0), layer(38): {compute_graph.l2l3_32_spill.out[0], compute_graph.l2l3_32_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1352), layer(38): compute_graph.flexml_layers[24].compute_node[0][0], compute_graph.flexml_layers[24].compute_node[0][1], compute_graph.flexml_layers[24].compute_node[0][2], compute_graph.flexml_layers[24].compute_node[0][3], compute_graph.flexml_layers[24].compute_node[1][0], compute_graph.flexml_layers[24].compute_node[1][1], compute_graph.flexml_layers[24].compute_node[1][2], compute_graph.flexml_layers[24].compute_node[1][3], compute_graph.flexml_layers[24].compute_node[2][0], compute_graph.flexml_layers[24].compute_node[2][1], compute_graph.flexml_layers[24].compute_node[2][2], compute_graph.flexml_layers[24].compute_node[2][3], compute_graph.flexml_layers[24].compute_node[3][0], compute_graph.flexml_layers[24].compute_node[3][1], compute_graph.flexml_layers[24].compute_node[3][2], compute_graph.flexml_layers[24].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.out[3], compute_graph.l2_33.in[0], compute_graph.l2_33.in[1], compute_graph.l2_33.in[2], compute_graph.l2_33.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(832): mode(3), layer(39): {compute_graph.Layer_34_wts_ddr.out[0], compute_graph.Layer_34_wts_ddr.out[1], compute_graph.Layer_34_l2_wts.in[0], compute_graph.Layer_34_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(832) is pipelined with KernelLayerNode(1352) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1353), layer(39): compute_graph.flexml_layers[25].compute_node[0][0], compute_graph.flexml_layers[25].compute_node[0][1], compute_graph.flexml_layers[25].compute_node[0][2], compute_graph.flexml_layers[25].compute_node[0][3], compute_graph.flexml_layers[25].compute_node[1][0], compute_graph.flexml_layers[25].compute_node[1][1], compute_graph.flexml_layers[25].compute_node[1][2], compute_graph.flexml_layers[25].compute_node[1][3], compute_graph.flexml_layers[25].compute_node[2][0], compute_graph.flexml_layers[25].compute_node[2][1], compute_graph.flexml_layers[25].compute_node[2][2], compute_graph.flexml_layers[25].compute_node[2][3], compute_graph.flexml_layers[25].compute_node[3][0], compute_graph.flexml_layers[25].compute_node[3][1], compute_graph.flexml_layers[25].compute_node[3][2], compute_graph.flexml_layers[25].compute_node[3][3], {compute_graph.l2_33.out[0], compute_graph.l2_33.out[1], compute_graph.l2_33.out[2], compute_graph.l2_33.out[3], compute_graph.Layer_34_l2_wts.out[0], compute_graph.Layer_34_l2_wts.out[1], compute_graph.Layer_34_l2_wts.out[2], compute_graph.Layer_34_l2_wts.out[3], compute_graph.l2_34.in[0], compute_graph.l2_34.in[1], compute_graph.l2_34.in[2], compute_graph.l2_34.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(833): mode(3), layer(40): {compute_graph.Layer_35_wts_ddr.out[0], compute_graph.Layer_35_wts_ddr.out[1], compute_graph.Layer_35_l2_wts.in[0], compute_graph.Layer_35_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(833) is pipelined with KernelLayerNode(1353) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1354), layer(40): compute_graph.flexml_layers[26].compute_node[0][0], compute_graph.flexml_layers[26].compute_node[0][1], compute_graph.flexml_layers[26].compute_node[0][2], compute_graph.flexml_layers[26].compute_node[0][3], compute_graph.flexml_layers[26].compute_node[1][0], compute_graph.flexml_layers[26].compute_node[1][1], compute_graph.flexml_layers[26].compute_node[1][2], compute_graph.flexml_layers[26].compute_node[1][3], compute_graph.flexml_layers[26].compute_node[2][0], compute_graph.flexml_layers[26].compute_node[2][1], compute_graph.flexml_layers[26].compute_node[2][2], compute_graph.flexml_layers[26].compute_node[2][3], compute_graph.flexml_layers[26].compute_node[3][0], compute_graph.flexml_layers[26].compute_node[3][1], compute_graph.flexml_layers[26].compute_node[3][2], compute_graph.flexml_layers[26].compute_node[3][3], {compute_graph.l2_34.out[0], compute_graph.l2_34.out[1], compute_graph.l2_34.out[2], compute_graph.l2_34.out[3], compute_graph.Layer_35_l2_wts.out[0], compute_graph.Layer_35_l2_wts.out[1], compute_graph.Layer_35_l2_wts.out[2], compute_graph.Layer_35_l2_wts.out[3], compute_graph.l2_35.in[0], compute_graph.l2_35.in[1], compute_graph.l2_35.in[2], compute_graph.l2_35.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(917): mode(3), layer(40): {compute_graph.l2_34.out[4], compute_graph.l2_34.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(917) is pipelined with KernelLayerNode(1354) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(834): mode(3), layer(41): {compute_graph.Layer_36_wts_ddr.out[0], compute_graph.Layer_36_wts_ddr.out[1], compute_graph.Layer_36_l2_wts.in[0], compute_graph.Layer_36_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(834) is pipelined with KernelLayerNode(1354) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1355), layer(41): compute_graph.flexml_layers[27].compute_node[0][0], compute_graph.flexml_layers[27].compute_node[0][1], compute_graph.flexml_layers[27].compute_node[0][2], compute_graph.flexml_layers[27].compute_node[0][3], compute_graph.flexml_layers[27].compute_node[1][0], compute_graph.flexml_layers[27].compute_node[1][1], compute_graph.flexml_layers[27].compute_node[1][2], compute_graph.flexml_layers[27].compute_node[1][3], compute_graph.flexml_layers[27].compute_node[2][0], compute_graph.flexml_layers[27].compute_node[2][1], compute_graph.flexml_layers[27].compute_node[2][2], compute_graph.flexml_layers[27].compute_node[2][3], compute_graph.flexml_layers[27].compute_node[3][0], compute_graph.flexml_layers[27].compute_node[3][1], compute_graph.flexml_layers[27].compute_node[3][2], compute_graph.flexml_layers[27].compute_node[3][3], {compute_graph.l2_35.out[0], compute_graph.l2_35.out[1], compute_graph.l2_35.out[2], compute_graph.l2_35.out[3], compute_graph.Layer_36_l2_wts.out[0], compute_graph.Layer_36_l2_wts.out[1], compute_graph.Layer_36_l2_wts.out[2], compute_graph.Layer_36_l2_wts.out[3], compute_graph.l2_36.in[0], compute_graph.l2_36.in[1], compute_graph.l2_36.in[2], compute_graph.l2_36.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(918): mode(0), layer(41): {compute_graph.l2_36.out[0], compute_graph.l2_36.out[1], compute_graph.l2l3_36_spill.in[0], compute_graph.l2l3_36_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(919): mode(0), layer(41): {compute_graph.l2_36.out[2], compute_graph.l2_36.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1054): mode(0), layer(42): {compute_graph.l2l3_36_spill.out[0], compute_graph.templated_graph_37.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1245): mode(0), layer(42): {compute_graph.templated_graph_37.trans_mt_ifm.out[0], compute_graph.templated_graph_37.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1246): mode(0), layer(42): {compute_graph.templated_graph_37.trans_mt_ofm.out[0], compute_graph.l2l3_37_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1055): mode(0), layer(43): {compute_graph.l2l3_37_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1356), layer(43): compute_graph.flexml_layers[28].compute_node[0][0], compute_graph.flexml_layers[28].compute_node[0][1], compute_graph.flexml_layers[28].compute_node[0][2], compute_graph.flexml_layers[28].compute_node[0][3], compute_graph.flexml_layers[28].compute_node[1][0], compute_graph.flexml_layers[28].compute_node[1][1], compute_graph.flexml_layers[28].compute_node[1][2], compute_graph.flexml_layers[28].compute_node[1][3], compute_graph.flexml_layers[28].compute_node[2][0], compute_graph.flexml_layers[28].compute_node[2][1], compute_graph.flexml_layers[28].compute_node[2][2], compute_graph.flexml_layers[28].compute_node[2][3], compute_graph.flexml_layers[28].compute_node[3][0], compute_graph.flexml_layers[28].compute_node[3][1], compute_graph.flexml_layers[28].compute_node[3][2], compute_graph.flexml_layers[28].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.out[3], compute_graph.l2_38.in[0], compute_graph.l2_38.in[1], compute_graph.l2_38.in[2], compute_graph.l2_38.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(835): mode(3), layer(44): {compute_graph.Layer_39_wts_ddr.out[0], compute_graph.Layer_39_wts_ddr.out[1], compute_graph.Layer_39_l2_wts.in[0], compute_graph.Layer_39_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(835) is pipelined with KernelLayerNode(1356) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1357), layer(44): compute_graph.flexml_layers[29].compute_node[0][0], compute_graph.flexml_layers[29].compute_node[0][1], compute_graph.flexml_layers[29].compute_node[0][2], compute_graph.flexml_layers[29].compute_node[0][3], compute_graph.flexml_layers[29].compute_node[1][0], compute_graph.flexml_layers[29].compute_node[1][1], compute_graph.flexml_layers[29].compute_node[1][2], compute_graph.flexml_layers[29].compute_node[1][3], compute_graph.flexml_layers[29].compute_node[2][0], compute_graph.flexml_layers[29].compute_node[2][1], compute_graph.flexml_layers[29].compute_node[2][2], compute_graph.flexml_layers[29].compute_node[2][3], compute_graph.flexml_layers[29].compute_node[3][0], compute_graph.flexml_layers[29].compute_node[3][1], compute_graph.flexml_layers[29].compute_node[3][2], compute_graph.flexml_layers[29].compute_node[3][3], {compute_graph.l2_38.out[0], compute_graph.l2_38.out[1], compute_graph.l2_38.out[2], compute_graph.l2_38.out[3], compute_graph.Layer_39_l2_wts.out[0], compute_graph.Layer_39_l2_wts.out[1], compute_graph.Layer_39_l2_wts.out[2], compute_graph.Layer_39_l2_wts.out[3], compute_graph.l2_39.in[0], compute_graph.l2_39.in[1], compute_graph.l2_39.in[2], compute_graph.l2_39.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(836): mode(3), layer(45): {compute_graph.Layer_40_wts_ddr.out[0], compute_graph.Layer_40_wts_ddr.out[1], compute_graph.Layer_40_l2_wts.in[0], compute_graph.Layer_40_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(836) is pipelined with KernelLayerNode(1357) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1358), layer(45): compute_graph.flexml_layers[30].compute_node[0][0], compute_graph.flexml_layers[30].compute_node[0][1], compute_graph.flexml_layers[30].compute_node[0][2], compute_graph.flexml_layers[30].compute_node[0][3], compute_graph.flexml_layers[30].compute_node[1][0], compute_graph.flexml_layers[30].compute_node[1][1], compute_graph.flexml_layers[30].compute_node[1][2], compute_graph.flexml_layers[30].compute_node[1][3], compute_graph.flexml_layers[30].compute_node[2][0], compute_graph.flexml_layers[30].compute_node[2][1], compute_graph.flexml_layers[30].compute_node[2][2], compute_graph.flexml_layers[30].compute_node[2][3], compute_graph.flexml_layers[30].compute_node[3][0], compute_graph.flexml_layers[30].compute_node[3][1], compute_graph.flexml_layers[30].compute_node[3][2], compute_graph.flexml_layers[30].compute_node[3][3], {compute_graph.l2_39.out[0], compute_graph.l2_39.out[1], compute_graph.l2_39.out[2], compute_graph.l2_39.out[3], compute_graph.Layer_40_l2_wts.out[0], compute_graph.Layer_40_l2_wts.out[1], compute_graph.Layer_40_l2_wts.out[2], compute_graph.Layer_40_l2_wts.out[3], compute_graph.l2_40.in[0], compute_graph.l2_40.in[1], compute_graph.l2_40.in[2], compute_graph.l2_40.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1359), layer(46): compute_graph.flexml_layers[31].compute_node[0][0], compute_graph.flexml_layers[31].compute_node[0][1], compute_graph.flexml_layers[31].compute_node[0][2], compute_graph.flexml_layers[31].compute_node[0][3], compute_graph.flexml_layers[31].compute_node[1][0], compute_graph.flexml_layers[31].compute_node[1][1], compute_graph.flexml_layers[31].compute_node[1][2], compute_graph.flexml_layers[31].compute_node[1][3], compute_graph.flexml_layers[31].compute_node[2][0], compute_graph.flexml_layers[31].compute_node[2][1], compute_graph.flexml_layers[31].compute_node[2][2], compute_graph.flexml_layers[31].compute_node[2][3], compute_graph.flexml_layers[31].compute_node[3][0], compute_graph.flexml_layers[31].compute_node[3][1], compute_graph.flexml_layers[31].compute_node[3][2], compute_graph.flexml_layers[31].compute_node[3][3], {compute_graph.l2_40.out[0], compute_graph.l2_40.out[1], compute_graph.l2_40.out[2], compute_graph.l2_40.out[3], compute_graph.l2_41.in[0], compute_graph.l2_41.in[1], compute_graph.l2_41.in[2], compute_graph.l2_41.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1360), layer(47): compute_graph.flexml_layers[32].compute_node[0][0], compute_graph.flexml_layers[32].compute_node[0][1], compute_graph.flexml_layers[32].compute_node[0][2], compute_graph.flexml_layers[32].compute_node[0][3], compute_graph.flexml_layers[32].compute_node[1][0], compute_graph.flexml_layers[32].compute_node[1][1], compute_graph.flexml_layers[32].compute_node[1][2], compute_graph.flexml_layers[32].compute_node[1][3], compute_graph.flexml_layers[32].compute_node[2][0], compute_graph.flexml_layers[32].compute_node[2][1], compute_graph.flexml_layers[32].compute_node[2][2], compute_graph.flexml_layers[32].compute_node[2][3], compute_graph.flexml_layers[32].compute_node[3][0], compute_graph.flexml_layers[32].compute_node[3][1], compute_graph.flexml_layers[32].compute_node[3][2], compute_graph.flexml_layers[32].compute_node[3][3], {compute_graph.l2_41.out[0], compute_graph.l2_41.out[1], compute_graph.l2_41.out[2], compute_graph.l2_41.out[3], compute_graph.l2_42.in[0], compute_graph.l2_42.in[1], compute_graph.l2_42.in[2], compute_graph.l2_42.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1361), layer(48): compute_graph.flexml_layers[33].compute_node[0][0], compute_graph.flexml_layers[33].compute_node[0][1], compute_graph.flexml_layers[33].compute_node[0][2], compute_graph.flexml_layers[33].compute_node[0][3], compute_graph.flexml_layers[33].compute_node[1][0], compute_graph.flexml_layers[33].compute_node[1][1], compute_graph.flexml_layers[33].compute_node[1][2], compute_graph.flexml_layers[33].compute_node[1][3], compute_graph.flexml_layers[33].compute_node[2][0], compute_graph.flexml_layers[33].compute_node[2][1], compute_graph.flexml_layers[33].compute_node[2][2], compute_graph.flexml_layers[33].compute_node[2][3], compute_graph.flexml_layers[33].compute_node[3][0], compute_graph.flexml_layers[33].compute_node[3][1], compute_graph.flexml_layers[33].compute_node[3][2], compute_graph.flexml_layers[33].compute_node[3][3], {compute_graph.l2_42.out[0], compute_graph.l2_42.out[1], compute_graph.l2_42.out[2], compute_graph.l2_42.out[3], compute_graph.l2_43.in[0], compute_graph.l2_43.in[1], compute_graph.l2_43.in[2], compute_graph.l2_43.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(920): mode(0), layer(48): {compute_graph.l2_43.out[1], compute_graph.l2l3_43_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1056): mode(0), layer(49): {compute_graph.l2l3_43_spill.out[0], compute_graph.templated_graph_44.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1247): mode(0), layer(49): {compute_graph.templated_graph_44.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_44_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1057): mode(0), layer(50): {compute_graph.l2l3_scratch_0_44_spill.out[0], compute_graph.templated_graph_44.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1248): mode(0), layer(50): {compute_graph.templated_graph_44.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_44_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1058): mode(0), layer(51): {compute_graph.l2l3_scratch_1_44_spill.out[0], compute_graph.templated_graph_44.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1249): mode(0), layer(51): {compute_graph.templated_graph_44.g2.ifm_mem.out[0], compute_graph.l2l3_44_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1053): mode(0), layer(52): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1059): mode(0), layer(52): {compute_graph.l2l3_44_spill.out[0], compute_graph.l2l3_44_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1362), layer(52): compute_graph.flexml_layers[34].compute_node[0][0], compute_graph.flexml_layers[34].compute_node[0][1], compute_graph.flexml_layers[34].compute_node[0][2], compute_graph.flexml_layers[34].compute_node[0][3], compute_graph.flexml_layers[34].compute_node[1][0], compute_graph.flexml_layers[34].compute_node[1][1], compute_graph.flexml_layers[34].compute_node[1][2], compute_graph.flexml_layers[34].compute_node[1][3], compute_graph.flexml_layers[34].compute_node[2][0], compute_graph.flexml_layers[34].compute_node[2][1], compute_graph.flexml_layers[34].compute_node[2][2], compute_graph.flexml_layers[34].compute_node[2][3], compute_graph.flexml_layers[34].compute_node[3][0], compute_graph.flexml_layers[34].compute_node[3][1], compute_graph.flexml_layers[34].compute_node[3][2], compute_graph.flexml_layers[34].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.out[3], compute_graph.l2_45.in[0], compute_graph.l2_45.in[1], compute_graph.l2_45.in[2], compute_graph.l2_45.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(837): mode(3), layer(53): {compute_graph.Layer_46_wts_ddr.out[0], compute_graph.Layer_46_wts_ddr.out[1], compute_graph.Layer_46_l2_wts.in[0], compute_graph.Layer_46_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(837) is pipelined with KernelLayerNode(1362) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434 overlaps with memory space used by compute_graph.l2_46. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1052): mode(0), layer(53): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1363), layer(53): compute_graph.flexml_layers[35].compute_node[0][0], compute_graph.flexml_layers[35].compute_node[0][1], compute_graph.flexml_layers[35].compute_node[0][2], compute_graph.flexml_layers[35].compute_node[0][3], compute_graph.flexml_layers[35].compute_node[1][0], compute_graph.flexml_layers[35].compute_node[1][1], compute_graph.flexml_layers[35].compute_node[1][2], compute_graph.flexml_layers[35].compute_node[1][3], compute_graph.flexml_layers[35].compute_node[2][0], compute_graph.flexml_layers[35].compute_node[2][1], compute_graph.flexml_layers[35].compute_node[2][2], compute_graph.flexml_layers[35].compute_node[2][3], compute_graph.flexml_layers[35].compute_node[3][0], compute_graph.flexml_layers[35].compute_node[3][1], compute_graph.flexml_layers[35].compute_node[3][2], compute_graph.flexml_layers[35].compute_node[3][3], {compute_graph.l2_45.out[0], compute_graph.l2_45.out[1], compute_graph.l2_45.out[2], compute_graph.l2_45.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.out[3], compute_graph.Layer_46_l2_wts.out[0], compute_graph.Layer_46_l2_wts.out[1], compute_graph.Layer_46_l2_wts.out[2], compute_graph.Layer_46_l2_wts.out[3], compute_graph.l2_46.in[0], compute_graph.l2_46.in[1], compute_graph.l2_46.in[2], compute_graph.l2_46.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(838): mode(3), layer(54): {compute_graph.Layer_47_wts_ddr.out[0], compute_graph.Layer_47_wts_ddr.out[1], compute_graph.Layer_47_l2_wts.in[0], compute_graph.Layer_47_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(838) is pipelined with KernelLayerNode(1363) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1364), layer(54): compute_graph.flexml_layers[36].compute_node[0][0], compute_graph.flexml_layers[36].compute_node[0][1], compute_graph.flexml_layers[36].compute_node[0][2], compute_graph.flexml_layers[36].compute_node[0][3], compute_graph.flexml_layers[36].compute_node[1][0], compute_graph.flexml_layers[36].compute_node[1][1], compute_graph.flexml_layers[36].compute_node[1][2], compute_graph.flexml_layers[36].compute_node[1][3], compute_graph.flexml_layers[36].compute_node[2][0], compute_graph.flexml_layers[36].compute_node[2][1], compute_graph.flexml_layers[36].compute_node[2][2], compute_graph.flexml_layers[36].compute_node[2][3], compute_graph.flexml_layers[36].compute_node[3][0], compute_graph.flexml_layers[36].compute_node[3][1], compute_graph.flexml_layers[36].compute_node[3][2], compute_graph.flexml_layers[36].compute_node[3][3], {compute_graph.l2_46.out[0], compute_graph.l2_46.out[1], compute_graph.l2_46.out[2], compute_graph.l2_46.out[3], compute_graph.Layer_47_l2_wts.out[0], compute_graph.Layer_47_l2_wts.out[1], compute_graph.Layer_47_l2_wts.out[2], compute_graph.Layer_47_l2_wts.out[3], compute_graph.l2_47.in[0], compute_graph.l2_47.in[1], compute_graph.l2_47.in[2], compute_graph.l2_47.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(921): mode(3), layer(54): {compute_graph.l2_46.out[4], compute_graph.l2_46.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(921) is pipelined with KernelLayerNode(1364) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(839): mode(3), layer(55): {compute_graph.Layer_48_wts_ddr.out[0], compute_graph.Layer_48_wts_ddr.out[1], compute_graph.Layer_48_l2_wts.in[0], compute_graph.Layer_48_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(839) is pipelined with KernelLayerNode(1364) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1365), layer(55): compute_graph.flexml_layers[37].compute_node[0][0], compute_graph.flexml_layers[37].compute_node[0][1], compute_graph.flexml_layers[37].compute_node[0][2], compute_graph.flexml_layers[37].compute_node[0][3], compute_graph.flexml_layers[37].compute_node[1][0], compute_graph.flexml_layers[37].compute_node[1][1], compute_graph.flexml_layers[37].compute_node[1][2], compute_graph.flexml_layers[37].compute_node[1][3], compute_graph.flexml_layers[37].compute_node[2][0], compute_graph.flexml_layers[37].compute_node[2][1], compute_graph.flexml_layers[37].compute_node[2][2], compute_graph.flexml_layers[37].compute_node[2][3], compute_graph.flexml_layers[37].compute_node[3][0], compute_graph.flexml_layers[37].compute_node[3][1], compute_graph.flexml_layers[37].compute_node[3][2], compute_graph.flexml_layers[37].compute_node[3][3], {compute_graph.l2_47.out[0], compute_graph.l2_47.out[1], compute_graph.l2_47.out[2], compute_graph.l2_47.out[3], compute_graph.Layer_48_l2_wts.out[0], compute_graph.Layer_48_l2_wts.out[1], compute_graph.Layer_48_l2_wts.out[2], compute_graph.Layer_48_l2_wts.out[3], compute_graph.l2_48.in[0], compute_graph.l2_48.in[1], compute_graph.l2_48.in[2], compute_graph.l2_48.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(923): mode(0), layer(55): {compute_graph.l2_48.out[2], compute_graph.l2_48.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(922): mode(0), layer(55): {compute_graph.l2_48.out[0], compute_graph.l2_48.out[1], compute_graph.l2l3_48_spill.in[0], compute_graph.l2l3_48_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1062): mode(0), layer(56): {compute_graph.l2l3_48_spill.out[0], compute_graph.templated_graph_49.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1250): mode(0), layer(56): {compute_graph.templated_graph_49.trans_mt_ifm.out[0], compute_graph.templated_graph_49.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1251): mode(0), layer(56): {compute_graph.templated_graph_49.trans_mt_ofm.out[0], compute_graph.l2l3_49_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1063): mode(0), layer(57): {compute_graph.l2l3_49_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1366), layer(57): compute_graph.flexml_layers[38].compute_node[0][0], compute_graph.flexml_layers[38].compute_node[0][1], compute_graph.flexml_layers[38].compute_node[0][2], compute_graph.flexml_layers[38].compute_node[0][3], compute_graph.flexml_layers[38].compute_node[1][0], compute_graph.flexml_layers[38].compute_node[1][1], compute_graph.flexml_layers[38].compute_node[1][2], compute_graph.flexml_layers[38].compute_node[1][3], compute_graph.flexml_layers[38].compute_node[2][0], compute_graph.flexml_layers[38].compute_node[2][1], compute_graph.flexml_layers[38].compute_node[2][2], compute_graph.flexml_layers[38].compute_node[2][3], compute_graph.flexml_layers[38].compute_node[3][0], compute_graph.flexml_layers[38].compute_node[3][1], compute_graph.flexml_layers[38].compute_node[3][2], compute_graph.flexml_layers[38].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.out[3], compute_graph.l2_50.in[0], compute_graph.l2_50.in[1], compute_graph.l2_50.in[2], compute_graph.l2_50.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(840): mode(3), layer(58): {compute_graph.Layer_51_wts_ddr.out[0], compute_graph.Layer_51_wts_ddr.out[1], compute_graph.Layer_51_l2_wts.in[0], compute_graph.Layer_51_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(840) is pipelined with KernelLayerNode(1366) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1367), layer(58): compute_graph.flexml_layers[39].compute_node[0][0], compute_graph.flexml_layers[39].compute_node[0][1], compute_graph.flexml_layers[39].compute_node[0][2], compute_graph.flexml_layers[39].compute_node[0][3], compute_graph.flexml_layers[39].compute_node[1][0], compute_graph.flexml_layers[39].compute_node[1][1], compute_graph.flexml_layers[39].compute_node[1][2], compute_graph.flexml_layers[39].compute_node[1][3], compute_graph.flexml_layers[39].compute_node[2][0], compute_graph.flexml_layers[39].compute_node[2][1], compute_graph.flexml_layers[39].compute_node[2][2], compute_graph.flexml_layers[39].compute_node[2][3], compute_graph.flexml_layers[39].compute_node[3][0], compute_graph.flexml_layers[39].compute_node[3][1], compute_graph.flexml_layers[39].compute_node[3][2], compute_graph.flexml_layers[39].compute_node[3][3], {compute_graph.l2_50.out[0], compute_graph.l2_50.out[1], compute_graph.l2_50.out[2], compute_graph.l2_50.out[3], compute_graph.Layer_51_l2_wts.out[0], compute_graph.Layer_51_l2_wts.out[1], compute_graph.Layer_51_l2_wts.out[2], compute_graph.Layer_51_l2_wts.out[3], compute_graph.l2_51.in[0], compute_graph.l2_51.in[1], compute_graph.l2_51.in[2], compute_graph.l2_51.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(841): mode(3), layer(59): {compute_graph.Layer_52_wts_ddr.out[0], compute_graph.Layer_52_wts_ddr.out[1], compute_graph.Layer_52_l2_wts.in[0], compute_graph.Layer_52_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(841) is pipelined with KernelLayerNode(1367) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1368), layer(59): compute_graph.flexml_layers[40].compute_node[0][0], compute_graph.flexml_layers[40].compute_node[0][1], compute_graph.flexml_layers[40].compute_node[0][2], compute_graph.flexml_layers[40].compute_node[0][3], compute_graph.flexml_layers[40].compute_node[1][0], compute_graph.flexml_layers[40].compute_node[1][1], compute_graph.flexml_layers[40].compute_node[1][2], compute_graph.flexml_layers[40].compute_node[1][3], compute_graph.flexml_layers[40].compute_node[2][0], compute_graph.flexml_layers[40].compute_node[2][1], compute_graph.flexml_layers[40].compute_node[2][2], compute_graph.flexml_layers[40].compute_node[2][3], compute_graph.flexml_layers[40].compute_node[3][0], compute_graph.flexml_layers[40].compute_node[3][1], compute_graph.flexml_layers[40].compute_node[3][2], compute_graph.flexml_layers[40].compute_node[3][3], {compute_graph.l2_51.out[0], compute_graph.l2_51.out[1], compute_graph.l2_51.out[2], compute_graph.l2_51.out[3], compute_graph.Layer_52_l2_wts.out[0], compute_graph.Layer_52_l2_wts.out[1], compute_graph.Layer_52_l2_wts.out[2], compute_graph.Layer_52_l2_wts.out[3], compute_graph.l2_52.in[0], compute_graph.l2_52.in[1], compute_graph.l2_52.in[2], compute_graph.l2_52.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1369), layer(60): compute_graph.flexml_layers[41].compute_node[0][0], compute_graph.flexml_layers[41].compute_node[0][1], compute_graph.flexml_layers[41].compute_node[0][2], compute_graph.flexml_layers[41].compute_node[0][3], compute_graph.flexml_layers[41].compute_node[1][0], compute_graph.flexml_layers[41].compute_node[1][1], compute_graph.flexml_layers[41].compute_node[1][2], compute_graph.flexml_layers[41].compute_node[1][3], compute_graph.flexml_layers[41].compute_node[2][0], compute_graph.flexml_layers[41].compute_node[2][1], compute_graph.flexml_layers[41].compute_node[2][2], compute_graph.flexml_layers[41].compute_node[2][3], compute_graph.flexml_layers[41].compute_node[3][0], compute_graph.flexml_layers[41].compute_node[3][1], compute_graph.flexml_layers[41].compute_node[3][2], compute_graph.flexml_layers[41].compute_node[3][3], {compute_graph.l2_52.out[0], compute_graph.l2_52.out[1], compute_graph.l2_52.out[2], compute_graph.l2_52.out[3], compute_graph.l2_53.in[0], compute_graph.l2_53.in[1], compute_graph.l2_53.in[2], compute_graph.l2_53.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1370), layer(61): compute_graph.flexml_layers[42].compute_node[0][0], compute_graph.flexml_layers[42].compute_node[0][1], compute_graph.flexml_layers[42].compute_node[0][2], compute_graph.flexml_layers[42].compute_node[0][3], compute_graph.flexml_layers[42].compute_node[1][0], compute_graph.flexml_layers[42].compute_node[1][1], compute_graph.flexml_layers[42].compute_node[1][2], compute_graph.flexml_layers[42].compute_node[1][3], compute_graph.flexml_layers[42].compute_node[2][0], compute_graph.flexml_layers[42].compute_node[2][1], compute_graph.flexml_layers[42].compute_node[2][2], compute_graph.flexml_layers[42].compute_node[2][3], compute_graph.flexml_layers[42].compute_node[3][0], compute_graph.flexml_layers[42].compute_node[3][1], compute_graph.flexml_layers[42].compute_node[3][2], compute_graph.flexml_layers[42].compute_node[3][3], {compute_graph.l2_53.out[0], compute_graph.l2_53.out[1], compute_graph.l2_53.out[2], compute_graph.l2_53.out[3], compute_graph.l2_54.in[0], compute_graph.l2_54.in[1], compute_graph.l2_54.in[2], compute_graph.l2_54.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1371), layer(62): compute_graph.flexml_layers[43].compute_node[0][0], compute_graph.flexml_layers[43].compute_node[0][1], compute_graph.flexml_layers[43].compute_node[0][2], compute_graph.flexml_layers[43].compute_node[0][3], compute_graph.flexml_layers[43].compute_node[1][0], compute_graph.flexml_layers[43].compute_node[1][1], compute_graph.flexml_layers[43].compute_node[1][2], compute_graph.flexml_layers[43].compute_node[1][3], compute_graph.flexml_layers[43].compute_node[2][0], compute_graph.flexml_layers[43].compute_node[2][1], compute_graph.flexml_layers[43].compute_node[2][2], compute_graph.flexml_layers[43].compute_node[2][3], compute_graph.flexml_layers[43].compute_node[3][0], compute_graph.flexml_layers[43].compute_node[3][1], compute_graph.flexml_layers[43].compute_node[3][2], compute_graph.flexml_layers[43].compute_node[3][3], {compute_graph.l2_54.out[0], compute_graph.l2_54.out[1], compute_graph.l2_54.out[2], compute_graph.l2_54.out[3], compute_graph.l2_55.in[0], compute_graph.l2_55.in[1], compute_graph.l2_55.in[2], compute_graph.l2_55.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(924): mode(0), layer(62): {compute_graph.l2_55.out[1], compute_graph.l2l3_55_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1064): mode(0), layer(63): {compute_graph.l2l3_55_spill.out[0], compute_graph.templated_graph_56.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1252): mode(0), layer(63): {compute_graph.templated_graph_56.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_56_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1065): mode(0), layer(64): {compute_graph.l2l3_scratch_0_56_spill.out[0], compute_graph.templated_graph_56.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1253): mode(0), layer(64): {compute_graph.templated_graph_56.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_56_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1066): mode(0), layer(65): {compute_graph.l2l3_scratch_1_56_spill.out[0], compute_graph.templated_graph_56.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1254): mode(0), layer(65): {compute_graph.templated_graph_56.g2.ifm_mem.out[0], compute_graph.l2l3_56_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1061): mode(0), layer(66): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1067): mode(0), layer(66): {compute_graph.l2l3_56_spill.out[0], compute_graph.l2l3_56_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1372), layer(66): compute_graph.flexml_layers[44].compute_node[0][0], compute_graph.flexml_layers[44].compute_node[0][1], compute_graph.flexml_layers[44].compute_node[0][2], compute_graph.flexml_layers[44].compute_node[0][3], compute_graph.flexml_layers[44].compute_node[1][0], compute_graph.flexml_layers[44].compute_node[1][1], compute_graph.flexml_layers[44].compute_node[1][2], compute_graph.flexml_layers[44].compute_node[1][3], compute_graph.flexml_layers[44].compute_node[2][0], compute_graph.flexml_layers[44].compute_node[2][1], compute_graph.flexml_layers[44].compute_node[2][2], compute_graph.flexml_layers[44].compute_node[2][3], compute_graph.flexml_layers[44].compute_node[3][0], compute_graph.flexml_layers[44].compute_node[3][1], compute_graph.flexml_layers[44].compute_node[3][2], compute_graph.flexml_layers[44].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.out[3], compute_graph.l2_57.in[0], compute_graph.l2_57.in[1], compute_graph.l2_57.in[2], compute_graph.l2_57.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(842): mode(3), layer(67): {compute_graph.Layer_58_wts_ddr.out[0], compute_graph.Layer_58_wts_ddr.out[1], compute_graph.Layer_58_l2_wts.in[0], compute_graph.Layer_58_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(842) is pipelined with KernelLayerNode(1372) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646 overlaps with memory space used by compute_graph.l2_58. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1060): mode(0), layer(67): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1373), layer(67): compute_graph.flexml_layers[45].compute_node[0][0], compute_graph.flexml_layers[45].compute_node[0][1], compute_graph.flexml_layers[45].compute_node[0][2], compute_graph.flexml_layers[45].compute_node[0][3], compute_graph.flexml_layers[45].compute_node[1][0], compute_graph.flexml_layers[45].compute_node[1][1], compute_graph.flexml_layers[45].compute_node[1][2], compute_graph.flexml_layers[45].compute_node[1][3], compute_graph.flexml_layers[45].compute_node[2][0], compute_graph.flexml_layers[45].compute_node[2][1], compute_graph.flexml_layers[45].compute_node[2][2], compute_graph.flexml_layers[45].compute_node[2][3], compute_graph.flexml_layers[45].compute_node[3][0], compute_graph.flexml_layers[45].compute_node[3][1], compute_graph.flexml_layers[45].compute_node[3][2], compute_graph.flexml_layers[45].compute_node[3][3], {compute_graph.l2_57.out[0], compute_graph.l2_57.out[1], compute_graph.l2_57.out[2], compute_graph.l2_57.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.out[3], compute_graph.Layer_58_l2_wts.out[0], compute_graph.Layer_58_l2_wts.out[1], compute_graph.Layer_58_l2_wts.out[2], compute_graph.Layer_58_l2_wts.out[3], compute_graph.l2_58.in[0], compute_graph.l2_58.in[1], compute_graph.l2_58.in[2], compute_graph.l2_58.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(843): mode(3), layer(68): {compute_graph.Layer_59_wts_ddr.out[0], compute_graph.Layer_59_wts_ddr.out[1], compute_graph.Layer_59_l2_wts.in[0], compute_graph.Layer_59_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(843) is pipelined with KernelLayerNode(1373) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2_58.out[0], compute_graph.l2_58.out[1], compute_graph.l2_58.out[2], compute_graph.l2_58.out[3], compute_graph.l2_59.in[0], compute_graph.l2_59.in[1], compute_graph.l2_59.in[2], compute_graph.l2_59.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1374), layer(68): compute_graph.flexml_layers[46].compute_node[0][0], compute_graph.flexml_layers[46].compute_node[0][1], compute_graph.flexml_layers[46].compute_node[0][2], compute_graph.flexml_layers[46].compute_node[0][3], compute_graph.flexml_layers[46].compute_node[1][0], compute_graph.flexml_layers[46].compute_node[1][1], compute_graph.flexml_layers[46].compute_node[1][2], compute_graph.flexml_layers[46].compute_node[1][3], compute_graph.flexml_layers[46].compute_node[2][0], compute_graph.flexml_layers[46].compute_node[2][1], compute_graph.flexml_layers[46].compute_node[2][2], compute_graph.flexml_layers[46].compute_node[2][3], compute_graph.flexml_layers[46].compute_node[3][0], compute_graph.flexml_layers[46].compute_node[3][1], compute_graph.flexml_layers[46].compute_node[3][2], compute_graph.flexml_layers[46].compute_node[3][3], {compute_graph.l2_58.out[0], compute_graph.l2_58.out[1], compute_graph.l2_58.out[2], compute_graph.l2_58.out[3], compute_graph.Layer_59_l2_wts.out[0], compute_graph.Layer_59_l2_wts.out[1], compute_graph.Layer_59_l2_wts.out[2], compute_graph.Layer_59_l2_wts.out[3], compute_graph.l2_59.in[0], compute_graph.l2_59.in[1], compute_graph.l2_59.in[2], compute_graph.l2_59.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(925): mode(3), layer(68): {compute_graph.l2_58.out[4], compute_graph.l2_58.out[5], compute_graph.spill_L3_Concat_Buffer_layer_236.in[2], compute_graph.spill_L3_Concat_Buffer_layer_236.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(925) is pipelined with KernelLayerNode(1374) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1375), layer(69): compute_graph.flexml_layers[47].compute_node[0][0], compute_graph.flexml_layers[47].compute_node[0][1], compute_graph.flexml_layers[47].compute_node[0][2], compute_graph.flexml_layers[47].compute_node[0][3], compute_graph.flexml_layers[47].compute_node[1][0], compute_graph.flexml_layers[47].compute_node[1][1], compute_graph.flexml_layers[47].compute_node[1][2], compute_graph.flexml_layers[47].compute_node[1][3], compute_graph.flexml_layers[47].compute_node[2][0], compute_graph.flexml_layers[47].compute_node[2][1], compute_graph.flexml_layers[47].compute_node[2][2], compute_graph.flexml_layers[47].compute_node[2][3], compute_graph.flexml_layers[47].compute_node[3][0], compute_graph.flexml_layers[47].compute_node[3][1], compute_graph.flexml_layers[47].compute_node[3][2], compute_graph.flexml_layers[47].compute_node[3][3], {compute_graph.l2_59.out[0], compute_graph.l2_59.out[1], compute_graph.l2_59.out[2], compute_graph.l2_59.out[3], compute_graph.l2_60.in[0], compute_graph.l2_60.in[1], compute_graph.l2_60.in[2], compute_graph.l2_60.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(926): mode(3), layer(69): {compute_graph.l2_59.out[4], compute_graph.l2_59.out[5], compute_graph.l2l3_59_spill.in[0], compute_graph.l2l3_59_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(926) is pipelined with KernelLayerNode(1375) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1376), layer(70): compute_graph.flexml_layers[48].compute_node[0][0], compute_graph.flexml_layers[48].compute_node[0][1], compute_graph.flexml_layers[48].compute_node[0][2], compute_graph.flexml_layers[48].compute_node[0][3], compute_graph.flexml_layers[48].compute_node[1][0], compute_graph.flexml_layers[48].compute_node[1][1], compute_graph.flexml_layers[48].compute_node[1][2], compute_graph.flexml_layers[48].compute_node[1][3], compute_graph.flexml_layers[48].compute_node[2][0], compute_graph.flexml_layers[48].compute_node[2][1], compute_graph.flexml_layers[48].compute_node[2][2], compute_graph.flexml_layers[48].compute_node[2][3], compute_graph.flexml_layers[48].compute_node[3][0], compute_graph.flexml_layers[48].compute_node[3][1], compute_graph.flexml_layers[48].compute_node[3][2], compute_graph.flexml_layers[48].compute_node[3][3], {compute_graph.l2_60.out[0], compute_graph.l2_60.out[1], compute_graph.l2_60.out[2], compute_graph.l2_60.out[3], compute_graph.l2_61.in[0], compute_graph.l2_61.in[1], compute_graph.l2_61.in[2], compute_graph.l2_61.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1377), layer(71): compute_graph.flexml_layers[49].compute_node[0][0], compute_graph.flexml_layers[49].compute_node[0][1], compute_graph.flexml_layers[49].compute_node[0][2], compute_graph.flexml_layers[49].compute_node[0][3], compute_graph.flexml_layers[49].compute_node[1][0], compute_graph.flexml_layers[49].compute_node[1][1], compute_graph.flexml_layers[49].compute_node[1][2], compute_graph.flexml_layers[49].compute_node[1][3], compute_graph.flexml_layers[49].compute_node[2][0], compute_graph.flexml_layers[49].compute_node[2][1], compute_graph.flexml_layers[49].compute_node[2][2], compute_graph.flexml_layers[49].compute_node[2][3], compute_graph.flexml_layers[49].compute_node[3][0], compute_graph.flexml_layers[49].compute_node[3][1], compute_graph.flexml_layers[49].compute_node[3][2], compute_graph.flexml_layers[49].compute_node[3][3], {compute_graph.l2_61.out[0], compute_graph.l2_61.out[1], compute_graph.l2_61.out[2], compute_graph.l2_61.out[3], compute_graph.l2_62.in[0], compute_graph.l2_62.in[1], compute_graph.l2_62.in[2], compute_graph.l2_62.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(927): mode(0), layer(71): {compute_graph.l2_62.out[0], compute_graph.l2_62.out[1], compute_graph.l2l3_62_spill.in[0], compute_graph.l2l3_62_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 928 will not be pipelined because it's in the same layer 72 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1585), layer(72): {compute_graph.l2l3_59_spill.out[0], compute_graph.l2l3_62_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1586), layer(72): {compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1587), layer(72): {compute_graph.l2l3_59_spill.out[1], compute_graph.l2l3_62_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1588), layer(72): {compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[3], compute_graph.l2_63.in[0], compute_graph.l2_63.in[1], compute_graph.l2_63.in[2], compute_graph.l2_63.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1378), layer(72): compute_graph.flexml_layers[50].compute_node[0][0], compute_graph.flexml_layers[50].compute_node[0][1], compute_graph.flexml_layers[50].compute_node[0][2], compute_graph.flexml_layers[50].compute_node[0][3], compute_graph.flexml_layers[50].compute_node[1][0], compute_graph.flexml_layers[50].compute_node[1][1], compute_graph.flexml_layers[50].compute_node[1][2], compute_graph.flexml_layers[50].compute_node[1][3], compute_graph.flexml_layers[50].compute_node[2][0], compute_graph.flexml_layers[50].compute_node[2][1], compute_graph.flexml_layers[50].compute_node[2][2], compute_graph.flexml_layers[50].compute_node[2][3], compute_graph.flexml_layers[50].compute_node[3][0], compute_graph.flexml_layers[50].compute_node[3][1], compute_graph.flexml_layers[50].compute_node[3][2], compute_graph.flexml_layers[50].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.out[3], compute_graph.l2_63.in[0], compute_graph.l2_63.in[1], compute_graph.l2_63.in[2], compute_graph.l2_63.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(928): mode(3), layer(72): {compute_graph.l2_63.out[0], compute_graph.l2_63.out[1], compute_graph.l2l3_63_spill.in[0], compute_graph.l2l3_63_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1070): mode(0), layer(73): {compute_graph.l2l3_63_spill.out[0], compute_graph.templated_graph_64.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1255): mode(0), layer(73): {compute_graph.templated_graph_64.ifm.out[0], compute_graph.templated_graph_64.ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1256): mode(0), layer(73): {compute_graph.templated_graph_64.ofm.out[0], compute_graph.l2l3_scratch_0_64_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1071): mode(0), layer(74): {compute_graph.l2l3_scratch_0_64_spill.out[0], compute_graph.templated_graph_64.ifm2.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1257): mode(0), layer(74): {compute_graph.templated_graph_64.ifm2.out[0], compute_graph.l2l3_64_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1072): mode(0), layer(75): {compute_graph.l2l3_64_spill.out[0], compute_graph.l2l3_64_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(844): mode(4), layer(75): {compute_graph.Layer_65_wts_ddr.out[0], compute_graph.Layer_65_wts_ddr.out[1], compute_graph.Layer_65_l2_wts.in[0], compute_graph.Layer_65_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(844) is pipelined with BufferToBufferNode(1072) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1379), layer(75): compute_graph.flexml_layers[51].compute_node[0][0], compute_graph.flexml_layers[51].compute_node[0][1], compute_graph.flexml_layers[51].compute_node[0][2], compute_graph.flexml_layers[51].compute_node[0][3], compute_graph.flexml_layers[51].compute_node[1][0], compute_graph.flexml_layers[51].compute_node[1][1], compute_graph.flexml_layers[51].compute_node[1][2], compute_graph.flexml_layers[51].compute_node[1][3], compute_graph.flexml_layers[51].compute_node[2][0], compute_graph.flexml_layers[51].compute_node[2][1], compute_graph.flexml_layers[51].compute_node[2][2], compute_graph.flexml_layers[51].compute_node[2][3], compute_graph.flexml_layers[51].compute_node[3][0], compute_graph.flexml_layers[51].compute_node[3][1], compute_graph.flexml_layers[51].compute_node[3][2], compute_graph.flexml_layers[51].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.out[3], compute_graph.Layer_65_l2_wts.out[0], compute_graph.Layer_65_l2_wts.out[1], compute_graph.Layer_65_l2_wts.out[2], compute_graph.Layer_65_l2_wts.out[3], compute_graph.l2_65.in[0], compute_graph.l2_65.in[1], compute_graph.l2_65.in[2], compute_graph.l2_65.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1380), layer(76): compute_graph.flexml_layers[52].compute_node[0][0], compute_graph.flexml_layers[52].compute_node[0][1], compute_graph.flexml_layers[52].compute_node[0][2], compute_graph.flexml_layers[52].compute_node[0][3], compute_graph.flexml_layers[52].compute_node[1][0], compute_graph.flexml_layers[52].compute_node[1][1], compute_graph.flexml_layers[52].compute_node[1][2], compute_graph.flexml_layers[52].compute_node[1][3], compute_graph.flexml_layers[52].compute_node[2][0], compute_graph.flexml_layers[52].compute_node[2][1], compute_graph.flexml_layers[52].compute_node[2][2], compute_graph.flexml_layers[52].compute_node[2][3], compute_graph.flexml_layers[52].compute_node[3][0], compute_graph.flexml_layers[52].compute_node[3][1], compute_graph.flexml_layers[52].compute_node[3][2], compute_graph.flexml_layers[52].compute_node[3][3], {compute_graph.l2_65.out[0], compute_graph.l2_65.out[1], compute_graph.l2_65.out[2], compute_graph.l2_65.out[3], compute_graph.l2_66.in[0], compute_graph.l2_66.in[1], compute_graph.l2_66.in[2], compute_graph.l2_66.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1381), layer(77): compute_graph.flexml_layers[53].compute_node[0][0], compute_graph.flexml_layers[53].compute_node[0][1], compute_graph.flexml_layers[53].compute_node[0][2], compute_graph.flexml_layers[53].compute_node[0][3], compute_graph.flexml_layers[53].compute_node[1][0], compute_graph.flexml_layers[53].compute_node[1][1], compute_graph.flexml_layers[53].compute_node[1][2], compute_graph.flexml_layers[53].compute_node[1][3], compute_graph.flexml_layers[53].compute_node[2][0], compute_graph.flexml_layers[53].compute_node[2][1], compute_graph.flexml_layers[53].compute_node[2][2], compute_graph.flexml_layers[53].compute_node[2][3], compute_graph.flexml_layers[53].compute_node[3][0], compute_graph.flexml_layers[53].compute_node[3][1], compute_graph.flexml_layers[53].compute_node[3][2], compute_graph.flexml_layers[53].compute_node[3][3], {compute_graph.l2_66.out[0], compute_graph.l2_66.out[1], compute_graph.l2_66.out[2], compute_graph.l2_66.out[3], compute_graph.l2_67.in[0], compute_graph.l2_67.in[1], compute_graph.l2_67.in[2], compute_graph.l2_67.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1382), layer(78): compute_graph.flexml_layers[54].compute_node[0][0], compute_graph.flexml_layers[54].compute_node[0][1], compute_graph.flexml_layers[54].compute_node[0][2], compute_graph.flexml_layers[54].compute_node[0][3], compute_graph.flexml_layers[54].compute_node[1][0], compute_graph.flexml_layers[54].compute_node[1][1], compute_graph.flexml_layers[54].compute_node[1][2], compute_graph.flexml_layers[54].compute_node[1][3], compute_graph.flexml_layers[54].compute_node[2][0], compute_graph.flexml_layers[54].compute_node[2][1], compute_graph.flexml_layers[54].compute_node[2][2], compute_graph.flexml_layers[54].compute_node[2][3], compute_graph.flexml_layers[54].compute_node[3][0], compute_graph.flexml_layers[54].compute_node[3][1], compute_graph.flexml_layers[54].compute_node[3][2], compute_graph.flexml_layers[54].compute_node[3][3], {compute_graph.l2_67.out[0], compute_graph.l2_67.out[1], compute_graph.l2_67.out[2], compute_graph.l2_67.out[3], compute_graph.l2_68.in[0], compute_graph.l2_68.in[1], compute_graph.l2_68.in[2], compute_graph.l2_68.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1383), layer(79): compute_graph.flexml_layers[55].compute_node[0][0], compute_graph.flexml_layers[55].compute_node[0][1], compute_graph.flexml_layers[55].compute_node[0][2], compute_graph.flexml_layers[55].compute_node[0][3], compute_graph.flexml_layers[55].compute_node[1][0], compute_graph.flexml_layers[55].compute_node[1][1], compute_graph.flexml_layers[55].compute_node[1][2], compute_graph.flexml_layers[55].compute_node[1][3], compute_graph.flexml_layers[55].compute_node[2][0], compute_graph.flexml_layers[55].compute_node[2][1], compute_graph.flexml_layers[55].compute_node[2][2], compute_graph.flexml_layers[55].compute_node[2][3], compute_graph.flexml_layers[55].compute_node[3][0], compute_graph.flexml_layers[55].compute_node[3][1], compute_graph.flexml_layers[55].compute_node[3][2], compute_graph.flexml_layers[55].compute_node[3][3], {compute_graph.l2_65.out[4], compute_graph.l2_65.out[5], compute_graph.l2_65.out[6], compute_graph.l2_65.out[7], compute_graph.l2_68.out[0], compute_graph.l2_68.out[1], compute_graph.l2_68.out[2], compute_graph.l2_68.out[3], compute_graph.l2_69.in[0], compute_graph.l2_69.in[1], compute_graph.l2_69.in[2], compute_graph.l2_69.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(845): mode(3), layer(80): {compute_graph.Layer_70_wts_ddr.out[0], compute_graph.Layer_70_wts_ddr.out[1], compute_graph.Layer_70_l2_wts.in[0], compute_graph.Layer_70_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(845) is pipelined with KernelLayerNode(1383) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1384), layer(80): compute_graph.flexml_layers[56].compute_node[0][0], compute_graph.flexml_layers[56].compute_node[0][1], compute_graph.flexml_layers[56].compute_node[0][2], compute_graph.flexml_layers[56].compute_node[0][3], compute_graph.flexml_layers[56].compute_node[1][0], compute_graph.flexml_layers[56].compute_node[1][1], compute_graph.flexml_layers[56].compute_node[1][2], compute_graph.flexml_layers[56].compute_node[1][3], compute_graph.flexml_layers[56].compute_node[2][0], compute_graph.flexml_layers[56].compute_node[2][1], compute_graph.flexml_layers[56].compute_node[2][2], compute_graph.flexml_layers[56].compute_node[2][3], compute_graph.flexml_layers[56].compute_node[3][0], compute_graph.flexml_layers[56].compute_node[3][1], compute_graph.flexml_layers[56].compute_node[3][2], compute_graph.flexml_layers[56].compute_node[3][3], {compute_graph.l2_69.out[0], compute_graph.l2_69.out[1], compute_graph.l2_69.out[2], compute_graph.l2_69.out[3], compute_graph.Layer_70_l2_wts.out[0], compute_graph.Layer_70_l2_wts.out[1], compute_graph.Layer_70_l2_wts.out[2], compute_graph.Layer_70_l2_wts.out[3], compute_graph.l2_70.in[0], compute_graph.l2_70.in[1], compute_graph.l2_70.in[2], compute_graph.l2_70.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(846): mode(3), layer(81): {compute_graph.Layer_71_wts_ddr.out[0], compute_graph.Layer_71_wts_ddr.out[1], compute_graph.Layer_71_l2_wts.in[0], compute_graph.Layer_71_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(846) is pipelined with KernelLayerNode(1384) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1385), layer(81): compute_graph.flexml_layers[57].compute_node[0][0], compute_graph.flexml_layers[57].compute_node[0][1], compute_graph.flexml_layers[57].compute_node[0][2], compute_graph.flexml_layers[57].compute_node[0][3], compute_graph.flexml_layers[57].compute_node[1][0], compute_graph.flexml_layers[57].compute_node[1][1], compute_graph.flexml_layers[57].compute_node[1][2], compute_graph.flexml_layers[57].compute_node[1][3], compute_graph.flexml_layers[57].compute_node[2][0], compute_graph.flexml_layers[57].compute_node[2][1], compute_graph.flexml_layers[57].compute_node[2][2], compute_graph.flexml_layers[57].compute_node[2][3], compute_graph.flexml_layers[57].compute_node[3][0], compute_graph.flexml_layers[57].compute_node[3][1], compute_graph.flexml_layers[57].compute_node[3][2], compute_graph.flexml_layers[57].compute_node[3][3], {compute_graph.l2_70.out[0], compute_graph.l2_70.out[1], compute_graph.l2_70.out[2], compute_graph.l2_70.out[3], compute_graph.Layer_71_l2_wts.out[0], compute_graph.Layer_71_l2_wts.out[1], compute_graph.Layer_71_l2_wts.out[2], compute_graph.Layer_71_l2_wts.out[3], compute_graph.l2_71.in[0], compute_graph.l2_71.in[1], compute_graph.l2_71.in[2], compute_graph.l2_71.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1386), layer(82): compute_graph.flexml_layers[58].compute_node[0][0], compute_graph.flexml_layers[58].compute_node[0][1], compute_graph.flexml_layers[58].compute_node[0][2], compute_graph.flexml_layers[58].compute_node[0][3], compute_graph.flexml_layers[58].compute_node[1][0], compute_graph.flexml_layers[58].compute_node[1][1], compute_graph.flexml_layers[58].compute_node[1][2], compute_graph.flexml_layers[58].compute_node[1][3], compute_graph.flexml_layers[58].compute_node[2][0], compute_graph.flexml_layers[58].compute_node[2][1], compute_graph.flexml_layers[58].compute_node[2][2], compute_graph.flexml_layers[58].compute_node[2][3], compute_graph.flexml_layers[58].compute_node[3][0], compute_graph.flexml_layers[58].compute_node[3][1], compute_graph.flexml_layers[58].compute_node[3][2], compute_graph.flexml_layers[58].compute_node[3][3], {compute_graph.l2_71.out[0], compute_graph.l2_71.out[1], compute_graph.l2_71.out[2], compute_graph.l2_71.out[3], compute_graph.l2_72.in[0], compute_graph.l2_72.in[1], compute_graph.l2_72.in[2], compute_graph.l2_72.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1387), layer(83): compute_graph.flexml_layers[59].compute_node[0][0], compute_graph.flexml_layers[59].compute_node[0][1], compute_graph.flexml_layers[59].compute_node[0][2], compute_graph.flexml_layers[59].compute_node[0][3], compute_graph.flexml_layers[59].compute_node[1][0], compute_graph.flexml_layers[59].compute_node[1][1], compute_graph.flexml_layers[59].compute_node[1][2], compute_graph.flexml_layers[59].compute_node[1][3], compute_graph.flexml_layers[59].compute_node[2][0], compute_graph.flexml_layers[59].compute_node[2][1], compute_graph.flexml_layers[59].compute_node[2][2], compute_graph.flexml_layers[59].compute_node[2][3], compute_graph.flexml_layers[59].compute_node[3][0], compute_graph.flexml_layers[59].compute_node[3][1], compute_graph.flexml_layers[59].compute_node[3][2], compute_graph.flexml_layers[59].compute_node[3][3], {compute_graph.l2_72.out[0], compute_graph.l2_72.out[1], compute_graph.l2_72.out[2], compute_graph.l2_72.out[3], compute_graph.l2_73.in[0], compute_graph.l2_73.in[1], compute_graph.l2_73.in[2], compute_graph.l2_73.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1388), layer(84): compute_graph.flexml_layers[60].compute_node[0][0], compute_graph.flexml_layers[60].compute_node[0][1], compute_graph.flexml_layers[60].compute_node[0][2], compute_graph.flexml_layers[60].compute_node[0][3], compute_graph.flexml_layers[60].compute_node[1][0], compute_graph.flexml_layers[60].compute_node[1][1], compute_graph.flexml_layers[60].compute_node[1][2], compute_graph.flexml_layers[60].compute_node[1][3], compute_graph.flexml_layers[60].compute_node[2][0], compute_graph.flexml_layers[60].compute_node[2][1], compute_graph.flexml_layers[60].compute_node[2][2], compute_graph.flexml_layers[60].compute_node[2][3], compute_graph.flexml_layers[60].compute_node[3][0], compute_graph.flexml_layers[60].compute_node[3][1], compute_graph.flexml_layers[60].compute_node[3][2], compute_graph.flexml_layers[60].compute_node[3][3], {compute_graph.l2_73.out[0], compute_graph.l2_73.out[1], compute_graph.l2_73.out[2], compute_graph.l2_73.out[3], compute_graph.l2_74.in[0], compute_graph.l2_74.in[1], compute_graph.l2_74.in[2], compute_graph.l2_74.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1389), layer(85): compute_graph.flexml_layers[61].compute_node[0][0], compute_graph.flexml_layers[61].compute_node[0][1], compute_graph.flexml_layers[61].compute_node[0][2], compute_graph.flexml_layers[61].compute_node[0][3], compute_graph.flexml_layers[61].compute_node[1][0], compute_graph.flexml_layers[61].compute_node[1][1], compute_graph.flexml_layers[61].compute_node[1][2], compute_graph.flexml_layers[61].compute_node[1][3], compute_graph.flexml_layers[61].compute_node[2][0], compute_graph.flexml_layers[61].compute_node[2][1], compute_graph.flexml_layers[61].compute_node[2][2], compute_graph.flexml_layers[61].compute_node[2][3], compute_graph.flexml_layers[61].compute_node[3][0], compute_graph.flexml_layers[61].compute_node[3][1], compute_graph.flexml_layers[61].compute_node[3][2], compute_graph.flexml_layers[61].compute_node[3][3], {compute_graph.l2_71.out[4], compute_graph.l2_71.out[5], compute_graph.l2_71.out[6], compute_graph.l2_71.out[7], compute_graph.l2_74.out[0], compute_graph.l2_74.out[1], compute_graph.l2_74.out[2], compute_graph.l2_74.out[3], compute_graph.l2_75.in[0], compute_graph.l2_75.in[1], compute_graph.l2_75.in[2], compute_graph.l2_75.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(847): mode(3), layer(86): {compute_graph.Layer_76_wts_ddr.out[0], compute_graph.Layer_76_wts_ddr.out[1], compute_graph.Layer_76_l2_wts.in[0], compute_graph.Layer_76_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(847) is pipelined with KernelLayerNode(1389) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1390), layer(86): compute_graph.flexml_layers[62].compute_node[0][0], compute_graph.flexml_layers[62].compute_node[0][1], compute_graph.flexml_layers[62].compute_node[0][2], compute_graph.flexml_layers[62].compute_node[0][3], compute_graph.flexml_layers[62].compute_node[1][0], compute_graph.flexml_layers[62].compute_node[1][1], compute_graph.flexml_layers[62].compute_node[1][2], compute_graph.flexml_layers[62].compute_node[1][3], compute_graph.flexml_layers[62].compute_node[2][0], compute_graph.flexml_layers[62].compute_node[2][1], compute_graph.flexml_layers[62].compute_node[2][2], compute_graph.flexml_layers[62].compute_node[2][3], compute_graph.flexml_layers[62].compute_node[3][0], compute_graph.flexml_layers[62].compute_node[3][1], compute_graph.flexml_layers[62].compute_node[3][2], compute_graph.flexml_layers[62].compute_node[3][3], {compute_graph.l2_75.out[0], compute_graph.l2_75.out[1], compute_graph.l2_75.out[2], compute_graph.l2_75.out[3], compute_graph.Layer_76_l2_wts.out[0], compute_graph.Layer_76_l2_wts.out[1], compute_graph.Layer_76_l2_wts.out[2], compute_graph.Layer_76_l2_wts.out[3], compute_graph.l2_76.in[0], compute_graph.l2_76.in[1], compute_graph.l2_76.in[2], compute_graph.l2_76.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1391), layer(87): compute_graph.flexml_layers[63].compute_node[0][0], compute_graph.flexml_layers[63].compute_node[0][1], compute_graph.flexml_layers[63].compute_node[0][2], compute_graph.flexml_layers[63].compute_node[0][3], compute_graph.flexml_layers[63].compute_node[1][0], compute_graph.flexml_layers[63].compute_node[1][1], compute_graph.flexml_layers[63].compute_node[1][2], compute_graph.flexml_layers[63].compute_node[1][3], compute_graph.flexml_layers[63].compute_node[2][0], compute_graph.flexml_layers[63].compute_node[2][1], compute_graph.flexml_layers[63].compute_node[2][2], compute_graph.flexml_layers[63].compute_node[2][3], compute_graph.flexml_layers[63].compute_node[3][0], compute_graph.flexml_layers[63].compute_node[3][1], compute_graph.flexml_layers[63].compute_node[3][2], compute_graph.flexml_layers[63].compute_node[3][3], {compute_graph.l2_76.out[0], compute_graph.l2_76.out[1], compute_graph.l2_76.out[2], compute_graph.l2_76.out[3], compute_graph.l2_77.in[0], compute_graph.l2_77.in[1], compute_graph.l2_77.in[2], compute_graph.l2_77.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1392), layer(88): compute_graph.flexml_layers[64].compute_node[0][0], compute_graph.flexml_layers[64].compute_node[0][1], compute_graph.flexml_layers[64].compute_node[0][2], compute_graph.flexml_layers[64].compute_node[0][3], compute_graph.flexml_layers[64].compute_node[1][0], compute_graph.flexml_layers[64].compute_node[1][1], compute_graph.flexml_layers[64].compute_node[1][2], compute_graph.flexml_layers[64].compute_node[1][3], compute_graph.flexml_layers[64].compute_node[2][0], compute_graph.flexml_layers[64].compute_node[2][1], compute_graph.flexml_layers[64].compute_node[2][2], compute_graph.flexml_layers[64].compute_node[2][3], compute_graph.flexml_layers[64].compute_node[3][0], compute_graph.flexml_layers[64].compute_node[3][1], compute_graph.flexml_layers[64].compute_node[3][2], compute_graph.flexml_layers[64].compute_node[3][3], {compute_graph.l2_77.out[0], compute_graph.l2_77.out[1], compute_graph.l2_77.out[2], compute_graph.l2_77.out[3], compute_graph.l2_78.in[0], compute_graph.l2_78.in[1], compute_graph.l2_78.in[2], compute_graph.l2_78.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1393), layer(89): compute_graph.flexml_layers[65].compute_node[0][0], compute_graph.flexml_layers[65].compute_node[0][1], compute_graph.flexml_layers[65].compute_node[0][2], compute_graph.flexml_layers[65].compute_node[0][3], compute_graph.flexml_layers[65].compute_node[1][0], compute_graph.flexml_layers[65].compute_node[1][1], compute_graph.flexml_layers[65].compute_node[1][2], compute_graph.flexml_layers[65].compute_node[1][3], compute_graph.flexml_layers[65].compute_node[2][0], compute_graph.flexml_layers[65].compute_node[2][1], compute_graph.flexml_layers[65].compute_node[2][2], compute_graph.flexml_layers[65].compute_node[2][3], compute_graph.flexml_layers[65].compute_node[3][0], compute_graph.flexml_layers[65].compute_node[3][1], compute_graph.flexml_layers[65].compute_node[3][2], compute_graph.flexml_layers[65].compute_node[3][3], {compute_graph.l2_78.out[0], compute_graph.l2_78.out[1], compute_graph.l2_78.out[2], compute_graph.l2_78.out[3], compute_graph.l2_79.in[0], compute_graph.l2_79.in[1], compute_graph.l2_79.in[2], compute_graph.l2_79.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1394), layer(90): compute_graph.flexml_layers[66].compute_node[0][0], compute_graph.flexml_layers[66].compute_node[0][1], compute_graph.flexml_layers[66].compute_node[0][2], compute_graph.flexml_layers[66].compute_node[0][3], compute_graph.flexml_layers[66].compute_node[1][0], compute_graph.flexml_layers[66].compute_node[1][1], compute_graph.flexml_layers[66].compute_node[1][2], compute_graph.flexml_layers[66].compute_node[1][3], compute_graph.flexml_layers[66].compute_node[2][0], compute_graph.flexml_layers[66].compute_node[2][1], compute_graph.flexml_layers[66].compute_node[2][2], compute_graph.flexml_layers[66].compute_node[2][3], compute_graph.flexml_layers[66].compute_node[3][0], compute_graph.flexml_layers[66].compute_node[3][1], compute_graph.flexml_layers[66].compute_node[3][2], compute_graph.flexml_layers[66].compute_node[3][3], {compute_graph.l2_76.out[4], compute_graph.l2_76.out[5], compute_graph.l2_76.out[6], compute_graph.l2_76.out[7], compute_graph.l2_79.out[0], compute_graph.l2_79.out[1], compute_graph.l2_79.out[2], compute_graph.l2_79.out[3], compute_graph.l2_80.in[0], compute_graph.l2_80.in[1], compute_graph.l2_80.in[2], compute_graph.l2_80.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(848): mode(3), layer(91): {compute_graph.Layer_81_wts_ddr.out[0], compute_graph.Layer_81_wts_ddr.out[1], compute_graph.Layer_81_l2_wts.in[0], compute_graph.Layer_81_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(848) is pipelined with KernelLayerNode(1394) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.l2_70 overlaps with memory space used by compute_graph.l2_81. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For KernelLayerNode(1395), layer(91): compute_graph.flexml_layers[67].compute_node[0][0], compute_graph.flexml_layers[67].compute_node[0][1], compute_graph.flexml_layers[67].compute_node[0][2], compute_graph.flexml_layers[67].compute_node[0][3], compute_graph.flexml_layers[67].compute_node[1][0], compute_graph.flexml_layers[67].compute_node[1][1], compute_graph.flexml_layers[67].compute_node[1][2], compute_graph.flexml_layers[67].compute_node[1][3], compute_graph.flexml_layers[67].compute_node[2][0], compute_graph.flexml_layers[67].compute_node[2][1], compute_graph.flexml_layers[67].compute_node[2][2], compute_graph.flexml_layers[67].compute_node[2][3], compute_graph.flexml_layers[67].compute_node[3][0], compute_graph.flexml_layers[67].compute_node[3][1], compute_graph.flexml_layers[67].compute_node[3][2], compute_graph.flexml_layers[67].compute_node[3][3], {compute_graph.l2_70.out[4], compute_graph.l2_70.out[5], compute_graph.l2_70.out[6], compute_graph.l2_70.out[7], compute_graph.l2_80.out[0], compute_graph.l2_80.out[1], compute_graph.l2_80.out[2], compute_graph.l2_80.out[3], compute_graph.Layer_81_l2_wts.out[0], compute_graph.Layer_81_l2_wts.out[1], compute_graph.Layer_81_l2_wts.out[2], compute_graph.Layer_81_l2_wts.out[3], compute_graph.l2_81.in[0], compute_graph.l2_81.in[1], compute_graph.l2_81.in[2], compute_graph.l2_81.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(849): mode(3), layer(92): {compute_graph.Layer_82_wts_ddr.out[0], compute_graph.Layer_82_wts_ddr.out[1], compute_graph.Layer_82_l2_wts.in[0], compute_graph.Layer_82_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(849) is pipelined with KernelLayerNode(1395) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1396), layer(92): compute_graph.flexml_layers[68].compute_node[0][0], compute_graph.flexml_layers[68].compute_node[0][1], compute_graph.flexml_layers[68].compute_node[0][2], compute_graph.flexml_layers[68].compute_node[0][3], compute_graph.flexml_layers[68].compute_node[1][0], compute_graph.flexml_layers[68].compute_node[1][1], compute_graph.flexml_layers[68].compute_node[1][2], compute_graph.flexml_layers[68].compute_node[1][3], compute_graph.flexml_layers[68].compute_node[2][0], compute_graph.flexml_layers[68].compute_node[2][1], compute_graph.flexml_layers[68].compute_node[2][2], compute_graph.flexml_layers[68].compute_node[2][3], compute_graph.flexml_layers[68].compute_node[3][0], compute_graph.flexml_layers[68].compute_node[3][1], compute_graph.flexml_layers[68].compute_node[3][2], compute_graph.flexml_layers[68].compute_node[3][3], {compute_graph.l2_81.out[0], compute_graph.l2_81.out[1], compute_graph.l2_81.out[2], compute_graph.l2_81.out[3], compute_graph.Layer_82_l2_wts.out[0], compute_graph.Layer_82_l2_wts.out[1], compute_graph.Layer_82_l2_wts.out[2], compute_graph.Layer_82_l2_wts.out[3], compute_graph.l2_82.in[0], compute_graph.l2_82.in[1], compute_graph.l2_82.in[2], compute_graph.l2_82.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1397), layer(93): compute_graph.flexml_layers[69].compute_node[0][0], compute_graph.flexml_layers[69].compute_node[0][1], compute_graph.flexml_layers[69].compute_node[0][2], compute_graph.flexml_layers[69].compute_node[0][3], compute_graph.flexml_layers[69].compute_node[1][0], compute_graph.flexml_layers[69].compute_node[1][1], compute_graph.flexml_layers[69].compute_node[1][2], compute_graph.flexml_layers[69].compute_node[1][3], compute_graph.flexml_layers[69].compute_node[2][0], compute_graph.flexml_layers[69].compute_node[2][1], compute_graph.flexml_layers[69].compute_node[2][2], compute_graph.flexml_layers[69].compute_node[2][3], compute_graph.flexml_layers[69].compute_node[3][0], compute_graph.flexml_layers[69].compute_node[3][1], compute_graph.flexml_layers[69].compute_node[3][2], compute_graph.flexml_layers[69].compute_node[3][3], {compute_graph.l2_82.out[0], compute_graph.l2_82.out[1], compute_graph.l2_82.out[2], compute_graph.l2_82.out[3], compute_graph.l2_83.in[0], compute_graph.l2_83.in[1], compute_graph.l2_83.in[2], compute_graph.l2_83.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1398), layer(94): compute_graph.flexml_layers[70].compute_node[0][0], compute_graph.flexml_layers[70].compute_node[0][1], compute_graph.flexml_layers[70].compute_node[0][2], compute_graph.flexml_layers[70].compute_node[0][3], compute_graph.flexml_layers[70].compute_node[1][0], compute_graph.flexml_layers[70].compute_node[1][1], compute_graph.flexml_layers[70].compute_node[1][2], compute_graph.flexml_layers[70].compute_node[1][3], compute_graph.flexml_layers[70].compute_node[2][0], compute_graph.flexml_layers[70].compute_node[2][1], compute_graph.flexml_layers[70].compute_node[2][2], compute_graph.flexml_layers[70].compute_node[2][3], compute_graph.flexml_layers[70].compute_node[3][0], compute_graph.flexml_layers[70].compute_node[3][1], compute_graph.flexml_layers[70].compute_node[3][2], compute_graph.flexml_layers[70].compute_node[3][3], {compute_graph.l2_83.out[0], compute_graph.l2_83.out[1], compute_graph.l2_83.out[2], compute_graph.l2_83.out[3], compute_graph.l2_84.in[0], compute_graph.l2_84.in[1], compute_graph.l2_84.in[2], compute_graph.l2_84.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1399), layer(95): compute_graph.flexml_layers[71].compute_node[0][0], compute_graph.flexml_layers[71].compute_node[0][1], compute_graph.flexml_layers[71].compute_node[0][2], compute_graph.flexml_layers[71].compute_node[0][3], compute_graph.flexml_layers[71].compute_node[1][0], compute_graph.flexml_layers[71].compute_node[1][1], compute_graph.flexml_layers[71].compute_node[1][2], compute_graph.flexml_layers[71].compute_node[1][3], compute_graph.flexml_layers[71].compute_node[2][0], compute_graph.flexml_layers[71].compute_node[2][1], compute_graph.flexml_layers[71].compute_node[2][2], compute_graph.flexml_layers[71].compute_node[2][3], compute_graph.flexml_layers[71].compute_node[3][0], compute_graph.flexml_layers[71].compute_node[3][1], compute_graph.flexml_layers[71].compute_node[3][2], compute_graph.flexml_layers[71].compute_node[3][3], {compute_graph.l2_84.out[0], compute_graph.l2_84.out[1], compute_graph.l2_84.out[2], compute_graph.l2_84.out[3], compute_graph.l2_85.in[0], compute_graph.l2_85.in[1], compute_graph.l2_85.in[2], compute_graph.l2_85.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1400), layer(96): compute_graph.flexml_layers[72].compute_node[0][0], compute_graph.flexml_layers[72].compute_node[0][1], compute_graph.flexml_layers[72].compute_node[0][2], compute_graph.flexml_layers[72].compute_node[0][3], compute_graph.flexml_layers[72].compute_node[1][0], compute_graph.flexml_layers[72].compute_node[1][1], compute_graph.flexml_layers[72].compute_node[1][2], compute_graph.flexml_layers[72].compute_node[1][3], compute_graph.flexml_layers[72].compute_node[2][0], compute_graph.flexml_layers[72].compute_node[2][1], compute_graph.flexml_layers[72].compute_node[2][2], compute_graph.flexml_layers[72].compute_node[2][3], compute_graph.flexml_layers[72].compute_node[3][0], compute_graph.flexml_layers[72].compute_node[3][1], compute_graph.flexml_layers[72].compute_node[3][2], compute_graph.flexml_layers[72].compute_node[3][3], {compute_graph.l2_82.out[4], compute_graph.l2_82.out[5], compute_graph.l2_82.out[6], compute_graph.l2_82.out[7], compute_graph.l2_85.out[0], compute_graph.l2_85.out[1], compute_graph.l2_85.out[2], compute_graph.l2_85.out[3], compute_graph.l2_86.in[0], compute_graph.l2_86.in[1], compute_graph.l2_86.in[2], compute_graph.l2_86.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(850): mode(3), layer(97): {compute_graph.Layer_87_wts_ddr.out[0], compute_graph.Layer_87_wts_ddr.out[1], compute_graph.Layer_87_l2_wts.in[0], compute_graph.Layer_87_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(850) is pipelined with KernelLayerNode(1400) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1401), layer(97): compute_graph.flexml_layers[73].compute_node[0][0], compute_graph.flexml_layers[73].compute_node[0][1], compute_graph.flexml_layers[73].compute_node[0][2], compute_graph.flexml_layers[73].compute_node[0][3], compute_graph.flexml_layers[73].compute_node[1][0], compute_graph.flexml_layers[73].compute_node[1][1], compute_graph.flexml_layers[73].compute_node[1][2], compute_graph.flexml_layers[73].compute_node[1][3], compute_graph.flexml_layers[73].compute_node[2][0], compute_graph.flexml_layers[73].compute_node[2][1], compute_graph.flexml_layers[73].compute_node[2][2], compute_graph.flexml_layers[73].compute_node[2][3], compute_graph.flexml_layers[73].compute_node[3][0], compute_graph.flexml_layers[73].compute_node[3][1], compute_graph.flexml_layers[73].compute_node[3][2], compute_graph.flexml_layers[73].compute_node[3][3], {compute_graph.l2_86.out[0], compute_graph.l2_86.out[1], compute_graph.l2_86.out[2], compute_graph.l2_86.out[3], compute_graph.Layer_87_l2_wts.out[0], compute_graph.Layer_87_l2_wts.out[1], compute_graph.Layer_87_l2_wts.out[2], compute_graph.Layer_87_l2_wts.out[3], compute_graph.l2_87.in[0], compute_graph.l2_87.in[1], compute_graph.l2_87.in[2], compute_graph.l2_87.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1402), layer(98): compute_graph.flexml_layers[74].compute_node[0][0], compute_graph.flexml_layers[74].compute_node[0][1], compute_graph.flexml_layers[74].compute_node[0][2], compute_graph.flexml_layers[74].compute_node[0][3], compute_graph.flexml_layers[74].compute_node[1][0], compute_graph.flexml_layers[74].compute_node[1][1], compute_graph.flexml_layers[74].compute_node[1][2], compute_graph.flexml_layers[74].compute_node[1][3], compute_graph.flexml_layers[74].compute_node[2][0], compute_graph.flexml_layers[74].compute_node[2][1], compute_graph.flexml_layers[74].compute_node[2][2], compute_graph.flexml_layers[74].compute_node[2][3], compute_graph.flexml_layers[74].compute_node[3][0], compute_graph.flexml_layers[74].compute_node[3][1], compute_graph.flexml_layers[74].compute_node[3][2], compute_graph.flexml_layers[74].compute_node[3][3], {compute_graph.l2_87.out[0], compute_graph.l2_87.out[1], compute_graph.l2_87.out[2], compute_graph.l2_87.out[3], compute_graph.l2_88.in[0], compute_graph.l2_88.in[1], compute_graph.l2_88.in[2], compute_graph.l2_88.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1403), layer(99): compute_graph.flexml_layers[75].compute_node[0][0], compute_graph.flexml_layers[75].compute_node[0][1], compute_graph.flexml_layers[75].compute_node[0][2], compute_graph.flexml_layers[75].compute_node[0][3], compute_graph.flexml_layers[75].compute_node[1][0], compute_graph.flexml_layers[75].compute_node[1][1], compute_graph.flexml_layers[75].compute_node[1][2], compute_graph.flexml_layers[75].compute_node[1][3], compute_graph.flexml_layers[75].compute_node[2][0], compute_graph.flexml_layers[75].compute_node[2][1], compute_graph.flexml_layers[75].compute_node[2][2], compute_graph.flexml_layers[75].compute_node[2][3], compute_graph.flexml_layers[75].compute_node[3][0], compute_graph.flexml_layers[75].compute_node[3][1], compute_graph.flexml_layers[75].compute_node[3][2], compute_graph.flexml_layers[75].compute_node[3][3], {compute_graph.l2_88.out[0], compute_graph.l2_88.out[1], compute_graph.l2_88.out[2], compute_graph.l2_88.out[3], compute_graph.l2_89.in[0], compute_graph.l2_89.in[1], compute_graph.l2_89.in[2], compute_graph.l2_89.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1404), layer(100): compute_graph.flexml_layers[76].compute_node[0][0], compute_graph.flexml_layers[76].compute_node[0][1], compute_graph.flexml_layers[76].compute_node[0][2], compute_graph.flexml_layers[76].compute_node[0][3], compute_graph.flexml_layers[76].compute_node[1][0], compute_graph.flexml_layers[76].compute_node[1][1], compute_graph.flexml_layers[76].compute_node[1][2], compute_graph.flexml_layers[76].compute_node[1][3], compute_graph.flexml_layers[76].compute_node[2][0], compute_graph.flexml_layers[76].compute_node[2][1], compute_graph.flexml_layers[76].compute_node[2][2], compute_graph.flexml_layers[76].compute_node[2][3], compute_graph.flexml_layers[76].compute_node[3][0], compute_graph.flexml_layers[76].compute_node[3][1], compute_graph.flexml_layers[76].compute_node[3][2], compute_graph.flexml_layers[76].compute_node[3][3], {compute_graph.l2_89.out[0], compute_graph.l2_89.out[1], compute_graph.l2_89.out[2], compute_graph.l2_89.out[3], compute_graph.l2_90.in[0], compute_graph.l2_90.in[1], compute_graph.l2_90.in[2], compute_graph.l2_90.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1405), layer(101): compute_graph.flexml_layers[77].compute_node[0][0], compute_graph.flexml_layers[77].compute_node[0][1], compute_graph.flexml_layers[77].compute_node[0][2], compute_graph.flexml_layers[77].compute_node[0][3], compute_graph.flexml_layers[77].compute_node[1][0], compute_graph.flexml_layers[77].compute_node[1][1], compute_graph.flexml_layers[77].compute_node[1][2], compute_graph.flexml_layers[77].compute_node[1][3], compute_graph.flexml_layers[77].compute_node[2][0], compute_graph.flexml_layers[77].compute_node[2][1], compute_graph.flexml_layers[77].compute_node[2][2], compute_graph.flexml_layers[77].compute_node[2][3], compute_graph.flexml_layers[77].compute_node[3][0], compute_graph.flexml_layers[77].compute_node[3][1], compute_graph.flexml_layers[77].compute_node[3][2], compute_graph.flexml_layers[77].compute_node[3][3], {compute_graph.l2_87.out[4], compute_graph.l2_87.out[5], compute_graph.l2_87.out[6], compute_graph.l2_87.out[7], compute_graph.l2_90.out[0], compute_graph.l2_90.out[1], compute_graph.l2_90.out[2], compute_graph.l2_90.out[3], compute_graph.l2_91.in[0], compute_graph.l2_91.in[1], compute_graph.l2_91.in[2], compute_graph.l2_91.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(851): mode(3), layer(102): {compute_graph.Layer_92_wts_ddr.out[0], compute_graph.Layer_92_wts_ddr.out[1], compute_graph.Layer_92_l2_wts.in[0], compute_graph.Layer_92_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(851) is pipelined with KernelLayerNode(1405) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.l2_81 overlaps with memory space used by compute_graph.l2_92. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For KernelLayerNode(1406), layer(102): compute_graph.flexml_layers[78].compute_node[0][0], compute_graph.flexml_layers[78].compute_node[0][1], compute_graph.flexml_layers[78].compute_node[0][2], compute_graph.flexml_layers[78].compute_node[0][3], compute_graph.flexml_layers[78].compute_node[1][0], compute_graph.flexml_layers[78].compute_node[1][1], compute_graph.flexml_layers[78].compute_node[1][2], compute_graph.flexml_layers[78].compute_node[1][3], compute_graph.flexml_layers[78].compute_node[2][0], compute_graph.flexml_layers[78].compute_node[2][1], compute_graph.flexml_layers[78].compute_node[2][2], compute_graph.flexml_layers[78].compute_node[2][3], compute_graph.flexml_layers[78].compute_node[3][0], compute_graph.flexml_layers[78].compute_node[3][1], compute_graph.flexml_layers[78].compute_node[3][2], compute_graph.flexml_layers[78].compute_node[3][3], {compute_graph.l2_81.out[4], compute_graph.l2_81.out[5], compute_graph.l2_81.out[6], compute_graph.l2_81.out[7], compute_graph.l2_91.out[0], compute_graph.l2_91.out[1], compute_graph.l2_91.out[2], compute_graph.l2_91.out[3], compute_graph.Layer_92_l2_wts.out[0], compute_graph.Layer_92_l2_wts.out[1], compute_graph.Layer_92_l2_wts.out[2], compute_graph.Layer_92_l2_wts.out[3], compute_graph.l2_92.in[0], compute_graph.l2_92.in[1], compute_graph.l2_92.in[2], compute_graph.l2_92.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(852): mode(3), layer(103): {compute_graph.Layer_93_wts_ddr.out[0], compute_graph.Layer_93_wts_ddr.out[1], compute_graph.Layer_93_l2_wts.in[0], compute_graph.Layer_93_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(852) is pipelined with KernelLayerNode(1406) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1407), layer(103): compute_graph.flexml_layers[79].compute_node[0][0], compute_graph.flexml_layers[79].compute_node[0][1], compute_graph.flexml_layers[79].compute_node[0][2], compute_graph.flexml_layers[79].compute_node[0][3], compute_graph.flexml_layers[79].compute_node[1][0], compute_graph.flexml_layers[79].compute_node[1][1], compute_graph.flexml_layers[79].compute_node[1][2], compute_graph.flexml_layers[79].compute_node[1][3], compute_graph.flexml_layers[79].compute_node[2][0], compute_graph.flexml_layers[79].compute_node[2][1], compute_graph.flexml_layers[79].compute_node[2][2], compute_graph.flexml_layers[79].compute_node[2][3], compute_graph.flexml_layers[79].compute_node[3][0], compute_graph.flexml_layers[79].compute_node[3][1], compute_graph.flexml_layers[79].compute_node[3][2], compute_graph.flexml_layers[79].compute_node[3][3], {compute_graph.l2_92.out[0], compute_graph.l2_92.out[1], compute_graph.l2_92.out[2], compute_graph.l2_92.out[3], compute_graph.Layer_93_l2_wts.out[0], compute_graph.Layer_93_l2_wts.out[1], compute_graph.Layer_93_l2_wts.out[2], compute_graph.Layer_93_l2_wts.out[3], compute_graph.l2_93.in[0], compute_graph.l2_93.in[1], compute_graph.l2_93.in[2], compute_graph.l2_93.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1408), layer(104): compute_graph.flexml_layers[80].compute_node[0][0], compute_graph.flexml_layers[80].compute_node[0][1], compute_graph.flexml_layers[80].compute_node[0][2], compute_graph.flexml_layers[80].compute_node[0][3], compute_graph.flexml_layers[80].compute_node[1][0], compute_graph.flexml_layers[80].compute_node[1][1], compute_graph.flexml_layers[80].compute_node[1][2], compute_graph.flexml_layers[80].compute_node[1][3], compute_graph.flexml_layers[80].compute_node[2][0], compute_graph.flexml_layers[80].compute_node[2][1], compute_graph.flexml_layers[80].compute_node[2][2], compute_graph.flexml_layers[80].compute_node[2][3], compute_graph.flexml_layers[80].compute_node[3][0], compute_graph.flexml_layers[80].compute_node[3][1], compute_graph.flexml_layers[80].compute_node[3][2], compute_graph.flexml_layers[80].compute_node[3][3], {compute_graph.l2_93.out[0], compute_graph.l2_93.out[1], compute_graph.l2_93.out[2], compute_graph.l2_93.out[3], compute_graph.l2_94.in[0], compute_graph.l2_94.in[1], compute_graph.l2_94.in[2], compute_graph.l2_94.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1409), layer(105): compute_graph.flexml_layers[81].compute_node[0][0], compute_graph.flexml_layers[81].compute_node[0][1], compute_graph.flexml_layers[81].compute_node[0][2], compute_graph.flexml_layers[81].compute_node[0][3], compute_graph.flexml_layers[81].compute_node[1][0], compute_graph.flexml_layers[81].compute_node[1][1], compute_graph.flexml_layers[81].compute_node[1][2], compute_graph.flexml_layers[81].compute_node[1][3], compute_graph.flexml_layers[81].compute_node[2][0], compute_graph.flexml_layers[81].compute_node[2][1], compute_graph.flexml_layers[81].compute_node[2][2], compute_graph.flexml_layers[81].compute_node[2][3], compute_graph.flexml_layers[81].compute_node[3][0], compute_graph.flexml_layers[81].compute_node[3][1], compute_graph.flexml_layers[81].compute_node[3][2], compute_graph.flexml_layers[81].compute_node[3][3], {compute_graph.l2_94.out[0], compute_graph.l2_94.out[1], compute_graph.l2_94.out[2], compute_graph.l2_94.out[3], compute_graph.l2_95.in[0], compute_graph.l2_95.in[1], compute_graph.l2_95.in[2], compute_graph.l2_95.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1410), layer(106): compute_graph.flexml_layers[82].compute_node[0][0], compute_graph.flexml_layers[82].compute_node[0][1], compute_graph.flexml_layers[82].compute_node[0][2], compute_graph.flexml_layers[82].compute_node[0][3], compute_graph.flexml_layers[82].compute_node[1][0], compute_graph.flexml_layers[82].compute_node[1][1], compute_graph.flexml_layers[82].compute_node[1][2], compute_graph.flexml_layers[82].compute_node[1][3], compute_graph.flexml_layers[82].compute_node[2][0], compute_graph.flexml_layers[82].compute_node[2][1], compute_graph.flexml_layers[82].compute_node[2][2], compute_graph.flexml_layers[82].compute_node[2][3], compute_graph.flexml_layers[82].compute_node[3][0], compute_graph.flexml_layers[82].compute_node[3][1], compute_graph.flexml_layers[82].compute_node[3][2], compute_graph.flexml_layers[82].compute_node[3][3], {compute_graph.l2_95.out[0], compute_graph.l2_95.out[1], compute_graph.l2_95.out[2], compute_graph.l2_95.out[3], compute_graph.l2_96.in[0], compute_graph.l2_96.in[1], compute_graph.l2_96.in[2], compute_graph.l2_96.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1411), layer(107): compute_graph.flexml_layers[83].compute_node[0][0], compute_graph.flexml_layers[83].compute_node[0][1], compute_graph.flexml_layers[83].compute_node[0][2], compute_graph.flexml_layers[83].compute_node[0][3], compute_graph.flexml_layers[83].compute_node[1][0], compute_graph.flexml_layers[83].compute_node[1][1], compute_graph.flexml_layers[83].compute_node[1][2], compute_graph.flexml_layers[83].compute_node[1][3], compute_graph.flexml_layers[83].compute_node[2][0], compute_graph.flexml_layers[83].compute_node[2][1], compute_graph.flexml_layers[83].compute_node[2][2], compute_graph.flexml_layers[83].compute_node[2][3], compute_graph.flexml_layers[83].compute_node[3][0], compute_graph.flexml_layers[83].compute_node[3][1], compute_graph.flexml_layers[83].compute_node[3][2], compute_graph.flexml_layers[83].compute_node[3][3], {compute_graph.l2_93.out[4], compute_graph.l2_93.out[5], compute_graph.l2_93.out[6], compute_graph.l2_93.out[7], compute_graph.l2_96.out[0], compute_graph.l2_96.out[1], compute_graph.l2_96.out[2], compute_graph.l2_96.out[3], compute_graph.l2_97.in[0], compute_graph.l2_97.in[1], compute_graph.l2_97.in[2], compute_graph.l2_97.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(853): mode(3), layer(108): {compute_graph.Layer_98_wts_ddr.out[0], compute_graph.Layer_98_wts_ddr.out[1], compute_graph.Layer_98_l2_wts.in[0], compute_graph.Layer_98_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(853) is pipelined with KernelLayerNode(1411) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1412), layer(108): compute_graph.flexml_layers[84].compute_node[0][0], compute_graph.flexml_layers[84].compute_node[0][1], compute_graph.flexml_layers[84].compute_node[0][2], compute_graph.flexml_layers[84].compute_node[0][3], compute_graph.flexml_layers[84].compute_node[1][0], compute_graph.flexml_layers[84].compute_node[1][1], compute_graph.flexml_layers[84].compute_node[1][2], compute_graph.flexml_layers[84].compute_node[1][3], compute_graph.flexml_layers[84].compute_node[2][0], compute_graph.flexml_layers[84].compute_node[2][1], compute_graph.flexml_layers[84].compute_node[2][2], compute_graph.flexml_layers[84].compute_node[2][3], compute_graph.flexml_layers[84].compute_node[3][0], compute_graph.flexml_layers[84].compute_node[3][1], compute_graph.flexml_layers[84].compute_node[3][2], compute_graph.flexml_layers[84].compute_node[3][3], {compute_graph.l2_97.out[0], compute_graph.l2_97.out[1], compute_graph.l2_97.out[2], compute_graph.l2_97.out[3], compute_graph.Layer_98_l2_wts.out[0], compute_graph.Layer_98_l2_wts.out[1], compute_graph.Layer_98_l2_wts.out[2], compute_graph.Layer_98_l2_wts.out[3], compute_graph.l2_98.in[0], compute_graph.l2_98.in[1], compute_graph.l2_98.in[2], compute_graph.l2_98.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1413), layer(109): compute_graph.flexml_layers[85].compute_node[0][0], compute_graph.flexml_layers[85].compute_node[0][1], compute_graph.flexml_layers[85].compute_node[0][2], compute_graph.flexml_layers[85].compute_node[0][3], compute_graph.flexml_layers[85].compute_node[1][0], compute_graph.flexml_layers[85].compute_node[1][1], compute_graph.flexml_layers[85].compute_node[1][2], compute_graph.flexml_layers[85].compute_node[1][3], compute_graph.flexml_layers[85].compute_node[2][0], compute_graph.flexml_layers[85].compute_node[2][1], compute_graph.flexml_layers[85].compute_node[2][2], compute_graph.flexml_layers[85].compute_node[2][3], compute_graph.flexml_layers[85].compute_node[3][0], compute_graph.flexml_layers[85].compute_node[3][1], compute_graph.flexml_layers[85].compute_node[3][2], compute_graph.flexml_layers[85].compute_node[3][3], {compute_graph.l2_98.out[0], compute_graph.l2_98.out[1], compute_graph.l2_98.out[2], compute_graph.l2_98.out[3], compute_graph.l2_99.in[0], compute_graph.l2_99.in[1], compute_graph.l2_99.in[2], compute_graph.l2_99.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1414), layer(110): compute_graph.flexml_layers[86].compute_node[0][0], compute_graph.flexml_layers[86].compute_node[0][1], compute_graph.flexml_layers[86].compute_node[0][2], compute_graph.flexml_layers[86].compute_node[0][3], compute_graph.flexml_layers[86].compute_node[1][0], compute_graph.flexml_layers[86].compute_node[1][1], compute_graph.flexml_layers[86].compute_node[1][2], compute_graph.flexml_layers[86].compute_node[1][3], compute_graph.flexml_layers[86].compute_node[2][0], compute_graph.flexml_layers[86].compute_node[2][1], compute_graph.flexml_layers[86].compute_node[2][2], compute_graph.flexml_layers[86].compute_node[2][3], compute_graph.flexml_layers[86].compute_node[3][0], compute_graph.flexml_layers[86].compute_node[3][1], compute_graph.flexml_layers[86].compute_node[3][2], compute_graph.flexml_layers[86].compute_node[3][3], {compute_graph.l2_99.out[0], compute_graph.l2_99.out[1], compute_graph.l2_99.out[2], compute_graph.l2_99.out[3], compute_graph.l2_100.in[0], compute_graph.l2_100.in[1], compute_graph.l2_100.in[2], compute_graph.l2_100.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1415), layer(111): compute_graph.flexml_layers[87].compute_node[0][0], compute_graph.flexml_layers[87].compute_node[0][1], compute_graph.flexml_layers[87].compute_node[0][2], compute_graph.flexml_layers[87].compute_node[0][3], compute_graph.flexml_layers[87].compute_node[1][0], compute_graph.flexml_layers[87].compute_node[1][1], compute_graph.flexml_layers[87].compute_node[1][2], compute_graph.flexml_layers[87].compute_node[1][3], compute_graph.flexml_layers[87].compute_node[2][0], compute_graph.flexml_layers[87].compute_node[2][1], compute_graph.flexml_layers[87].compute_node[2][2], compute_graph.flexml_layers[87].compute_node[2][3], compute_graph.flexml_layers[87].compute_node[3][0], compute_graph.flexml_layers[87].compute_node[3][1], compute_graph.flexml_layers[87].compute_node[3][2], compute_graph.flexml_layers[87].compute_node[3][3], {compute_graph.l2_100.out[0], compute_graph.l2_100.out[1], compute_graph.l2_100.out[2], compute_graph.l2_100.out[3], compute_graph.l2_101.in[0], compute_graph.l2_101.in[1], compute_graph.l2_101.in[2], compute_graph.l2_101.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1416), layer(112): compute_graph.flexml_layers[88].compute_node[0][0], compute_graph.flexml_layers[88].compute_node[0][1], compute_graph.flexml_layers[88].compute_node[0][2], compute_graph.flexml_layers[88].compute_node[0][3], compute_graph.flexml_layers[88].compute_node[1][0], compute_graph.flexml_layers[88].compute_node[1][1], compute_graph.flexml_layers[88].compute_node[1][2], compute_graph.flexml_layers[88].compute_node[1][3], compute_graph.flexml_layers[88].compute_node[2][0], compute_graph.flexml_layers[88].compute_node[2][1], compute_graph.flexml_layers[88].compute_node[2][2], compute_graph.flexml_layers[88].compute_node[2][3], compute_graph.flexml_layers[88].compute_node[3][0], compute_graph.flexml_layers[88].compute_node[3][1], compute_graph.flexml_layers[88].compute_node[3][2], compute_graph.flexml_layers[88].compute_node[3][3], {compute_graph.l2_98.out[4], compute_graph.l2_98.out[5], compute_graph.l2_98.out[6], compute_graph.l2_98.out[7], compute_graph.l2_101.out[0], compute_graph.l2_101.out[1], compute_graph.l2_101.out[2], compute_graph.l2_101.out[3], compute_graph.l2_102.in[0], compute_graph.l2_102.in[1], compute_graph.l2_102.in[2], compute_graph.l2_102.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(854): mode(3), layer(113): {compute_graph.Layer_103_wts_ddr.out[0], compute_graph.Layer_103_wts_ddr.out[1], compute_graph.Layer_103_l2_wts.in[0], compute_graph.Layer_103_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(854) is pipelined with KernelLayerNode(1416) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.l2_92 overlaps with memory space used by compute_graph.l2_103. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For KernelLayerNode(1417), layer(113): compute_graph.flexml_layers[89].compute_node[0][0], compute_graph.flexml_layers[89].compute_node[0][1], compute_graph.flexml_layers[89].compute_node[0][2], compute_graph.flexml_layers[89].compute_node[0][3], compute_graph.flexml_layers[89].compute_node[1][0], compute_graph.flexml_layers[89].compute_node[1][1], compute_graph.flexml_layers[89].compute_node[1][2], compute_graph.flexml_layers[89].compute_node[1][3], compute_graph.flexml_layers[89].compute_node[2][0], compute_graph.flexml_layers[89].compute_node[2][1], compute_graph.flexml_layers[89].compute_node[2][2], compute_graph.flexml_layers[89].compute_node[2][3], compute_graph.flexml_layers[89].compute_node[3][0], compute_graph.flexml_layers[89].compute_node[3][1], compute_graph.flexml_layers[89].compute_node[3][2], compute_graph.flexml_layers[89].compute_node[3][3], {compute_graph.l2_92.out[4], compute_graph.l2_92.out[5], compute_graph.l2_92.out[6], compute_graph.l2_92.out[7], compute_graph.l2_102.out[0], compute_graph.l2_102.out[1], compute_graph.l2_102.out[2], compute_graph.l2_102.out[3], compute_graph.Layer_103_l2_wts.out[0], compute_graph.Layer_103_l2_wts.out[1], compute_graph.Layer_103_l2_wts.out[2], compute_graph.Layer_103_l2_wts.out[3], compute_graph.l2_103.in[0], compute_graph.l2_103.in[1], compute_graph.l2_103.in[2], compute_graph.l2_103.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(855): mode(3), layer(114): {compute_graph.Layer_104_wts_ddr.out[0], compute_graph.Layer_104_wts_ddr.out[1], compute_graph.Layer_104_l2_wts.in[0], compute_graph.Layer_104_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(855) is pipelined with KernelLayerNode(1417) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1418), layer(114): compute_graph.flexml_layers[90].compute_node[0][0], compute_graph.flexml_layers[90].compute_node[0][1], compute_graph.flexml_layers[90].compute_node[0][2], compute_graph.flexml_layers[90].compute_node[0][3], compute_graph.flexml_layers[90].compute_node[1][0], compute_graph.flexml_layers[90].compute_node[1][1], compute_graph.flexml_layers[90].compute_node[1][2], compute_graph.flexml_layers[90].compute_node[1][3], compute_graph.flexml_layers[90].compute_node[2][0], compute_graph.flexml_layers[90].compute_node[2][1], compute_graph.flexml_layers[90].compute_node[2][2], compute_graph.flexml_layers[90].compute_node[2][3], compute_graph.flexml_layers[90].compute_node[3][0], compute_graph.flexml_layers[90].compute_node[3][1], compute_graph.flexml_layers[90].compute_node[3][2], compute_graph.flexml_layers[90].compute_node[3][3], {compute_graph.l2_103.out[0], compute_graph.l2_103.out[1], compute_graph.l2_103.out[2], compute_graph.l2_103.out[3], compute_graph.Layer_104_l2_wts.out[0], compute_graph.Layer_104_l2_wts.out[1], compute_graph.Layer_104_l2_wts.out[2], compute_graph.Layer_104_l2_wts.out[3], compute_graph.l2_104.in[0], compute_graph.l2_104.in[1], compute_graph.l2_104.in[2], compute_graph.l2_104.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1419), layer(115): compute_graph.flexml_layers[91].compute_node[0][0], compute_graph.flexml_layers[91].compute_node[0][1], compute_graph.flexml_layers[91].compute_node[0][2], compute_graph.flexml_layers[91].compute_node[0][3], compute_graph.flexml_layers[91].compute_node[1][0], compute_graph.flexml_layers[91].compute_node[1][1], compute_graph.flexml_layers[91].compute_node[1][2], compute_graph.flexml_layers[91].compute_node[1][3], compute_graph.flexml_layers[91].compute_node[2][0], compute_graph.flexml_layers[91].compute_node[2][1], compute_graph.flexml_layers[91].compute_node[2][2], compute_graph.flexml_layers[91].compute_node[2][3], compute_graph.flexml_layers[91].compute_node[3][0], compute_graph.flexml_layers[91].compute_node[3][1], compute_graph.flexml_layers[91].compute_node[3][2], compute_graph.flexml_layers[91].compute_node[3][3], {compute_graph.l2_104.out[0], compute_graph.l2_104.out[1], compute_graph.l2_104.out[2], compute_graph.l2_104.out[3], compute_graph.l2_105.in[0], compute_graph.l2_105.in[1], compute_graph.l2_105.in[2], compute_graph.l2_105.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1420), layer(116): compute_graph.flexml_layers[92].compute_node[0][0], compute_graph.flexml_layers[92].compute_node[0][1], compute_graph.flexml_layers[92].compute_node[0][2], compute_graph.flexml_layers[92].compute_node[0][3], compute_graph.flexml_layers[92].compute_node[1][0], compute_graph.flexml_layers[92].compute_node[1][1], compute_graph.flexml_layers[92].compute_node[1][2], compute_graph.flexml_layers[92].compute_node[1][3], compute_graph.flexml_layers[92].compute_node[2][0], compute_graph.flexml_layers[92].compute_node[2][1], compute_graph.flexml_layers[92].compute_node[2][2], compute_graph.flexml_layers[92].compute_node[2][3], compute_graph.flexml_layers[92].compute_node[3][0], compute_graph.flexml_layers[92].compute_node[3][1], compute_graph.flexml_layers[92].compute_node[3][2], compute_graph.flexml_layers[92].compute_node[3][3], {compute_graph.l2_105.out[0], compute_graph.l2_105.out[1], compute_graph.l2_105.out[2], compute_graph.l2_105.out[3], compute_graph.l2_106.in[0], compute_graph.l2_106.in[1], compute_graph.l2_106.in[2], compute_graph.l2_106.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1421), layer(117): compute_graph.flexml_layers[93].compute_node[0][0], compute_graph.flexml_layers[93].compute_node[0][1], compute_graph.flexml_layers[93].compute_node[0][2], compute_graph.flexml_layers[93].compute_node[0][3], compute_graph.flexml_layers[93].compute_node[1][0], compute_graph.flexml_layers[93].compute_node[1][1], compute_graph.flexml_layers[93].compute_node[1][2], compute_graph.flexml_layers[93].compute_node[1][3], compute_graph.flexml_layers[93].compute_node[2][0], compute_graph.flexml_layers[93].compute_node[2][1], compute_graph.flexml_layers[93].compute_node[2][2], compute_graph.flexml_layers[93].compute_node[2][3], compute_graph.flexml_layers[93].compute_node[3][0], compute_graph.flexml_layers[93].compute_node[3][1], compute_graph.flexml_layers[93].compute_node[3][2], compute_graph.flexml_layers[93].compute_node[3][3], {compute_graph.l2_106.out[0], compute_graph.l2_106.out[1], compute_graph.l2_106.out[2], compute_graph.l2_106.out[3], compute_graph.l2_107.in[0], compute_graph.l2_107.in[1], compute_graph.l2_107.in[2], compute_graph.l2_107.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1422), layer(118): compute_graph.flexml_layers[94].compute_node[0][0], compute_graph.flexml_layers[94].compute_node[0][1], compute_graph.flexml_layers[94].compute_node[0][2], compute_graph.flexml_layers[94].compute_node[0][3], compute_graph.flexml_layers[94].compute_node[1][0], compute_graph.flexml_layers[94].compute_node[1][1], compute_graph.flexml_layers[94].compute_node[1][2], compute_graph.flexml_layers[94].compute_node[1][3], compute_graph.flexml_layers[94].compute_node[2][0], compute_graph.flexml_layers[94].compute_node[2][1], compute_graph.flexml_layers[94].compute_node[2][2], compute_graph.flexml_layers[94].compute_node[2][3], compute_graph.flexml_layers[94].compute_node[3][0], compute_graph.flexml_layers[94].compute_node[3][1], compute_graph.flexml_layers[94].compute_node[3][2], compute_graph.flexml_layers[94].compute_node[3][3], {compute_graph.l2_104.out[4], compute_graph.l2_104.out[5], compute_graph.l2_104.out[6], compute_graph.l2_104.out[7], compute_graph.l2_107.out[0], compute_graph.l2_107.out[1], compute_graph.l2_107.out[2], compute_graph.l2_107.out[3], compute_graph.l2_108.in[0], compute_graph.l2_108.in[1], compute_graph.l2_108.in[2], compute_graph.l2_108.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(856): mode(3), layer(119): {compute_graph.Layer_109_wts_ddr.out[0], compute_graph.Layer_109_wts_ddr.out[1], compute_graph.Layer_109_l2_wts.in[0], compute_graph.Layer_109_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(856) is pipelined with KernelLayerNode(1422) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1423), layer(119): compute_graph.flexml_layers[95].compute_node[0][0], compute_graph.flexml_layers[95].compute_node[0][1], compute_graph.flexml_layers[95].compute_node[0][2], compute_graph.flexml_layers[95].compute_node[0][3], compute_graph.flexml_layers[95].compute_node[1][0], compute_graph.flexml_layers[95].compute_node[1][1], compute_graph.flexml_layers[95].compute_node[1][2], compute_graph.flexml_layers[95].compute_node[1][3], compute_graph.flexml_layers[95].compute_node[2][0], compute_graph.flexml_layers[95].compute_node[2][1], compute_graph.flexml_layers[95].compute_node[2][2], compute_graph.flexml_layers[95].compute_node[2][3], compute_graph.flexml_layers[95].compute_node[3][0], compute_graph.flexml_layers[95].compute_node[3][1], compute_graph.flexml_layers[95].compute_node[3][2], compute_graph.flexml_layers[95].compute_node[3][3], {compute_graph.l2_108.out[0], compute_graph.l2_108.out[1], compute_graph.l2_108.out[2], compute_graph.l2_108.out[3], compute_graph.Layer_109_l2_wts.out[0], compute_graph.Layer_109_l2_wts.out[1], compute_graph.Layer_109_l2_wts.out[2], compute_graph.Layer_109_l2_wts.out[3], compute_graph.l2_109.in[0], compute_graph.l2_109.in[1], compute_graph.l2_109.in[2], compute_graph.l2_109.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1424), layer(120): compute_graph.flexml_layers[96].compute_node[0][0], compute_graph.flexml_layers[96].compute_node[0][1], compute_graph.flexml_layers[96].compute_node[0][2], compute_graph.flexml_layers[96].compute_node[0][3], compute_graph.flexml_layers[96].compute_node[1][0], compute_graph.flexml_layers[96].compute_node[1][1], compute_graph.flexml_layers[96].compute_node[1][2], compute_graph.flexml_layers[96].compute_node[1][3], compute_graph.flexml_layers[96].compute_node[2][0], compute_graph.flexml_layers[96].compute_node[2][1], compute_graph.flexml_layers[96].compute_node[2][2], compute_graph.flexml_layers[96].compute_node[2][3], compute_graph.flexml_layers[96].compute_node[3][0], compute_graph.flexml_layers[96].compute_node[3][1], compute_graph.flexml_layers[96].compute_node[3][2], compute_graph.flexml_layers[96].compute_node[3][3], {compute_graph.l2_109.out[0], compute_graph.l2_109.out[1], compute_graph.l2_109.out[2], compute_graph.l2_109.out[3], compute_graph.l2_110.in[0], compute_graph.l2_110.in[1], compute_graph.l2_110.in[2], compute_graph.l2_110.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1425), layer(121): compute_graph.flexml_layers[97].compute_node[0][0], compute_graph.flexml_layers[97].compute_node[0][1], compute_graph.flexml_layers[97].compute_node[0][2], compute_graph.flexml_layers[97].compute_node[0][3], compute_graph.flexml_layers[97].compute_node[1][0], compute_graph.flexml_layers[97].compute_node[1][1], compute_graph.flexml_layers[97].compute_node[1][2], compute_graph.flexml_layers[97].compute_node[1][3], compute_graph.flexml_layers[97].compute_node[2][0], compute_graph.flexml_layers[97].compute_node[2][1], compute_graph.flexml_layers[97].compute_node[2][2], compute_graph.flexml_layers[97].compute_node[2][3], compute_graph.flexml_layers[97].compute_node[3][0], compute_graph.flexml_layers[97].compute_node[3][1], compute_graph.flexml_layers[97].compute_node[3][2], compute_graph.flexml_layers[97].compute_node[3][3], {compute_graph.l2_110.out[0], compute_graph.l2_110.out[1], compute_graph.l2_110.out[2], compute_graph.l2_110.out[3], compute_graph.l2_111.in[0], compute_graph.l2_111.in[1], compute_graph.l2_111.in[2], compute_graph.l2_111.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1426), layer(122): compute_graph.flexml_layers[98].compute_node[0][0], compute_graph.flexml_layers[98].compute_node[0][1], compute_graph.flexml_layers[98].compute_node[0][2], compute_graph.flexml_layers[98].compute_node[0][3], compute_graph.flexml_layers[98].compute_node[1][0], compute_graph.flexml_layers[98].compute_node[1][1], compute_graph.flexml_layers[98].compute_node[1][2], compute_graph.flexml_layers[98].compute_node[1][3], compute_graph.flexml_layers[98].compute_node[2][0], compute_graph.flexml_layers[98].compute_node[2][1], compute_graph.flexml_layers[98].compute_node[2][2], compute_graph.flexml_layers[98].compute_node[2][3], compute_graph.flexml_layers[98].compute_node[3][0], compute_graph.flexml_layers[98].compute_node[3][1], compute_graph.flexml_layers[98].compute_node[3][2], compute_graph.flexml_layers[98].compute_node[3][3], {compute_graph.l2_111.out[0], compute_graph.l2_111.out[1], compute_graph.l2_111.out[2], compute_graph.l2_111.out[3], compute_graph.l2_112.in[0], compute_graph.l2_112.in[1], compute_graph.l2_112.in[2], compute_graph.l2_112.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1427), layer(123): compute_graph.flexml_layers[99].compute_node[0][0], compute_graph.flexml_layers[99].compute_node[0][1], compute_graph.flexml_layers[99].compute_node[0][2], compute_graph.flexml_layers[99].compute_node[0][3], compute_graph.flexml_layers[99].compute_node[1][0], compute_graph.flexml_layers[99].compute_node[1][1], compute_graph.flexml_layers[99].compute_node[1][2], compute_graph.flexml_layers[99].compute_node[1][3], compute_graph.flexml_layers[99].compute_node[2][0], compute_graph.flexml_layers[99].compute_node[2][1], compute_graph.flexml_layers[99].compute_node[2][2], compute_graph.flexml_layers[99].compute_node[2][3], compute_graph.flexml_layers[99].compute_node[3][0], compute_graph.flexml_layers[99].compute_node[3][1], compute_graph.flexml_layers[99].compute_node[3][2], compute_graph.flexml_layers[99].compute_node[3][3], {compute_graph.l2_109.out[4], compute_graph.l2_109.out[5], compute_graph.l2_109.out[6], compute_graph.l2_109.out[7], compute_graph.l2_112.out[0], compute_graph.l2_112.out[1], compute_graph.l2_112.out[2], compute_graph.l2_112.out[3], compute_graph.l2_113.in[0], compute_graph.l2_113.in[1], compute_graph.l2_113.in[2], compute_graph.l2_113.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(930): mode(0), layer(123): {compute_graph.l2_113.out[2], compute_graph.l2_113.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(929): mode(0), layer(123): {compute_graph.l2_113.out[0], compute_graph.l2_113.out[1], compute_graph.l2l3_113_spill.in[0], compute_graph.l2l3_113_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1074): mode(0), layer(124): {compute_graph.l2l3_113_spill.out[0], compute_graph.templated_graph_114.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1258): mode(0), layer(124): {compute_graph.templated_graph_114.trans_mt_ifm.out[0], compute_graph.templated_graph_114.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1259): mode(0), layer(124): {compute_graph.templated_graph_114.trans_mt_ofm.out[0], compute_graph.l2l3_114_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1075): mode(0), layer(125): {compute_graph.l2l3_114_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1428), layer(125): compute_graph.flexml_layers[100].compute_node[0][0], compute_graph.flexml_layers[100].compute_node[0][1], compute_graph.flexml_layers[100].compute_node[0][2], compute_graph.flexml_layers[100].compute_node[0][3], compute_graph.flexml_layers[100].compute_node[1][0], compute_graph.flexml_layers[100].compute_node[1][1], compute_graph.flexml_layers[100].compute_node[1][2], compute_graph.flexml_layers[100].compute_node[1][3], compute_graph.flexml_layers[100].compute_node[2][0], compute_graph.flexml_layers[100].compute_node[2][1], compute_graph.flexml_layers[100].compute_node[2][2], compute_graph.flexml_layers[100].compute_node[2][3], compute_graph.flexml_layers[100].compute_node[3][0], compute_graph.flexml_layers[100].compute_node[3][1], compute_graph.flexml_layers[100].compute_node[3][2], compute_graph.flexml_layers[100].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.out[3], compute_graph.l2_115.in[0], compute_graph.l2_115.in[1], compute_graph.l2_115.in[2], compute_graph.l2_115.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(857): mode(3), layer(126): {compute_graph.Layer_116_wts_ddr.out[0], compute_graph.Layer_116_wts_ddr.out[1], compute_graph.Layer_116_l2_wts.in[0], compute_graph.Layer_116_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(857) is pipelined with KernelLayerNode(1428) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1429), layer(126): compute_graph.flexml_layers[101].compute_node[0][0], compute_graph.flexml_layers[101].compute_node[0][1], compute_graph.flexml_layers[101].compute_node[0][2], compute_graph.flexml_layers[101].compute_node[0][3], compute_graph.flexml_layers[101].compute_node[1][0], compute_graph.flexml_layers[101].compute_node[1][1], compute_graph.flexml_layers[101].compute_node[1][2], compute_graph.flexml_layers[101].compute_node[1][3], compute_graph.flexml_layers[101].compute_node[2][0], compute_graph.flexml_layers[101].compute_node[2][1], compute_graph.flexml_layers[101].compute_node[2][2], compute_graph.flexml_layers[101].compute_node[2][3], compute_graph.flexml_layers[101].compute_node[3][0], compute_graph.flexml_layers[101].compute_node[3][1], compute_graph.flexml_layers[101].compute_node[3][2], compute_graph.flexml_layers[101].compute_node[3][3], {compute_graph.l2_115.out[0], compute_graph.l2_115.out[1], compute_graph.l2_115.out[2], compute_graph.l2_115.out[3], compute_graph.Layer_116_l2_wts.out[0], compute_graph.Layer_116_l2_wts.out[1], compute_graph.Layer_116_l2_wts.out[2], compute_graph.Layer_116_l2_wts.out[3], compute_graph.l2_116.in[0], compute_graph.l2_116.in[1], compute_graph.l2_116.in[2], compute_graph.l2_116.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(858): mode(3), layer(127): {compute_graph.Layer_117_wts_ddr.out[0], compute_graph.Layer_117_wts_ddr.out[1], compute_graph.Layer_117_l2_wts.in[0], compute_graph.Layer_117_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(858) is pipelined with KernelLayerNode(1429) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1430), layer(127): compute_graph.flexml_layers[102].compute_node[0][0], compute_graph.flexml_layers[102].compute_node[0][1], compute_graph.flexml_layers[102].compute_node[0][2], compute_graph.flexml_layers[102].compute_node[0][3], compute_graph.flexml_layers[102].compute_node[1][0], compute_graph.flexml_layers[102].compute_node[1][1], compute_graph.flexml_layers[102].compute_node[1][2], compute_graph.flexml_layers[102].compute_node[1][3], compute_graph.flexml_layers[102].compute_node[2][0], compute_graph.flexml_layers[102].compute_node[2][1], compute_graph.flexml_layers[102].compute_node[2][2], compute_graph.flexml_layers[102].compute_node[2][3], compute_graph.flexml_layers[102].compute_node[3][0], compute_graph.flexml_layers[102].compute_node[3][1], compute_graph.flexml_layers[102].compute_node[3][2], compute_graph.flexml_layers[102].compute_node[3][3], {compute_graph.l2_116.out[0], compute_graph.l2_116.out[1], compute_graph.l2_116.out[2], compute_graph.l2_116.out[3], compute_graph.Layer_117_l2_wts.out[0], compute_graph.Layer_117_l2_wts.out[1], compute_graph.Layer_117_l2_wts.out[2], compute_graph.Layer_117_l2_wts.out[3], compute_graph.l2_117.in[0], compute_graph.l2_117.in[1], compute_graph.l2_117.in[2], compute_graph.l2_117.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1431), layer(128): compute_graph.flexml_layers[103].compute_node[0][0], compute_graph.flexml_layers[103].compute_node[0][1], compute_graph.flexml_layers[103].compute_node[0][2], compute_graph.flexml_layers[103].compute_node[0][3], compute_graph.flexml_layers[103].compute_node[1][0], compute_graph.flexml_layers[103].compute_node[1][1], compute_graph.flexml_layers[103].compute_node[1][2], compute_graph.flexml_layers[103].compute_node[1][3], compute_graph.flexml_layers[103].compute_node[2][0], compute_graph.flexml_layers[103].compute_node[2][1], compute_graph.flexml_layers[103].compute_node[2][2], compute_graph.flexml_layers[103].compute_node[2][3], compute_graph.flexml_layers[103].compute_node[3][0], compute_graph.flexml_layers[103].compute_node[3][1], compute_graph.flexml_layers[103].compute_node[3][2], compute_graph.flexml_layers[103].compute_node[3][3], {compute_graph.l2_117.out[0], compute_graph.l2_117.out[1], compute_graph.l2_117.out[2], compute_graph.l2_117.out[3], compute_graph.l2_118.in[0], compute_graph.l2_118.in[1], compute_graph.l2_118.in[2], compute_graph.l2_118.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1432), layer(129): compute_graph.flexml_layers[104].compute_node[0][0], compute_graph.flexml_layers[104].compute_node[0][1], compute_graph.flexml_layers[104].compute_node[0][2], compute_graph.flexml_layers[104].compute_node[0][3], compute_graph.flexml_layers[104].compute_node[1][0], compute_graph.flexml_layers[104].compute_node[1][1], compute_graph.flexml_layers[104].compute_node[1][2], compute_graph.flexml_layers[104].compute_node[1][3], compute_graph.flexml_layers[104].compute_node[2][0], compute_graph.flexml_layers[104].compute_node[2][1], compute_graph.flexml_layers[104].compute_node[2][2], compute_graph.flexml_layers[104].compute_node[2][3], compute_graph.flexml_layers[104].compute_node[3][0], compute_graph.flexml_layers[104].compute_node[3][1], compute_graph.flexml_layers[104].compute_node[3][2], compute_graph.flexml_layers[104].compute_node[3][3], {compute_graph.l2_118.out[0], compute_graph.l2_118.out[1], compute_graph.l2_118.out[2], compute_graph.l2_118.out[3], compute_graph.l2_119.in[0], compute_graph.l2_119.in[1], compute_graph.l2_119.in[2], compute_graph.l2_119.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1433), layer(130): compute_graph.flexml_layers[105].compute_node[0][0], compute_graph.flexml_layers[105].compute_node[0][1], compute_graph.flexml_layers[105].compute_node[0][2], compute_graph.flexml_layers[105].compute_node[0][3], compute_graph.flexml_layers[105].compute_node[1][0], compute_graph.flexml_layers[105].compute_node[1][1], compute_graph.flexml_layers[105].compute_node[1][2], compute_graph.flexml_layers[105].compute_node[1][3], compute_graph.flexml_layers[105].compute_node[2][0], compute_graph.flexml_layers[105].compute_node[2][1], compute_graph.flexml_layers[105].compute_node[2][2], compute_graph.flexml_layers[105].compute_node[2][3], compute_graph.flexml_layers[105].compute_node[3][0], compute_graph.flexml_layers[105].compute_node[3][1], compute_graph.flexml_layers[105].compute_node[3][2], compute_graph.flexml_layers[105].compute_node[3][3], {compute_graph.l2_119.out[0], compute_graph.l2_119.out[1], compute_graph.l2_119.out[2], compute_graph.l2_119.out[3], compute_graph.l2_120.in[0], compute_graph.l2_120.in[1], compute_graph.l2_120.in[2], compute_graph.l2_120.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(931): mode(0), layer(130): {compute_graph.l2_120.out[1], compute_graph.l2l3_120_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1076): mode(0), layer(131): {compute_graph.l2l3_120_spill.out[0], compute_graph.templated_graph_121.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1260): mode(0), layer(131): {compute_graph.templated_graph_121.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_121_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1077): mode(0), layer(132): {compute_graph.l2l3_scratch_0_121_spill.out[0], compute_graph.templated_graph_121.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1261): mode(0), layer(132): {compute_graph.templated_graph_121.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_121_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1078): mode(0), layer(133): {compute_graph.l2l3_scratch_1_121_spill.out[0], compute_graph.templated_graph_121.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1262): mode(0), layer(133): {compute_graph.templated_graph_121.g2.ifm_mem.out[0], compute_graph.l2l3_121_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1073): mode(0), layer(134): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1079): mode(0), layer(134): {compute_graph.l2l3_121_spill.out[0], compute_graph.l2l3_121_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1434), layer(134): compute_graph.flexml_layers[106].compute_node[0][0], compute_graph.flexml_layers[106].compute_node[0][1], compute_graph.flexml_layers[106].compute_node[0][2], compute_graph.flexml_layers[106].compute_node[0][3], compute_graph.flexml_layers[106].compute_node[1][0], compute_graph.flexml_layers[106].compute_node[1][1], compute_graph.flexml_layers[106].compute_node[1][2], compute_graph.flexml_layers[106].compute_node[1][3], compute_graph.flexml_layers[106].compute_node[2][0], compute_graph.flexml_layers[106].compute_node[2][1], compute_graph.flexml_layers[106].compute_node[2][2], compute_graph.flexml_layers[106].compute_node[2][3], compute_graph.flexml_layers[106].compute_node[3][0], compute_graph.flexml_layers[106].compute_node[3][1], compute_graph.flexml_layers[106].compute_node[3][2], compute_graph.flexml_layers[106].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.out[3], compute_graph.l2_122.in[0], compute_graph.l2_122.in[1], compute_graph.l2_122.in[2], compute_graph.l2_122.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(859): mode(3), layer(135): {compute_graph.Layer_123_wts_ddr.out[0], compute_graph.Layer_123_wts_ddr.out[1], compute_graph.Layer_123_l2_wts.in[0], compute_graph.Layer_123_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(859) is pipelined with KernelLayerNode(1434) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1435), layer(135): compute_graph.flexml_layers[107].compute_node[0][0], compute_graph.flexml_layers[107].compute_node[0][1], compute_graph.flexml_layers[107].compute_node[0][2], compute_graph.flexml_layers[107].compute_node[0][3], compute_graph.flexml_layers[107].compute_node[1][0], compute_graph.flexml_layers[107].compute_node[1][1], compute_graph.flexml_layers[107].compute_node[1][2], compute_graph.flexml_layers[107].compute_node[1][3], compute_graph.flexml_layers[107].compute_node[2][0], compute_graph.flexml_layers[107].compute_node[2][1], compute_graph.flexml_layers[107].compute_node[2][2], compute_graph.flexml_layers[107].compute_node[2][3], compute_graph.flexml_layers[107].compute_node[3][0], compute_graph.flexml_layers[107].compute_node[3][1], compute_graph.flexml_layers[107].compute_node[3][2], compute_graph.flexml_layers[107].compute_node[3][3], {compute_graph.l2_122.out[0], compute_graph.l2_122.out[1], compute_graph.l2_122.out[2], compute_graph.l2_122.out[3], compute_graph.Layer_123_l2_wts.out[0], compute_graph.Layer_123_l2_wts.out[1], compute_graph.Layer_123_l2_wts.out[2], compute_graph.Layer_123_l2_wts.out[3], compute_graph.l2_123.in[0], compute_graph.l2_123.in[1], compute_graph.l2_123.in[2], compute_graph.l2_123.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(860): mode(3), layer(136): {compute_graph.Layer_124_wts_ddr.out[0], compute_graph.Layer_124_wts_ddr.out[1], compute_graph.Layer_124_l2_wts.in[0], compute_graph.Layer_124_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(860) is pipelined with KernelLayerNode(1435) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1436), layer(136): compute_graph.flexml_layers[108].compute_node[0][0], compute_graph.flexml_layers[108].compute_node[0][1], compute_graph.flexml_layers[108].compute_node[0][2], compute_graph.flexml_layers[108].compute_node[0][3], compute_graph.flexml_layers[108].compute_node[1][0], compute_graph.flexml_layers[108].compute_node[1][1], compute_graph.flexml_layers[108].compute_node[1][2], compute_graph.flexml_layers[108].compute_node[1][3], compute_graph.flexml_layers[108].compute_node[2][0], compute_graph.flexml_layers[108].compute_node[2][1], compute_graph.flexml_layers[108].compute_node[2][2], compute_graph.flexml_layers[108].compute_node[2][3], compute_graph.flexml_layers[108].compute_node[3][0], compute_graph.flexml_layers[108].compute_node[3][1], compute_graph.flexml_layers[108].compute_node[3][2], compute_graph.flexml_layers[108].compute_node[3][3], {compute_graph.l2_123.out[0], compute_graph.l2_123.out[1], compute_graph.l2_123.out[2], compute_graph.l2_123.out[3], compute_graph.Layer_124_l2_wts.out[0], compute_graph.Layer_124_l2_wts.out[1], compute_graph.Layer_124_l2_wts.out[2], compute_graph.Layer_124_l2_wts.out[3], compute_graph.l2_124.in[0], compute_graph.l2_124.in[1], compute_graph.l2_124.in[2], compute_graph.l2_124.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(932): mode(3), layer(136): {compute_graph.l2_123.out[4], compute_graph.l2_123.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(932) is pipelined with KernelLayerNode(1436) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1437), layer(137): compute_graph.flexml_layers[109].compute_node[0][0], compute_graph.flexml_layers[109].compute_node[0][1], compute_graph.flexml_layers[109].compute_node[0][2], compute_graph.flexml_layers[109].compute_node[0][3], compute_graph.flexml_layers[109].compute_node[1][0], compute_graph.flexml_layers[109].compute_node[1][1], compute_graph.flexml_layers[109].compute_node[1][2], compute_graph.flexml_layers[109].compute_node[1][3], compute_graph.flexml_layers[109].compute_node[2][0], compute_graph.flexml_layers[109].compute_node[2][1], compute_graph.flexml_layers[109].compute_node[2][2], compute_graph.flexml_layers[109].compute_node[2][3], compute_graph.flexml_layers[109].compute_node[3][0], compute_graph.flexml_layers[109].compute_node[3][1], compute_graph.flexml_layers[109].compute_node[3][2], compute_graph.flexml_layers[109].compute_node[3][3], {compute_graph.l2_124.out[0], compute_graph.l2_124.out[1], compute_graph.l2_124.out[2], compute_graph.l2_124.out[3], compute_graph.l2_125.in[0], compute_graph.l2_125.in[1], compute_graph.l2_125.in[2], compute_graph.l2_125.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1438), layer(138): compute_graph.flexml_layers[110].compute_node[0][0], compute_graph.flexml_layers[110].compute_node[0][1], compute_graph.flexml_layers[110].compute_node[0][2], compute_graph.flexml_layers[110].compute_node[0][3], compute_graph.flexml_layers[110].compute_node[1][0], compute_graph.flexml_layers[110].compute_node[1][1], compute_graph.flexml_layers[110].compute_node[1][2], compute_graph.flexml_layers[110].compute_node[1][3], compute_graph.flexml_layers[110].compute_node[2][0], compute_graph.flexml_layers[110].compute_node[2][1], compute_graph.flexml_layers[110].compute_node[2][2], compute_graph.flexml_layers[110].compute_node[2][3], compute_graph.flexml_layers[110].compute_node[3][0], compute_graph.flexml_layers[110].compute_node[3][1], compute_graph.flexml_layers[110].compute_node[3][2], compute_graph.flexml_layers[110].compute_node[3][3], {compute_graph.l2_125.out[0], compute_graph.l2_125.out[1], compute_graph.l2_125.out[2], compute_graph.l2_125.out[3], compute_graph.l2_126.in[0], compute_graph.l2_126.in[1], compute_graph.l2_126.in[2], compute_graph.l2_126.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1439), layer(139): compute_graph.flexml_layers[111].compute_node[0][0], compute_graph.flexml_layers[111].compute_node[0][1], compute_graph.flexml_layers[111].compute_node[0][2], compute_graph.flexml_layers[111].compute_node[0][3], compute_graph.flexml_layers[111].compute_node[1][0], compute_graph.flexml_layers[111].compute_node[1][1], compute_graph.flexml_layers[111].compute_node[1][2], compute_graph.flexml_layers[111].compute_node[1][3], compute_graph.flexml_layers[111].compute_node[2][0], compute_graph.flexml_layers[111].compute_node[2][1], compute_graph.flexml_layers[111].compute_node[2][2], compute_graph.flexml_layers[111].compute_node[2][3], compute_graph.flexml_layers[111].compute_node[3][0], compute_graph.flexml_layers[111].compute_node[3][1], compute_graph.flexml_layers[111].compute_node[3][2], compute_graph.flexml_layers[111].compute_node[3][3], {compute_graph.l2_126.out[0], compute_graph.l2_126.out[1], compute_graph.l2_126.out[2], compute_graph.l2_126.out[3], compute_graph.l2_127.in[0], compute_graph.l2_127.in[1], compute_graph.l2_127.in[2], compute_graph.l2_127.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1440), layer(140): compute_graph.flexml_layers[112].compute_node[0][0], compute_graph.flexml_layers[112].compute_node[0][1], compute_graph.flexml_layers[112].compute_node[0][2], compute_graph.flexml_layers[112].compute_node[0][3], compute_graph.flexml_layers[112].compute_node[1][0], compute_graph.flexml_layers[112].compute_node[1][1], compute_graph.flexml_layers[112].compute_node[1][2], compute_graph.flexml_layers[112].compute_node[1][3], compute_graph.flexml_layers[112].compute_node[2][0], compute_graph.flexml_layers[112].compute_node[2][1], compute_graph.flexml_layers[112].compute_node[2][2], compute_graph.flexml_layers[112].compute_node[2][3], compute_graph.flexml_layers[112].compute_node[3][0], compute_graph.flexml_layers[112].compute_node[3][1], compute_graph.flexml_layers[112].compute_node[3][2], compute_graph.flexml_layers[112].compute_node[3][3], {compute_graph.l2_124.out[4], compute_graph.l2_124.out[5], compute_graph.l2_124.out[6], compute_graph.l2_124.out[7], compute_graph.l2_127.out[0], compute_graph.l2_127.out[1], compute_graph.l2_127.out[2], compute_graph.l2_127.out[3], compute_graph.l2_128.in[0], compute_graph.l2_128.in[1], compute_graph.l2_128.in[2], compute_graph.l2_128.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(861): mode(3), layer(141): {compute_graph.Layer_129_wts_ddr.out[0], compute_graph.Layer_129_wts_ddr.out[1], compute_graph.Layer_129_l2_wts.in[0], compute_graph.Layer_129_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(861) is pipelined with KernelLayerNode(1440) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1441), layer(141): compute_graph.flexml_layers[113].compute_node[0][0], compute_graph.flexml_layers[113].compute_node[0][1], compute_graph.flexml_layers[113].compute_node[0][2], compute_graph.flexml_layers[113].compute_node[0][3], compute_graph.flexml_layers[113].compute_node[1][0], compute_graph.flexml_layers[113].compute_node[1][1], compute_graph.flexml_layers[113].compute_node[1][2], compute_graph.flexml_layers[113].compute_node[1][3], compute_graph.flexml_layers[113].compute_node[2][0], compute_graph.flexml_layers[113].compute_node[2][1], compute_graph.flexml_layers[113].compute_node[2][2], compute_graph.flexml_layers[113].compute_node[2][3], compute_graph.flexml_layers[113].compute_node[3][0], compute_graph.flexml_layers[113].compute_node[3][1], compute_graph.flexml_layers[113].compute_node[3][2], compute_graph.flexml_layers[113].compute_node[3][3], {compute_graph.l2_128.out[0], compute_graph.l2_128.out[1], compute_graph.l2_128.out[2], compute_graph.l2_128.out[3], compute_graph.Layer_129_l2_wts.out[0], compute_graph.Layer_129_l2_wts.out[1], compute_graph.Layer_129_l2_wts.out[2], compute_graph.Layer_129_l2_wts.out[3], compute_graph.l2_129.in[0], compute_graph.l2_129.in[1], compute_graph.l2_129.in[2], compute_graph.l2_129.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1442), layer(142): compute_graph.flexml_layers[114].compute_node[0][0], compute_graph.flexml_layers[114].compute_node[0][1], compute_graph.flexml_layers[114].compute_node[0][2], compute_graph.flexml_layers[114].compute_node[0][3], compute_graph.flexml_layers[114].compute_node[1][0], compute_graph.flexml_layers[114].compute_node[1][1], compute_graph.flexml_layers[114].compute_node[1][2], compute_graph.flexml_layers[114].compute_node[1][3], compute_graph.flexml_layers[114].compute_node[2][0], compute_graph.flexml_layers[114].compute_node[2][1], compute_graph.flexml_layers[114].compute_node[2][2], compute_graph.flexml_layers[114].compute_node[2][3], compute_graph.flexml_layers[114].compute_node[3][0], compute_graph.flexml_layers[114].compute_node[3][1], compute_graph.flexml_layers[114].compute_node[3][2], compute_graph.flexml_layers[114].compute_node[3][3], {compute_graph.l2_129.out[0], compute_graph.l2_129.out[1], compute_graph.l2_129.out[2], compute_graph.l2_129.out[3], compute_graph.l2_130.in[0], compute_graph.l2_130.in[1], compute_graph.l2_130.in[2], compute_graph.l2_130.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1443), layer(143): compute_graph.flexml_layers[115].compute_node[0][0], compute_graph.flexml_layers[115].compute_node[0][1], compute_graph.flexml_layers[115].compute_node[0][2], compute_graph.flexml_layers[115].compute_node[0][3], compute_graph.flexml_layers[115].compute_node[1][0], compute_graph.flexml_layers[115].compute_node[1][1], compute_graph.flexml_layers[115].compute_node[1][2], compute_graph.flexml_layers[115].compute_node[1][3], compute_graph.flexml_layers[115].compute_node[2][0], compute_graph.flexml_layers[115].compute_node[2][1], compute_graph.flexml_layers[115].compute_node[2][2], compute_graph.flexml_layers[115].compute_node[2][3], compute_graph.flexml_layers[115].compute_node[3][0], compute_graph.flexml_layers[115].compute_node[3][1], compute_graph.flexml_layers[115].compute_node[3][2], compute_graph.flexml_layers[115].compute_node[3][3], {compute_graph.l2_130.out[0], compute_graph.l2_130.out[1], compute_graph.l2_130.out[2], compute_graph.l2_130.out[3], compute_graph.l2_131.in[0], compute_graph.l2_131.in[1], compute_graph.l2_131.in[2], compute_graph.l2_131.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1444), layer(144): compute_graph.flexml_layers[116].compute_node[0][0], compute_graph.flexml_layers[116].compute_node[0][1], compute_graph.flexml_layers[116].compute_node[0][2], compute_graph.flexml_layers[116].compute_node[0][3], compute_graph.flexml_layers[116].compute_node[1][0], compute_graph.flexml_layers[116].compute_node[1][1], compute_graph.flexml_layers[116].compute_node[1][2], compute_graph.flexml_layers[116].compute_node[1][3], compute_graph.flexml_layers[116].compute_node[2][0], compute_graph.flexml_layers[116].compute_node[2][1], compute_graph.flexml_layers[116].compute_node[2][2], compute_graph.flexml_layers[116].compute_node[2][3], compute_graph.flexml_layers[116].compute_node[3][0], compute_graph.flexml_layers[116].compute_node[3][1], compute_graph.flexml_layers[116].compute_node[3][2], compute_graph.flexml_layers[116].compute_node[3][3], {compute_graph.l2_131.out[0], compute_graph.l2_131.out[1], compute_graph.l2_131.out[2], compute_graph.l2_131.out[3], compute_graph.l2_132.in[0], compute_graph.l2_132.in[1], compute_graph.l2_132.in[2], compute_graph.l2_132.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1445), layer(145): compute_graph.flexml_layers[117].compute_node[0][0], compute_graph.flexml_layers[117].compute_node[0][1], compute_graph.flexml_layers[117].compute_node[0][2], compute_graph.flexml_layers[117].compute_node[0][3], compute_graph.flexml_layers[117].compute_node[1][0], compute_graph.flexml_layers[117].compute_node[1][1], compute_graph.flexml_layers[117].compute_node[1][2], compute_graph.flexml_layers[117].compute_node[1][3], compute_graph.flexml_layers[117].compute_node[2][0], compute_graph.flexml_layers[117].compute_node[2][1], compute_graph.flexml_layers[117].compute_node[2][2], compute_graph.flexml_layers[117].compute_node[2][3], compute_graph.flexml_layers[117].compute_node[3][0], compute_graph.flexml_layers[117].compute_node[3][1], compute_graph.flexml_layers[117].compute_node[3][2], compute_graph.flexml_layers[117].compute_node[3][3], {compute_graph.l2_129.out[4], compute_graph.l2_129.out[5], compute_graph.l2_129.out[6], compute_graph.l2_129.out[7], compute_graph.l2_132.out[0], compute_graph.l2_132.out[1], compute_graph.l2_132.out[2], compute_graph.l2_132.out[3], compute_graph.l2_133.in[0], compute_graph.l2_133.in[1], compute_graph.l2_133.in[2], compute_graph.l2_133.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(933): mode(0), layer(145): {compute_graph.l2_133.out[0], compute_graph.l2_133.out[1], compute_graph.l2l3_133_spill.in[0], compute_graph.l2l3_133_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(934): mode(0), layer(145): {compute_graph.l2_133.out[2], compute_graph.l2_133.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1082): mode(0), layer(146): {compute_graph.l2l3_133_spill.out[0], compute_graph.templated_graph_134.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1263): mode(0), layer(146): {compute_graph.templated_graph_134.trans_mt_ifm.out[0], compute_graph.templated_graph_134.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1264): mode(0), layer(146): {compute_graph.templated_graph_134.trans_mt_ofm.out[0], compute_graph.l2l3_134_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1083): mode(0), layer(147): {compute_graph.l2l3_134_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1446), layer(147): compute_graph.flexml_layers[118].compute_node[0][0], compute_graph.flexml_layers[118].compute_node[0][1], compute_graph.flexml_layers[118].compute_node[0][2], compute_graph.flexml_layers[118].compute_node[0][3], compute_graph.flexml_layers[118].compute_node[1][0], compute_graph.flexml_layers[118].compute_node[1][1], compute_graph.flexml_layers[118].compute_node[1][2], compute_graph.flexml_layers[118].compute_node[1][3], compute_graph.flexml_layers[118].compute_node[2][0], compute_graph.flexml_layers[118].compute_node[2][1], compute_graph.flexml_layers[118].compute_node[2][2], compute_graph.flexml_layers[118].compute_node[2][3], compute_graph.flexml_layers[118].compute_node[3][0], compute_graph.flexml_layers[118].compute_node[3][1], compute_graph.flexml_layers[118].compute_node[3][2], compute_graph.flexml_layers[118].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.out[3], compute_graph.l2_135.in[0], compute_graph.l2_135.in[1], compute_graph.l2_135.in[2], compute_graph.l2_135.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(862): mode(3), layer(148): {compute_graph.Layer_136_wts_ddr.out[0], compute_graph.Layer_136_wts_ddr.out[1], compute_graph.Layer_136_l2_wts.in[0], compute_graph.Layer_136_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(862) is pipelined with KernelLayerNode(1446) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1447), layer(148): compute_graph.flexml_layers[119].compute_node[0][0], compute_graph.flexml_layers[119].compute_node[0][1], compute_graph.flexml_layers[119].compute_node[0][2], compute_graph.flexml_layers[119].compute_node[0][3], compute_graph.flexml_layers[119].compute_node[1][0], compute_graph.flexml_layers[119].compute_node[1][1], compute_graph.flexml_layers[119].compute_node[1][2], compute_graph.flexml_layers[119].compute_node[1][3], compute_graph.flexml_layers[119].compute_node[2][0], compute_graph.flexml_layers[119].compute_node[2][1], compute_graph.flexml_layers[119].compute_node[2][2], compute_graph.flexml_layers[119].compute_node[2][3], compute_graph.flexml_layers[119].compute_node[3][0], compute_graph.flexml_layers[119].compute_node[3][1], compute_graph.flexml_layers[119].compute_node[3][2], compute_graph.flexml_layers[119].compute_node[3][3], {compute_graph.l2_135.out[0], compute_graph.l2_135.out[1], compute_graph.l2_135.out[2], compute_graph.l2_135.out[3], compute_graph.Layer_136_l2_wts.out[0], compute_graph.Layer_136_l2_wts.out[1], compute_graph.Layer_136_l2_wts.out[2], compute_graph.Layer_136_l2_wts.out[3], compute_graph.l2_136.in[0], compute_graph.l2_136.in[1], compute_graph.l2_136.in[2], compute_graph.l2_136.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(863): mode(3), layer(149): {compute_graph.Layer_137_wts_ddr.out[0], compute_graph.Layer_137_wts_ddr.out[1], compute_graph.Layer_137_l2_wts.in[0], compute_graph.Layer_137_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(863) is pipelined with KernelLayerNode(1447) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1448), layer(149): compute_graph.flexml_layers[120].compute_node[0][0], compute_graph.flexml_layers[120].compute_node[0][1], compute_graph.flexml_layers[120].compute_node[0][2], compute_graph.flexml_layers[120].compute_node[0][3], compute_graph.flexml_layers[120].compute_node[1][0], compute_graph.flexml_layers[120].compute_node[1][1], compute_graph.flexml_layers[120].compute_node[1][2], compute_graph.flexml_layers[120].compute_node[1][3], compute_graph.flexml_layers[120].compute_node[2][0], compute_graph.flexml_layers[120].compute_node[2][1], compute_graph.flexml_layers[120].compute_node[2][2], compute_graph.flexml_layers[120].compute_node[2][3], compute_graph.flexml_layers[120].compute_node[3][0], compute_graph.flexml_layers[120].compute_node[3][1], compute_graph.flexml_layers[120].compute_node[3][2], compute_graph.flexml_layers[120].compute_node[3][3], {compute_graph.l2_136.out[0], compute_graph.l2_136.out[1], compute_graph.l2_136.out[2], compute_graph.l2_136.out[3], compute_graph.Layer_137_l2_wts.out[0], compute_graph.Layer_137_l2_wts.out[1], compute_graph.Layer_137_l2_wts.out[2], compute_graph.Layer_137_l2_wts.out[3], compute_graph.l2_137.in[0], compute_graph.l2_137.in[1], compute_graph.l2_137.in[2], compute_graph.l2_137.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1449), layer(150): compute_graph.flexml_layers[121].compute_node[0][0], compute_graph.flexml_layers[121].compute_node[0][1], compute_graph.flexml_layers[121].compute_node[0][2], compute_graph.flexml_layers[121].compute_node[0][3], compute_graph.flexml_layers[121].compute_node[1][0], compute_graph.flexml_layers[121].compute_node[1][1], compute_graph.flexml_layers[121].compute_node[1][2], compute_graph.flexml_layers[121].compute_node[1][3], compute_graph.flexml_layers[121].compute_node[2][0], compute_graph.flexml_layers[121].compute_node[2][1], compute_graph.flexml_layers[121].compute_node[2][2], compute_graph.flexml_layers[121].compute_node[2][3], compute_graph.flexml_layers[121].compute_node[3][0], compute_graph.flexml_layers[121].compute_node[3][1], compute_graph.flexml_layers[121].compute_node[3][2], compute_graph.flexml_layers[121].compute_node[3][3], {compute_graph.l2_137.out[0], compute_graph.l2_137.out[1], compute_graph.l2_137.out[2], compute_graph.l2_137.out[3], compute_graph.l2_138.in[0], compute_graph.l2_138.in[1], compute_graph.l2_138.in[2], compute_graph.l2_138.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1450), layer(151): compute_graph.flexml_layers[122].compute_node[0][0], compute_graph.flexml_layers[122].compute_node[0][1], compute_graph.flexml_layers[122].compute_node[0][2], compute_graph.flexml_layers[122].compute_node[0][3], compute_graph.flexml_layers[122].compute_node[1][0], compute_graph.flexml_layers[122].compute_node[1][1], compute_graph.flexml_layers[122].compute_node[1][2], compute_graph.flexml_layers[122].compute_node[1][3], compute_graph.flexml_layers[122].compute_node[2][0], compute_graph.flexml_layers[122].compute_node[2][1], compute_graph.flexml_layers[122].compute_node[2][2], compute_graph.flexml_layers[122].compute_node[2][3], compute_graph.flexml_layers[122].compute_node[3][0], compute_graph.flexml_layers[122].compute_node[3][1], compute_graph.flexml_layers[122].compute_node[3][2], compute_graph.flexml_layers[122].compute_node[3][3], {compute_graph.l2_138.out[0], compute_graph.l2_138.out[1], compute_graph.l2_138.out[2], compute_graph.l2_138.out[3], compute_graph.l2_139.in[0], compute_graph.l2_139.in[1], compute_graph.l2_139.in[2], compute_graph.l2_139.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1451), layer(152): compute_graph.flexml_layers[123].compute_node[0][0], compute_graph.flexml_layers[123].compute_node[0][1], compute_graph.flexml_layers[123].compute_node[0][2], compute_graph.flexml_layers[123].compute_node[0][3], compute_graph.flexml_layers[123].compute_node[1][0], compute_graph.flexml_layers[123].compute_node[1][1], compute_graph.flexml_layers[123].compute_node[1][2], compute_graph.flexml_layers[123].compute_node[1][3], compute_graph.flexml_layers[123].compute_node[2][0], compute_graph.flexml_layers[123].compute_node[2][1], compute_graph.flexml_layers[123].compute_node[2][2], compute_graph.flexml_layers[123].compute_node[2][3], compute_graph.flexml_layers[123].compute_node[3][0], compute_graph.flexml_layers[123].compute_node[3][1], compute_graph.flexml_layers[123].compute_node[3][2], compute_graph.flexml_layers[123].compute_node[3][3], {compute_graph.l2_139.out[0], compute_graph.l2_139.out[1], compute_graph.l2_139.out[2], compute_graph.l2_139.out[3], compute_graph.l2_140.in[0], compute_graph.l2_140.in[1], compute_graph.l2_140.in[2], compute_graph.l2_140.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(935): mode(0), layer(152): {compute_graph.l2_140.out[1], compute_graph.l2l3_140_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1084): mode(0), layer(153): {compute_graph.l2l3_140_spill.out[0], compute_graph.templated_graph_141.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1265): mode(0), layer(153): {compute_graph.templated_graph_141.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_141_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1085): mode(0), layer(154): {compute_graph.l2l3_scratch_0_141_spill.out[0], compute_graph.templated_graph_141.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1266): mode(0), layer(154): {compute_graph.templated_graph_141.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_141_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1086): mode(0), layer(155): {compute_graph.l2l3_scratch_1_141_spill.out[0], compute_graph.templated_graph_141.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1267): mode(0), layer(155): {compute_graph.templated_graph_141.g2.ifm_mem.out[0], compute_graph.l2l3_141_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1081): mode(0), layer(156): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1087): mode(0), layer(156): {compute_graph.l2l3_141_spill.out[0], compute_graph.l2l3_141_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1452), layer(156): compute_graph.flexml_layers[124].compute_node[0][0], compute_graph.flexml_layers[124].compute_node[0][1], compute_graph.flexml_layers[124].compute_node[0][2], compute_graph.flexml_layers[124].compute_node[0][3], compute_graph.flexml_layers[124].compute_node[1][0], compute_graph.flexml_layers[124].compute_node[1][1], compute_graph.flexml_layers[124].compute_node[1][2], compute_graph.flexml_layers[124].compute_node[1][3], compute_graph.flexml_layers[124].compute_node[2][0], compute_graph.flexml_layers[124].compute_node[2][1], compute_graph.flexml_layers[124].compute_node[2][2], compute_graph.flexml_layers[124].compute_node[2][3], compute_graph.flexml_layers[124].compute_node[3][0], compute_graph.flexml_layers[124].compute_node[3][1], compute_graph.flexml_layers[124].compute_node[3][2], compute_graph.flexml_layers[124].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.out[3], compute_graph.l2_142.in[0], compute_graph.l2_142.in[1], compute_graph.l2_142.in[2], compute_graph.l2_142.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(864): mode(3), layer(157): {compute_graph.Layer_143_wts_ddr.out[0], compute_graph.Layer_143_wts_ddr.out[1], compute_graph.Layer_143_l2_wts.in[0], compute_graph.Layer_143_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(864) is pipelined with KernelLayerNode(1452) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123 overlaps with memory space used by compute_graph.l2_143. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1080): mode(0), layer(157): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1453), layer(157): compute_graph.flexml_layers[125].compute_node[0][0], compute_graph.flexml_layers[125].compute_node[0][1], compute_graph.flexml_layers[125].compute_node[0][2], compute_graph.flexml_layers[125].compute_node[0][3], compute_graph.flexml_layers[125].compute_node[1][0], compute_graph.flexml_layers[125].compute_node[1][1], compute_graph.flexml_layers[125].compute_node[1][2], compute_graph.flexml_layers[125].compute_node[1][3], compute_graph.flexml_layers[125].compute_node[2][0], compute_graph.flexml_layers[125].compute_node[2][1], compute_graph.flexml_layers[125].compute_node[2][2], compute_graph.flexml_layers[125].compute_node[2][3], compute_graph.flexml_layers[125].compute_node[3][0], compute_graph.flexml_layers[125].compute_node[3][1], compute_graph.flexml_layers[125].compute_node[3][2], compute_graph.flexml_layers[125].compute_node[3][3], {compute_graph.l2_142.out[0], compute_graph.l2_142.out[1], compute_graph.l2_142.out[2], compute_graph.l2_142.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.out[3], compute_graph.Layer_143_l2_wts.out[0], compute_graph.Layer_143_l2_wts.out[1], compute_graph.Layer_143_l2_wts.out[2], compute_graph.Layer_143_l2_wts.out[3], compute_graph.l2_143.in[0], compute_graph.l2_143.in[1], compute_graph.l2_143.in[2], compute_graph.l2_143.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(865): mode(3), layer(158): {compute_graph.Layer_144_wts_ddr.out[0], compute_graph.Layer_144_wts_ddr.out[1], compute_graph.Layer_144_l2_wts.in[0], compute_graph.Layer_144_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(865) is pipelined with KernelLayerNode(1453) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1454), layer(158): compute_graph.flexml_layers[126].compute_node[0][0], compute_graph.flexml_layers[126].compute_node[0][1], compute_graph.flexml_layers[126].compute_node[0][2], compute_graph.flexml_layers[126].compute_node[0][3], compute_graph.flexml_layers[126].compute_node[1][0], compute_graph.flexml_layers[126].compute_node[1][1], compute_graph.flexml_layers[126].compute_node[1][2], compute_graph.flexml_layers[126].compute_node[1][3], compute_graph.flexml_layers[126].compute_node[2][0], compute_graph.flexml_layers[126].compute_node[2][1], compute_graph.flexml_layers[126].compute_node[2][2], compute_graph.flexml_layers[126].compute_node[2][3], compute_graph.flexml_layers[126].compute_node[3][0], compute_graph.flexml_layers[126].compute_node[3][1], compute_graph.flexml_layers[126].compute_node[3][2], compute_graph.flexml_layers[126].compute_node[3][3], {compute_graph.l2_143.out[0], compute_graph.l2_143.out[1], compute_graph.l2_143.out[2], compute_graph.l2_143.out[3], compute_graph.Layer_144_l2_wts.out[0], compute_graph.Layer_144_l2_wts.out[1], compute_graph.Layer_144_l2_wts.out[2], compute_graph.Layer_144_l2_wts.out[3], compute_graph.l2_144.in[0], compute_graph.l2_144.in[1], compute_graph.l2_144.in[2], compute_graph.l2_144.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1455), layer(159): compute_graph.flexml_layers[127].compute_node[0][0], compute_graph.flexml_layers[127].compute_node[0][1], compute_graph.flexml_layers[127].compute_node[0][2], compute_graph.flexml_layers[127].compute_node[0][3], compute_graph.flexml_layers[127].compute_node[1][0], compute_graph.flexml_layers[127].compute_node[1][1], compute_graph.flexml_layers[127].compute_node[1][2], compute_graph.flexml_layers[127].compute_node[1][3], compute_graph.flexml_layers[127].compute_node[2][0], compute_graph.flexml_layers[127].compute_node[2][1], compute_graph.flexml_layers[127].compute_node[2][2], compute_graph.flexml_layers[127].compute_node[2][3], compute_graph.flexml_layers[127].compute_node[3][0], compute_graph.flexml_layers[127].compute_node[3][1], compute_graph.flexml_layers[127].compute_node[3][2], compute_graph.flexml_layers[127].compute_node[3][3], {compute_graph.l2_144.out[0], compute_graph.l2_144.out[1], compute_graph.l2_144.out[2], compute_graph.l2_144.out[3], compute_graph.l2_145.in[0], compute_graph.l2_145.in[1], compute_graph.l2_145.in[2], compute_graph.l2_145.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1456), layer(160): compute_graph.flexml_layers[128].compute_node[0][0], compute_graph.flexml_layers[128].compute_node[0][1], compute_graph.flexml_layers[128].compute_node[0][2], compute_graph.flexml_layers[128].compute_node[0][3], compute_graph.flexml_layers[128].compute_node[1][0], compute_graph.flexml_layers[128].compute_node[1][1], compute_graph.flexml_layers[128].compute_node[1][2], compute_graph.flexml_layers[128].compute_node[1][3], compute_graph.flexml_layers[128].compute_node[2][0], compute_graph.flexml_layers[128].compute_node[2][1], compute_graph.flexml_layers[128].compute_node[2][2], compute_graph.flexml_layers[128].compute_node[2][3], compute_graph.flexml_layers[128].compute_node[3][0], compute_graph.flexml_layers[128].compute_node[3][1], compute_graph.flexml_layers[128].compute_node[3][2], compute_graph.flexml_layers[128].compute_node[3][3], {compute_graph.l2_145.out[0], compute_graph.l2_145.out[1], compute_graph.l2_145.out[2], compute_graph.l2_145.out[3], compute_graph.l2_146.in[0], compute_graph.l2_146.in[1], compute_graph.l2_146.in[2], compute_graph.l2_146.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1457), layer(161): compute_graph.flexml_layers[129].compute_node[0][0], compute_graph.flexml_layers[129].compute_node[0][1], compute_graph.flexml_layers[129].compute_node[0][2], compute_graph.flexml_layers[129].compute_node[0][3], compute_graph.flexml_layers[129].compute_node[1][0], compute_graph.flexml_layers[129].compute_node[1][1], compute_graph.flexml_layers[129].compute_node[1][2], compute_graph.flexml_layers[129].compute_node[1][3], compute_graph.flexml_layers[129].compute_node[2][0], compute_graph.flexml_layers[129].compute_node[2][1], compute_graph.flexml_layers[129].compute_node[2][2], compute_graph.flexml_layers[129].compute_node[2][3], compute_graph.flexml_layers[129].compute_node[3][0], compute_graph.flexml_layers[129].compute_node[3][1], compute_graph.flexml_layers[129].compute_node[3][2], compute_graph.flexml_layers[129].compute_node[3][3], {compute_graph.l2_146.out[0], compute_graph.l2_146.out[1], compute_graph.l2_146.out[2], compute_graph.l2_146.out[3], compute_graph.l2_147.in[0], compute_graph.l2_147.in[1], compute_graph.l2_147.in[2], compute_graph.l2_147.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1458), layer(162): compute_graph.flexml_layers[130].compute_node[0][0], compute_graph.flexml_layers[130].compute_node[0][1], compute_graph.flexml_layers[130].compute_node[0][2], compute_graph.flexml_layers[130].compute_node[0][3], compute_graph.flexml_layers[130].compute_node[1][0], compute_graph.flexml_layers[130].compute_node[1][1], compute_graph.flexml_layers[130].compute_node[1][2], compute_graph.flexml_layers[130].compute_node[1][3], compute_graph.flexml_layers[130].compute_node[2][0], compute_graph.flexml_layers[130].compute_node[2][1], compute_graph.flexml_layers[130].compute_node[2][2], compute_graph.flexml_layers[130].compute_node[2][3], compute_graph.flexml_layers[130].compute_node[3][0], compute_graph.flexml_layers[130].compute_node[3][1], compute_graph.flexml_layers[130].compute_node[3][2], compute_graph.flexml_layers[130].compute_node[3][3], {compute_graph.l2_144.out[4], compute_graph.l2_144.out[5], compute_graph.l2_144.out[6], compute_graph.l2_144.out[7], compute_graph.l2_147.out[0], compute_graph.l2_147.out[1], compute_graph.l2_147.out[2], compute_graph.l2_147.out[3], compute_graph.l2_148.in[0], compute_graph.l2_148.in[1], compute_graph.l2_148.in[2], compute_graph.l2_148.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(866): mode(3), layer(163): {compute_graph.Layer_149_wts_ddr.out[0], compute_graph.Layer_149_wts_ddr.out[1], compute_graph.Layer_149_l2_wts.in[0], compute_graph.Layer_149_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(866) is pipelined with KernelLayerNode(1458) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1459), layer(163): compute_graph.flexml_layers[131].compute_node[0][0], compute_graph.flexml_layers[131].compute_node[0][1], compute_graph.flexml_layers[131].compute_node[0][2], compute_graph.flexml_layers[131].compute_node[0][3], compute_graph.flexml_layers[131].compute_node[1][0], compute_graph.flexml_layers[131].compute_node[1][1], compute_graph.flexml_layers[131].compute_node[1][2], compute_graph.flexml_layers[131].compute_node[1][3], compute_graph.flexml_layers[131].compute_node[2][0], compute_graph.flexml_layers[131].compute_node[2][1], compute_graph.flexml_layers[131].compute_node[2][2], compute_graph.flexml_layers[131].compute_node[2][3], compute_graph.flexml_layers[131].compute_node[3][0], compute_graph.flexml_layers[131].compute_node[3][1], compute_graph.flexml_layers[131].compute_node[3][2], compute_graph.flexml_layers[131].compute_node[3][3], {compute_graph.l2_148.out[0], compute_graph.l2_148.out[1], compute_graph.l2_148.out[2], compute_graph.l2_148.out[3], compute_graph.Layer_149_l2_wts.out[0], compute_graph.Layer_149_l2_wts.out[1], compute_graph.Layer_149_l2_wts.out[2], compute_graph.Layer_149_l2_wts.out[3], compute_graph.l2_149.in[0], compute_graph.l2_149.in[1], compute_graph.l2_149.in[2], compute_graph.l2_149.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1460), layer(164): compute_graph.flexml_layers[132].compute_node[0][0], compute_graph.flexml_layers[132].compute_node[0][1], compute_graph.flexml_layers[132].compute_node[0][2], compute_graph.flexml_layers[132].compute_node[0][3], compute_graph.flexml_layers[132].compute_node[1][0], compute_graph.flexml_layers[132].compute_node[1][1], compute_graph.flexml_layers[132].compute_node[1][2], compute_graph.flexml_layers[132].compute_node[1][3], compute_graph.flexml_layers[132].compute_node[2][0], compute_graph.flexml_layers[132].compute_node[2][1], compute_graph.flexml_layers[132].compute_node[2][2], compute_graph.flexml_layers[132].compute_node[2][3], compute_graph.flexml_layers[132].compute_node[3][0], compute_graph.flexml_layers[132].compute_node[3][1], compute_graph.flexml_layers[132].compute_node[3][2], compute_graph.flexml_layers[132].compute_node[3][3], {compute_graph.l2_149.out[0], compute_graph.l2_149.out[1], compute_graph.l2_149.out[2], compute_graph.l2_149.out[3], compute_graph.l2_150.in[0], compute_graph.l2_150.in[1], compute_graph.l2_150.in[2], compute_graph.l2_150.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1461), layer(165): compute_graph.flexml_layers[133].compute_node[0][0], compute_graph.flexml_layers[133].compute_node[0][1], compute_graph.flexml_layers[133].compute_node[0][2], compute_graph.flexml_layers[133].compute_node[0][3], compute_graph.flexml_layers[133].compute_node[1][0], compute_graph.flexml_layers[133].compute_node[1][1], compute_graph.flexml_layers[133].compute_node[1][2], compute_graph.flexml_layers[133].compute_node[1][3], compute_graph.flexml_layers[133].compute_node[2][0], compute_graph.flexml_layers[133].compute_node[2][1], compute_graph.flexml_layers[133].compute_node[2][2], compute_graph.flexml_layers[133].compute_node[2][3], compute_graph.flexml_layers[133].compute_node[3][0], compute_graph.flexml_layers[133].compute_node[3][1], compute_graph.flexml_layers[133].compute_node[3][2], compute_graph.flexml_layers[133].compute_node[3][3], {compute_graph.l2_150.out[0], compute_graph.l2_150.out[1], compute_graph.l2_150.out[2], compute_graph.l2_150.out[3], compute_graph.l2_151.in[0], compute_graph.l2_151.in[1], compute_graph.l2_151.in[2], compute_graph.l2_151.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1462), layer(166): compute_graph.flexml_layers[134].compute_node[0][0], compute_graph.flexml_layers[134].compute_node[0][1], compute_graph.flexml_layers[134].compute_node[0][2], compute_graph.flexml_layers[134].compute_node[0][3], compute_graph.flexml_layers[134].compute_node[1][0], compute_graph.flexml_layers[134].compute_node[1][1], compute_graph.flexml_layers[134].compute_node[1][2], compute_graph.flexml_layers[134].compute_node[1][3], compute_graph.flexml_layers[134].compute_node[2][0], compute_graph.flexml_layers[134].compute_node[2][1], compute_graph.flexml_layers[134].compute_node[2][2], compute_graph.flexml_layers[134].compute_node[2][3], compute_graph.flexml_layers[134].compute_node[3][0], compute_graph.flexml_layers[134].compute_node[3][1], compute_graph.flexml_layers[134].compute_node[3][2], compute_graph.flexml_layers[134].compute_node[3][3], {compute_graph.l2_151.out[0], compute_graph.l2_151.out[1], compute_graph.l2_151.out[2], compute_graph.l2_151.out[3], compute_graph.l2_152.in[0], compute_graph.l2_152.in[1], compute_graph.l2_152.in[2], compute_graph.l2_152.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1463), layer(167): compute_graph.flexml_layers[135].compute_node[0][0], compute_graph.flexml_layers[135].compute_node[0][1], compute_graph.flexml_layers[135].compute_node[0][2], compute_graph.flexml_layers[135].compute_node[0][3], compute_graph.flexml_layers[135].compute_node[1][0], compute_graph.flexml_layers[135].compute_node[1][1], compute_graph.flexml_layers[135].compute_node[1][2], compute_graph.flexml_layers[135].compute_node[1][3], compute_graph.flexml_layers[135].compute_node[2][0], compute_graph.flexml_layers[135].compute_node[2][1], compute_graph.flexml_layers[135].compute_node[2][2], compute_graph.flexml_layers[135].compute_node[2][3], compute_graph.flexml_layers[135].compute_node[3][0], compute_graph.flexml_layers[135].compute_node[3][1], compute_graph.flexml_layers[135].compute_node[3][2], compute_graph.flexml_layers[135].compute_node[3][3], {compute_graph.l2_149.out[4], compute_graph.l2_149.out[5], compute_graph.l2_149.out[6], compute_graph.l2_149.out[7], compute_graph.l2_152.out[0], compute_graph.l2_152.out[1], compute_graph.l2_152.out[2], compute_graph.l2_152.out[3], compute_graph.l2_153.in[0], compute_graph.l2_153.in[1], compute_graph.l2_153.in[2], compute_graph.l2_153.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(937): mode(0), layer(167): {compute_graph.l2_153.out[2], compute_graph.l2_153.out[3], compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(936): mode(0), layer(167): {compute_graph.l2_153.out[0], compute_graph.l2_153.out[1], compute_graph.l2l3_153_spill.in[0], compute_graph.l2l3_153_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1089): mode(0), layer(168): {compute_graph.l2l3_153_spill.out[0], compute_graph.templated_graph_154.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1268): mode(0), layer(168): {compute_graph.templated_graph_154.trans_mt_ifm.out[0], compute_graph.templated_graph_154.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1269): mode(0), layer(168): {compute_graph.templated_graph_154.trans_mt_ofm.out[0], compute_graph.l2l3_154_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1090): mode(0), layer(169): {compute_graph.l2l3_154_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1464), layer(169): compute_graph.flexml_layers[136].compute_node[0][0], compute_graph.flexml_layers[136].compute_node[0][1], compute_graph.flexml_layers[136].compute_node[0][2], compute_graph.flexml_layers[136].compute_node[0][3], compute_graph.flexml_layers[136].compute_node[1][0], compute_graph.flexml_layers[136].compute_node[1][1], compute_graph.flexml_layers[136].compute_node[1][2], compute_graph.flexml_layers[136].compute_node[1][3], compute_graph.flexml_layers[136].compute_node[2][0], compute_graph.flexml_layers[136].compute_node[2][1], compute_graph.flexml_layers[136].compute_node[2][2], compute_graph.flexml_layers[136].compute_node[2][3], compute_graph.flexml_layers[136].compute_node[3][0], compute_graph.flexml_layers[136].compute_node[3][1], compute_graph.flexml_layers[136].compute_node[3][2], compute_graph.flexml_layers[136].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.out[3], compute_graph.l2_155.in[0], compute_graph.l2_155.in[1], compute_graph.l2_155.in[2], compute_graph.l2_155.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(867): mode(3), layer(170): {compute_graph.Layer_156_wts_ddr.out[0], compute_graph.Layer_156_wts_ddr.out[1], compute_graph.Layer_156_l2_wts.in[0], compute_graph.Layer_156_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(867) is pipelined with KernelLayerNode(1464) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1465), layer(170): compute_graph.flexml_layers[137].compute_node[0][0], compute_graph.flexml_layers[137].compute_node[0][1], compute_graph.flexml_layers[137].compute_node[0][2], compute_graph.flexml_layers[137].compute_node[0][3], compute_graph.flexml_layers[137].compute_node[1][0], compute_graph.flexml_layers[137].compute_node[1][1], compute_graph.flexml_layers[137].compute_node[1][2], compute_graph.flexml_layers[137].compute_node[1][3], compute_graph.flexml_layers[137].compute_node[2][0], compute_graph.flexml_layers[137].compute_node[2][1], compute_graph.flexml_layers[137].compute_node[2][2], compute_graph.flexml_layers[137].compute_node[2][3], compute_graph.flexml_layers[137].compute_node[3][0], compute_graph.flexml_layers[137].compute_node[3][1], compute_graph.flexml_layers[137].compute_node[3][2], compute_graph.flexml_layers[137].compute_node[3][3], {compute_graph.l2_155.out[0], compute_graph.l2_155.out[1], compute_graph.l2_155.out[2], compute_graph.l2_155.out[3], compute_graph.Layer_156_l2_wts.out[0], compute_graph.Layer_156_l2_wts.out[1], compute_graph.Layer_156_l2_wts.out[2], compute_graph.Layer_156_l2_wts.out[3], compute_graph.l2_156.in[0], compute_graph.l2_156.in[1], compute_graph.l2_156.in[2], compute_graph.l2_156.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(868): mode(3), layer(171): {compute_graph.Layer_157_wts_ddr.out[0], compute_graph.Layer_157_wts_ddr.out[1], compute_graph.Layer_157_l2_wts.in[0], compute_graph.Layer_157_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(868) is pipelined with KernelLayerNode(1465) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1466), layer(171): compute_graph.flexml_layers[138].compute_node[0][0], compute_graph.flexml_layers[138].compute_node[0][1], compute_graph.flexml_layers[138].compute_node[0][2], compute_graph.flexml_layers[138].compute_node[0][3], compute_graph.flexml_layers[138].compute_node[1][0], compute_graph.flexml_layers[138].compute_node[1][1], compute_graph.flexml_layers[138].compute_node[1][2], compute_graph.flexml_layers[138].compute_node[1][3], compute_graph.flexml_layers[138].compute_node[2][0], compute_graph.flexml_layers[138].compute_node[2][1], compute_graph.flexml_layers[138].compute_node[2][2], compute_graph.flexml_layers[138].compute_node[2][3], compute_graph.flexml_layers[138].compute_node[3][0], compute_graph.flexml_layers[138].compute_node[3][1], compute_graph.flexml_layers[138].compute_node[3][2], compute_graph.flexml_layers[138].compute_node[3][3], {compute_graph.l2_156.out[0], compute_graph.l2_156.out[1], compute_graph.l2_156.out[2], compute_graph.l2_156.out[3], compute_graph.Layer_157_l2_wts.out[0], compute_graph.Layer_157_l2_wts.out[1], compute_graph.Layer_157_l2_wts.out[2], compute_graph.Layer_157_l2_wts.out[3], compute_graph.l2_157.in[0], compute_graph.l2_157.in[1], compute_graph.l2_157.in[2], compute_graph.l2_157.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1467), layer(172): compute_graph.flexml_layers[139].compute_node[0][0], compute_graph.flexml_layers[139].compute_node[0][1], compute_graph.flexml_layers[139].compute_node[0][2], compute_graph.flexml_layers[139].compute_node[0][3], compute_graph.flexml_layers[139].compute_node[1][0], compute_graph.flexml_layers[139].compute_node[1][1], compute_graph.flexml_layers[139].compute_node[1][2], compute_graph.flexml_layers[139].compute_node[1][3], compute_graph.flexml_layers[139].compute_node[2][0], compute_graph.flexml_layers[139].compute_node[2][1], compute_graph.flexml_layers[139].compute_node[2][2], compute_graph.flexml_layers[139].compute_node[2][3], compute_graph.flexml_layers[139].compute_node[3][0], compute_graph.flexml_layers[139].compute_node[3][1], compute_graph.flexml_layers[139].compute_node[3][2], compute_graph.flexml_layers[139].compute_node[3][3], {compute_graph.l2_157.out[0], compute_graph.l2_157.out[1], compute_graph.l2_157.out[2], compute_graph.l2_157.out[3], compute_graph.l2_158.in[0], compute_graph.l2_158.in[1], compute_graph.l2_158.in[2], compute_graph.l2_158.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1468), layer(173): compute_graph.flexml_layers[140].compute_node[0][0], compute_graph.flexml_layers[140].compute_node[0][1], compute_graph.flexml_layers[140].compute_node[0][2], compute_graph.flexml_layers[140].compute_node[0][3], compute_graph.flexml_layers[140].compute_node[1][0], compute_graph.flexml_layers[140].compute_node[1][1], compute_graph.flexml_layers[140].compute_node[1][2], compute_graph.flexml_layers[140].compute_node[1][3], compute_graph.flexml_layers[140].compute_node[2][0], compute_graph.flexml_layers[140].compute_node[2][1], compute_graph.flexml_layers[140].compute_node[2][2], compute_graph.flexml_layers[140].compute_node[2][3], compute_graph.flexml_layers[140].compute_node[3][0], compute_graph.flexml_layers[140].compute_node[3][1], compute_graph.flexml_layers[140].compute_node[3][2], compute_graph.flexml_layers[140].compute_node[3][3], {compute_graph.l2_158.out[0], compute_graph.l2_158.out[1], compute_graph.l2_158.out[2], compute_graph.l2_158.out[3], compute_graph.l2_159.in[0], compute_graph.l2_159.in[1], compute_graph.l2_159.in[2], compute_graph.l2_159.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1469), layer(174): compute_graph.flexml_layers[141].compute_node[0][0], compute_graph.flexml_layers[141].compute_node[0][1], compute_graph.flexml_layers[141].compute_node[0][2], compute_graph.flexml_layers[141].compute_node[0][3], compute_graph.flexml_layers[141].compute_node[1][0], compute_graph.flexml_layers[141].compute_node[1][1], compute_graph.flexml_layers[141].compute_node[1][2], compute_graph.flexml_layers[141].compute_node[1][3], compute_graph.flexml_layers[141].compute_node[2][0], compute_graph.flexml_layers[141].compute_node[2][1], compute_graph.flexml_layers[141].compute_node[2][2], compute_graph.flexml_layers[141].compute_node[2][3], compute_graph.flexml_layers[141].compute_node[3][0], compute_graph.flexml_layers[141].compute_node[3][1], compute_graph.flexml_layers[141].compute_node[3][2], compute_graph.flexml_layers[141].compute_node[3][3], {compute_graph.l2_159.out[0], compute_graph.l2_159.out[1], compute_graph.l2_159.out[2], compute_graph.l2_159.out[3], compute_graph.l2_160.in[0], compute_graph.l2_160.in[1], compute_graph.l2_160.in[2], compute_graph.l2_160.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(938): mode(0), layer(174): {compute_graph.l2_160.out[1], compute_graph.l2l3_160_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1091): mode(0), layer(175): {compute_graph.l2l3_160_spill.out[0], compute_graph.templated_graph_161.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1270): mode(0), layer(175): {compute_graph.templated_graph_161.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_161_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1093): mode(0), layer(176): {compute_graph.l2l3_scratch_0_161_spill.out[0], compute_graph.templated_graph_161.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1271): mode(0), layer(176): {compute_graph.templated_graph_161.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_161_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1092): mode(0), layer(177): {compute_graph.l2l3_scratch_1_161_spill.out[0], compute_graph.templated_graph_161.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1272): mode(0), layer(177): {compute_graph.templated_graph_161.g2.ifm_mem.out[0], compute_graph.l2l3_161_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1088): mode(0), layer(178): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1094): mode(0), layer(178): {compute_graph.l2l3_161_spill.out[0], compute_graph.l2l3_161_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1470), layer(178): compute_graph.flexml_layers[142].compute_node[0][0], compute_graph.flexml_layers[142].compute_node[0][1], compute_graph.flexml_layers[142].compute_node[0][2], compute_graph.flexml_layers[142].compute_node[0][3], compute_graph.flexml_layers[142].compute_node[1][0], compute_graph.flexml_layers[142].compute_node[1][1], compute_graph.flexml_layers[142].compute_node[1][2], compute_graph.flexml_layers[142].compute_node[1][3], compute_graph.flexml_layers[142].compute_node[2][0], compute_graph.flexml_layers[142].compute_node[2][1], compute_graph.flexml_layers[142].compute_node[2][2], compute_graph.flexml_layers[142].compute_node[2][3], compute_graph.flexml_layers[142].compute_node[3][0], compute_graph.flexml_layers[142].compute_node[3][1], compute_graph.flexml_layers[142].compute_node[3][2], compute_graph.flexml_layers[142].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.out[3], compute_graph.l2_162.in[0], compute_graph.l2_162.in[1], compute_graph.l2_162.in[2], compute_graph.l2_162.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(869): mode(3), layer(179): {compute_graph.Layer_163_wts_ddr.out[0], compute_graph.Layer_163_wts_ddr.out[1], compute_graph.Layer_163_l2_wts.in[0], compute_graph.Layer_163_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(869) is pipelined with KernelLayerNode(1470) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1471), layer(179): compute_graph.flexml_layers[143].compute_node[0][0], compute_graph.flexml_layers[143].compute_node[0][1], compute_graph.flexml_layers[143].compute_node[0][2], compute_graph.flexml_layers[143].compute_node[0][3], compute_graph.flexml_layers[143].compute_node[1][0], compute_graph.flexml_layers[143].compute_node[1][1], compute_graph.flexml_layers[143].compute_node[1][2], compute_graph.flexml_layers[143].compute_node[1][3], compute_graph.flexml_layers[143].compute_node[2][0], compute_graph.flexml_layers[143].compute_node[2][1], compute_graph.flexml_layers[143].compute_node[2][2], compute_graph.flexml_layers[143].compute_node[2][3], compute_graph.flexml_layers[143].compute_node[3][0], compute_graph.flexml_layers[143].compute_node[3][1], compute_graph.flexml_layers[143].compute_node[3][2], compute_graph.flexml_layers[143].compute_node[3][3], {compute_graph.Layer_163_l2_wts.out[0], compute_graph.Layer_163_l2_wts.out[1], compute_graph.Layer_163_l2_wts.out[2], compute_graph.Layer_163_l2_wts.out[3], compute_graph.l2_162.out[0], compute_graph.l2_162.out[1], compute_graph.l2_162.out[2], compute_graph.l2_162.out[3], compute_graph.l2_163.in[0], compute_graph.l2_163.in[1], compute_graph.l2_163.in[2], compute_graph.l2_163.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(870): mode(0), layer(180): {compute_graph.Layer_164_wts_ddr.out[0], compute_graph.Layer_164_wts_ddr.out[1], compute_graph.Layer_164_l2_wts.in[0], compute_graph.Layer_164_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.Layer_164_l2_wts.out[0], compute_graph.Layer_164_l2_wts.out[1], compute_graph.Layer_164_l2_wts.out[2], compute_graph.Layer_164_l2_wts.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1472), layer(180): compute_graph.flexml_layers[144].compute_node[0][0], compute_graph.flexml_layers[144].compute_node[0][1], compute_graph.flexml_layers[144].compute_node[0][2], compute_graph.flexml_layers[144].compute_node[0][3], compute_graph.flexml_layers[144].compute_node[1][0], compute_graph.flexml_layers[144].compute_node[1][1], compute_graph.flexml_layers[144].compute_node[1][2], compute_graph.flexml_layers[144].compute_node[1][3], compute_graph.flexml_layers[144].compute_node[2][0], compute_graph.flexml_layers[144].compute_node[2][1], compute_graph.flexml_layers[144].compute_node[2][2], compute_graph.flexml_layers[144].compute_node[2][3], compute_graph.flexml_layers[144].compute_node[3][0], compute_graph.flexml_layers[144].compute_node[3][1], compute_graph.flexml_layers[144].compute_node[3][2], compute_graph.flexml_layers[144].compute_node[3][3], {compute_graph.Layer_164_l2_wts.out[0], compute_graph.Layer_164_l2_wts.out[1], compute_graph.Layer_164_l2_wts.out[2], compute_graph.Layer_164_l2_wts.out[3], compute_graph.l2_163.out[0], compute_graph.l2_163.out[1], compute_graph.l2_163.out[2], compute_graph.l2_163.out[3], compute_graph.l2_164.in[0], compute_graph.l2_164.in[1], compute_graph.l2_164.in[2], compute_graph.l2_164.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(939): mode(3), layer(180): {compute_graph.l2_163.out[4], compute_graph.l2_163.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(939) is pipelined with KernelLayerNode(1472) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1473), layer(181): compute_graph.flexml_layers[145].compute_node[0][0], compute_graph.flexml_layers[145].compute_node[0][1], compute_graph.flexml_layers[145].compute_node[0][2], compute_graph.flexml_layers[145].compute_node[0][3], compute_graph.flexml_layers[145].compute_node[1][0], compute_graph.flexml_layers[145].compute_node[1][1], compute_graph.flexml_layers[145].compute_node[1][2], compute_graph.flexml_layers[145].compute_node[1][3], compute_graph.flexml_layers[145].compute_node[2][0], compute_graph.flexml_layers[145].compute_node[2][1], compute_graph.flexml_layers[145].compute_node[2][2], compute_graph.flexml_layers[145].compute_node[2][3], compute_graph.flexml_layers[145].compute_node[3][0], compute_graph.flexml_layers[145].compute_node[3][1], compute_graph.flexml_layers[145].compute_node[3][2], compute_graph.flexml_layers[145].compute_node[3][3], {compute_graph.l2_164.out[0], compute_graph.l2_164.out[1], compute_graph.l2_164.out[2], compute_graph.l2_164.out[3], compute_graph.l2_165.in[0], compute_graph.l2_165.in[1], compute_graph.l2_165.in[2], compute_graph.l2_165.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(940): mode(3), layer(181): {compute_graph.l2_164.out[4], compute_graph.l2_164.out[5], compute_graph.l2l3_164_spill.in[0], compute_graph.l2l3_164_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(940) is pipelined with KernelLayerNode(1473) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1474), layer(182): compute_graph.flexml_layers[146].compute_node[0][0], compute_graph.flexml_layers[146].compute_node[0][1], compute_graph.flexml_layers[146].compute_node[0][2], compute_graph.flexml_layers[146].compute_node[0][3], compute_graph.flexml_layers[146].compute_node[1][0], compute_graph.flexml_layers[146].compute_node[1][1], compute_graph.flexml_layers[146].compute_node[1][2], compute_graph.flexml_layers[146].compute_node[1][3], compute_graph.flexml_layers[146].compute_node[2][0], compute_graph.flexml_layers[146].compute_node[2][1], compute_graph.flexml_layers[146].compute_node[2][2], compute_graph.flexml_layers[146].compute_node[2][3], compute_graph.flexml_layers[146].compute_node[3][0], compute_graph.flexml_layers[146].compute_node[3][1], compute_graph.flexml_layers[146].compute_node[3][2], compute_graph.flexml_layers[146].compute_node[3][3], {compute_graph.l2_165.out[0], compute_graph.l2_165.out[1], compute_graph.l2_165.out[2], compute_graph.l2_165.out[3], compute_graph.l2_166.in[0], compute_graph.l2_166.in[1], compute_graph.l2_166.in[2], compute_graph.l2_166.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1475), layer(183): compute_graph.flexml_layers[147].compute_node[0][0], compute_graph.flexml_layers[147].compute_node[0][1], compute_graph.flexml_layers[147].compute_node[0][2], compute_graph.flexml_layers[147].compute_node[0][3], compute_graph.flexml_layers[147].compute_node[1][0], compute_graph.flexml_layers[147].compute_node[1][1], compute_graph.flexml_layers[147].compute_node[1][2], compute_graph.flexml_layers[147].compute_node[1][3], compute_graph.flexml_layers[147].compute_node[2][0], compute_graph.flexml_layers[147].compute_node[2][1], compute_graph.flexml_layers[147].compute_node[2][2], compute_graph.flexml_layers[147].compute_node[2][3], compute_graph.flexml_layers[147].compute_node[3][0], compute_graph.flexml_layers[147].compute_node[3][1], compute_graph.flexml_layers[147].compute_node[3][2], compute_graph.flexml_layers[147].compute_node[3][3], {compute_graph.l2_166.out[0], compute_graph.l2_166.out[1], compute_graph.l2_166.out[2], compute_graph.l2_166.out[3], compute_graph.l2_167.in[0], compute_graph.l2_167.in[1], compute_graph.l2_167.in[2], compute_graph.l2_167.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(941): mode(0), layer(183): {compute_graph.l2_167.out[0], compute_graph.l2_167.out[1], compute_graph.l2l3_167_spill.in[0], compute_graph.l2l3_167_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 942 will not be pipelined because it's in the same layer 184 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1589), layer(184): {compute_graph.l2l3_164_spill.out[0], compute_graph.l2l3_167_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1590), layer(184): {compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1591), layer(184): {compute_graph.l2l3_164_spill.out[1], compute_graph.l2l3_167_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1592), layer(184): {compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1476), layer(184): compute_graph.flexml_layers[148].compute_node[0][0], compute_graph.flexml_layers[148].compute_node[0][1], compute_graph.flexml_layers[148].compute_node[0][2], compute_graph.flexml_layers[148].compute_node[0][3], compute_graph.flexml_layers[148].compute_node[1][0], compute_graph.flexml_layers[148].compute_node[1][1], compute_graph.flexml_layers[148].compute_node[1][2], compute_graph.flexml_layers[148].compute_node[1][3], compute_graph.flexml_layers[148].compute_node[2][0], compute_graph.flexml_layers[148].compute_node[2][1], compute_graph.flexml_layers[148].compute_node[2][2], compute_graph.flexml_layers[148].compute_node[2][3], compute_graph.flexml_layers[148].compute_node[3][0], compute_graph.flexml_layers[148].compute_node[3][1], compute_graph.flexml_layers[148].compute_node[3][2], compute_graph.flexml_layers[148].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.out[3], compute_graph.l2_168.in[0], compute_graph.l2_168.in[1], compute_graph.l2_168.in[2], compute_graph.l2_168.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(871): mode(3), layer(185): {compute_graph.Layer_169_wts_ddr.out[0], compute_graph.Layer_169_wts_ddr.out[1], compute_graph.Layer_169_l2_wts.in[0], compute_graph.Layer_169_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(871) is pipelined with KernelLayerNode(1476) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(942): mode(3), layer(184): {compute_graph.l2_168.out[0], compute_graph.l2_168.out[1], compute_graph.l2l3_168_spill.in[0], compute_graph.l2l3_168_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1098): mode(0), layer(185): {compute_graph.l2l3_168_spill.out[0], compute_graph.l2l3_168_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1477), layer(185): compute_graph.flexml_layers[149].compute_node[0][0], compute_graph.flexml_layers[149].compute_node[0][1], compute_graph.flexml_layers[149].compute_node[0][2], compute_graph.flexml_layers[149].compute_node[0][3], compute_graph.flexml_layers[149].compute_node[1][0], compute_graph.flexml_layers[149].compute_node[1][1], compute_graph.flexml_layers[149].compute_node[1][2], compute_graph.flexml_layers[149].compute_node[1][3], compute_graph.flexml_layers[149].compute_node[2][0], compute_graph.flexml_layers[149].compute_node[2][1], compute_graph.flexml_layers[149].compute_node[2][2], compute_graph.flexml_layers[149].compute_node[2][3], compute_graph.flexml_layers[149].compute_node[3][0], compute_graph.flexml_layers[149].compute_node[3][1], compute_graph.flexml_layers[149].compute_node[3][2], compute_graph.flexml_layers[149].compute_node[3][3], {compute_graph.Layer_169_l2_wts.out[0], compute_graph.Layer_169_l2_wts.out[1], compute_graph.Layer_169_l2_wts.out[2], compute_graph.Layer_169_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.out[3], compute_graph.l2_169.in[0], compute_graph.l2_169.in[1], compute_graph.l2_169.in[2], compute_graph.l2_169.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1478), layer(186): compute_graph.flexml_layers[150].compute_node[0][0], compute_graph.flexml_layers[150].compute_node[0][1], compute_graph.flexml_layers[150].compute_node[0][2], compute_graph.flexml_layers[150].compute_node[0][3], compute_graph.flexml_layers[150].compute_node[1][0], compute_graph.flexml_layers[150].compute_node[1][1], compute_graph.flexml_layers[150].compute_node[1][2], compute_graph.flexml_layers[150].compute_node[1][3], compute_graph.flexml_layers[150].compute_node[2][0], compute_graph.flexml_layers[150].compute_node[2][1], compute_graph.flexml_layers[150].compute_node[2][2], compute_graph.flexml_layers[150].compute_node[2][3], compute_graph.flexml_layers[150].compute_node[3][0], compute_graph.flexml_layers[150].compute_node[3][1], compute_graph.flexml_layers[150].compute_node[3][2], compute_graph.flexml_layers[150].compute_node[3][3], {compute_graph.l2_169.out[0], compute_graph.l2_169.out[1], compute_graph.l2_169.out[2], compute_graph.l2_169.out[3], compute_graph.l2_170.in[0], compute_graph.l2_170.in[1], compute_graph.l2_170.in[2], compute_graph.l2_170.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(943): mode(3), layer(186): {compute_graph.l2_169.out[4], compute_graph.l2_169.out[5], compute_graph.l2l3_169_spill.in[0], compute_graph.l2l3_169_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(943) is pipelined with KernelLayerNode(1478) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1479), layer(187): compute_graph.flexml_layers[151].compute_node[0][0], compute_graph.flexml_layers[151].compute_node[0][1], compute_graph.flexml_layers[151].compute_node[0][2], compute_graph.flexml_layers[151].compute_node[0][3], compute_graph.flexml_layers[151].compute_node[1][0], compute_graph.flexml_layers[151].compute_node[1][1], compute_graph.flexml_layers[151].compute_node[1][2], compute_graph.flexml_layers[151].compute_node[1][3], compute_graph.flexml_layers[151].compute_node[2][0], compute_graph.flexml_layers[151].compute_node[2][1], compute_graph.flexml_layers[151].compute_node[2][2], compute_graph.flexml_layers[151].compute_node[2][3], compute_graph.flexml_layers[151].compute_node[3][0], compute_graph.flexml_layers[151].compute_node[3][1], compute_graph.flexml_layers[151].compute_node[3][2], compute_graph.flexml_layers[151].compute_node[3][3], {compute_graph.l2_170.out[0], compute_graph.l2_170.out[1], compute_graph.l2_170.out[2], compute_graph.l2_170.out[3], compute_graph.l2_171.in[0], compute_graph.l2_171.in[1], compute_graph.l2_171.in[2], compute_graph.l2_171.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1480), layer(188): compute_graph.flexml_layers[152].compute_node[0][0], compute_graph.flexml_layers[152].compute_node[0][1], compute_graph.flexml_layers[152].compute_node[0][2], compute_graph.flexml_layers[152].compute_node[0][3], compute_graph.flexml_layers[152].compute_node[1][0], compute_graph.flexml_layers[152].compute_node[1][1], compute_graph.flexml_layers[152].compute_node[1][2], compute_graph.flexml_layers[152].compute_node[1][3], compute_graph.flexml_layers[152].compute_node[2][0], compute_graph.flexml_layers[152].compute_node[2][1], compute_graph.flexml_layers[152].compute_node[2][2], compute_graph.flexml_layers[152].compute_node[2][3], compute_graph.flexml_layers[152].compute_node[3][0], compute_graph.flexml_layers[152].compute_node[3][1], compute_graph.flexml_layers[152].compute_node[3][2], compute_graph.flexml_layers[152].compute_node[3][3], {compute_graph.l2_171.out[0], compute_graph.l2_171.out[1], compute_graph.l2_171.out[2], compute_graph.l2_171.out[3], compute_graph.l2_172.in[0], compute_graph.l2_172.in[1], compute_graph.l2_172.in[2], compute_graph.l2_172.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(944): mode(0), layer(188): {compute_graph.l2_172.out[0], compute_graph.l2_172.out[1], compute_graph.l2l3_172_spill.in[0], compute_graph.l2l3_172_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 945 will not be pipelined because it's in the same layer 189 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1593), layer(189): {compute_graph.l2l3_169_spill.out[0], compute_graph.l2l3_172_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1594), layer(189): {compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1595), layer(189): {compute_graph.l2l3_169_spill.out[1], compute_graph.l2l3_172_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1596), layer(189): {compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1481), layer(189): compute_graph.flexml_layers[153].compute_node[0][0], compute_graph.flexml_layers[153].compute_node[0][1], compute_graph.flexml_layers[153].compute_node[0][2], compute_graph.flexml_layers[153].compute_node[0][3], compute_graph.flexml_layers[153].compute_node[1][0], compute_graph.flexml_layers[153].compute_node[1][1], compute_graph.flexml_layers[153].compute_node[1][2], compute_graph.flexml_layers[153].compute_node[1][3], compute_graph.flexml_layers[153].compute_node[2][0], compute_graph.flexml_layers[153].compute_node[2][1], compute_graph.flexml_layers[153].compute_node[2][2], compute_graph.flexml_layers[153].compute_node[2][3], compute_graph.flexml_layers[153].compute_node[3][0], compute_graph.flexml_layers[153].compute_node[3][1], compute_graph.flexml_layers[153].compute_node[3][2], compute_graph.flexml_layers[153].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.out[3], compute_graph.l2_173.in[0], compute_graph.l2_173.in[1], compute_graph.l2_173.in[2], compute_graph.l2_173.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(945): mode(3), layer(189): {compute_graph.l2_173.out[0], compute_graph.l2_173.out[1], compute_graph.l2l3_173_spill.in[0], compute_graph.l2l3_173_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1101): mode(0), layer(190): {compute_graph.l2l3_173_spill.out[0], compute_graph.templated_graph_174.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1273): mode(0), layer(190): {compute_graph.templated_graph_174.trans_mt_ifm.out[0], compute_graph.templated_graph_174.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1274): mode(0), layer(190): {compute_graph.templated_graph_174.trans_mt_ofm.out[0], compute_graph.l2l3_174_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1103): mode(0), layer(191): {compute_graph.l2l3_174_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1482), layer(191): compute_graph.flexml_layers[154].compute_node[0][0], compute_graph.flexml_layers[154].compute_node[0][1], compute_graph.flexml_layers[154].compute_node[0][2], compute_graph.flexml_layers[154].compute_node[0][3], compute_graph.flexml_layers[154].compute_node[1][0], compute_graph.flexml_layers[154].compute_node[1][1], compute_graph.flexml_layers[154].compute_node[1][2], compute_graph.flexml_layers[154].compute_node[1][3], compute_graph.flexml_layers[154].compute_node[2][0], compute_graph.flexml_layers[154].compute_node[2][1], compute_graph.flexml_layers[154].compute_node[2][2], compute_graph.flexml_layers[154].compute_node[2][3], compute_graph.flexml_layers[154].compute_node[3][0], compute_graph.flexml_layers[154].compute_node[3][1], compute_graph.flexml_layers[154].compute_node[3][2], compute_graph.flexml_layers[154].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.out[3], compute_graph.l2_175.in[0], compute_graph.l2_175.in[1], compute_graph.l2_175.in[2], compute_graph.l2_175.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(872): mode(3), layer(192): {compute_graph.Layer_176_wts_ddr.out[0], compute_graph.Layer_176_wts_ddr.out[1], compute_graph.Layer_176_l2_wts.in[0], compute_graph.Layer_176_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(872) is pipelined with KernelLayerNode(1482) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1483), layer(192): compute_graph.flexml_layers[155].compute_node[0][0], compute_graph.flexml_layers[155].compute_node[0][1], compute_graph.flexml_layers[155].compute_node[0][2], compute_graph.flexml_layers[155].compute_node[0][3], compute_graph.flexml_layers[155].compute_node[1][0], compute_graph.flexml_layers[155].compute_node[1][1], compute_graph.flexml_layers[155].compute_node[1][2], compute_graph.flexml_layers[155].compute_node[1][3], compute_graph.flexml_layers[155].compute_node[2][0], compute_graph.flexml_layers[155].compute_node[2][1], compute_graph.flexml_layers[155].compute_node[2][2], compute_graph.flexml_layers[155].compute_node[2][3], compute_graph.flexml_layers[155].compute_node[3][0], compute_graph.flexml_layers[155].compute_node[3][1], compute_graph.flexml_layers[155].compute_node[3][2], compute_graph.flexml_layers[155].compute_node[3][3], {compute_graph.Layer_176_l2_wts.out[0], compute_graph.Layer_176_l2_wts.out[1], compute_graph.Layer_176_l2_wts.out[2], compute_graph.Layer_176_l2_wts.out[3], compute_graph.l2_175.out[0], compute_graph.l2_175.out[1], compute_graph.l2_175.out[2], compute_graph.l2_175.out[3], compute_graph.l2_176.in[0], compute_graph.l2_176.in[1], compute_graph.l2_176.in[2], compute_graph.l2_176.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(873): mode(0), layer(193): {compute_graph.Layer_177_wts_ddr.out[0], compute_graph.Layer_177_wts_ddr.out[1], compute_graph.Layer_177_l2_wts.in[0], compute_graph.Layer_177_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1484), layer(193): compute_graph.flexml_layers[156].compute_node[0][0], compute_graph.flexml_layers[156].compute_node[0][1], compute_graph.flexml_layers[156].compute_node[0][2], compute_graph.flexml_layers[156].compute_node[0][3], compute_graph.flexml_layers[156].compute_node[1][0], compute_graph.flexml_layers[156].compute_node[1][1], compute_graph.flexml_layers[156].compute_node[1][2], compute_graph.flexml_layers[156].compute_node[1][3], compute_graph.flexml_layers[156].compute_node[2][0], compute_graph.flexml_layers[156].compute_node[2][1], compute_graph.flexml_layers[156].compute_node[2][2], compute_graph.flexml_layers[156].compute_node[2][3], compute_graph.flexml_layers[156].compute_node[3][0], compute_graph.flexml_layers[156].compute_node[3][1], compute_graph.flexml_layers[156].compute_node[3][2], compute_graph.flexml_layers[156].compute_node[3][3], {compute_graph.Layer_177_l2_wts.out[0], compute_graph.Layer_177_l2_wts.out[1], compute_graph.Layer_177_l2_wts.out[2], compute_graph.Layer_177_l2_wts.out[3], compute_graph.l2_176.out[0], compute_graph.l2_176.out[1], compute_graph.l2_176.out[2], compute_graph.l2_176.out[3], compute_graph.l2_177.in[0], compute_graph.l2_177.in[1], compute_graph.l2_177.in[2], compute_graph.l2_177.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1485), layer(194): compute_graph.flexml_layers[157].compute_node[0][0], compute_graph.flexml_layers[157].compute_node[0][1], compute_graph.flexml_layers[157].compute_node[0][2], compute_graph.flexml_layers[157].compute_node[0][3], compute_graph.flexml_layers[157].compute_node[1][0], compute_graph.flexml_layers[157].compute_node[1][1], compute_graph.flexml_layers[157].compute_node[1][2], compute_graph.flexml_layers[157].compute_node[1][3], compute_graph.flexml_layers[157].compute_node[2][0], compute_graph.flexml_layers[157].compute_node[2][1], compute_graph.flexml_layers[157].compute_node[2][2], compute_graph.flexml_layers[157].compute_node[2][3], compute_graph.flexml_layers[157].compute_node[3][0], compute_graph.flexml_layers[157].compute_node[3][1], compute_graph.flexml_layers[157].compute_node[3][2], compute_graph.flexml_layers[157].compute_node[3][3], {compute_graph.l2_177.out[0], compute_graph.l2_177.out[1], compute_graph.l2_177.out[2], compute_graph.l2_177.out[3], compute_graph.l2_178.in[0], compute_graph.l2_178.in[1], compute_graph.l2_178.in[2], compute_graph.l2_178.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1486), layer(195): compute_graph.flexml_layers[158].compute_node[0][0], compute_graph.flexml_layers[158].compute_node[0][1], compute_graph.flexml_layers[158].compute_node[0][2], compute_graph.flexml_layers[158].compute_node[0][3], compute_graph.flexml_layers[158].compute_node[1][0], compute_graph.flexml_layers[158].compute_node[1][1], compute_graph.flexml_layers[158].compute_node[1][2], compute_graph.flexml_layers[158].compute_node[1][3], compute_graph.flexml_layers[158].compute_node[2][0], compute_graph.flexml_layers[158].compute_node[2][1], compute_graph.flexml_layers[158].compute_node[2][2], compute_graph.flexml_layers[158].compute_node[2][3], compute_graph.flexml_layers[158].compute_node[3][0], compute_graph.flexml_layers[158].compute_node[3][1], compute_graph.flexml_layers[158].compute_node[3][2], compute_graph.flexml_layers[158].compute_node[3][3], {compute_graph.l2_178.out[0], compute_graph.l2_178.out[1], compute_graph.l2_178.out[2], compute_graph.l2_178.out[3], compute_graph.l2_179.in[0], compute_graph.l2_179.in[1], compute_graph.l2_179.in[2], compute_graph.l2_179.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1487), layer(196): compute_graph.flexml_layers[159].compute_node[0][0], compute_graph.flexml_layers[159].compute_node[0][1], compute_graph.flexml_layers[159].compute_node[0][2], compute_graph.flexml_layers[159].compute_node[0][3], compute_graph.flexml_layers[159].compute_node[1][0], compute_graph.flexml_layers[159].compute_node[1][1], compute_graph.flexml_layers[159].compute_node[1][2], compute_graph.flexml_layers[159].compute_node[1][3], compute_graph.flexml_layers[159].compute_node[2][0], compute_graph.flexml_layers[159].compute_node[2][1], compute_graph.flexml_layers[159].compute_node[2][2], compute_graph.flexml_layers[159].compute_node[2][3], compute_graph.flexml_layers[159].compute_node[3][0], compute_graph.flexml_layers[159].compute_node[3][1], compute_graph.flexml_layers[159].compute_node[3][2], compute_graph.flexml_layers[159].compute_node[3][3], {compute_graph.l2_179.out[0], compute_graph.l2_179.out[1], compute_graph.l2_179.out[2], compute_graph.l2_179.out[3], compute_graph.l2_180.in[0], compute_graph.l2_180.in[1], compute_graph.l2_180.in[2], compute_graph.l2_180.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(946): mode(0), layer(196): {compute_graph.l2_180.out[1], compute_graph.l2l3_180_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1104): mode(0), layer(197): {compute_graph.l2l3_180_spill.out[0], compute_graph.templated_graph_181.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1275): mode(0), layer(197): {compute_graph.templated_graph_181.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_181_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1105): mode(0), layer(198): {compute_graph.l2l3_scratch_0_181_spill.out[0], compute_graph.templated_graph_181.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1276): mode(0), layer(198): {compute_graph.templated_graph_181.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_181_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1106): mode(0), layer(199): {compute_graph.l2l3_scratch_1_181_spill.out[0], compute_graph.templated_graph_181.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1277): mode(0), layer(199): {compute_graph.templated_graph_181.g2.ifm_mem.out[0], compute_graph.l2l3_181_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1102): mode(0), layer(200): {compute_graph.l2l3_173_spill.out[1], compute_graph.l2l3_173_spill.out[2], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1107): mode(0), layer(200): {compute_graph.l2l3_181_spill.out[0], compute_graph.l2l3_181_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1488), layer(200): compute_graph.flexml_layers[160].compute_node[0][0], compute_graph.flexml_layers[160].compute_node[0][1], compute_graph.flexml_layers[160].compute_node[0][2], compute_graph.flexml_layers[160].compute_node[0][3], compute_graph.flexml_layers[160].compute_node[1][0], compute_graph.flexml_layers[160].compute_node[1][1], compute_graph.flexml_layers[160].compute_node[1][2], compute_graph.flexml_layers[160].compute_node[1][3], compute_graph.flexml_layers[160].compute_node[2][0], compute_graph.flexml_layers[160].compute_node[2][1], compute_graph.flexml_layers[160].compute_node[2][2], compute_graph.flexml_layers[160].compute_node[2][3], compute_graph.flexml_layers[160].compute_node[3][0], compute_graph.flexml_layers[160].compute_node[3][1], compute_graph.flexml_layers[160].compute_node[3][2], compute_graph.flexml_layers[160].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.out[3], compute_graph.l2_182.in[0], compute_graph.l2_182.in[1], compute_graph.l2_182.in[2], compute_graph.l2_182.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(874): mode(3), layer(201): {compute_graph.Layer_183_wts_ddr.out[0], compute_graph.Layer_183_wts_ddr.out[1], compute_graph.Layer_183_l2_wts.in[0], compute_graph.Layer_183_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(874) is pipelined with KernelLayerNode(1488) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163 overlaps with memory space used by compute_graph.l2_183. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1095): mode(0), layer(201): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[3], compute_graph.l2_182.out[0], compute_graph.l2_182.out[1], compute_graph.l2_182.out[2], compute_graph.l2_182.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1489), layer(201): compute_graph.flexml_layers[161].compute_node[0][0], compute_graph.flexml_layers[161].compute_node[0][1], compute_graph.flexml_layers[161].compute_node[0][2], compute_graph.flexml_layers[161].compute_node[0][3], compute_graph.flexml_layers[161].compute_node[1][0], compute_graph.flexml_layers[161].compute_node[1][1], compute_graph.flexml_layers[161].compute_node[1][2], compute_graph.flexml_layers[161].compute_node[1][3], compute_graph.flexml_layers[161].compute_node[2][0], compute_graph.flexml_layers[161].compute_node[2][1], compute_graph.flexml_layers[161].compute_node[2][2], compute_graph.flexml_layers[161].compute_node[2][3], compute_graph.flexml_layers[161].compute_node[3][0], compute_graph.flexml_layers[161].compute_node[3][1], compute_graph.flexml_layers[161].compute_node[3][2], compute_graph.flexml_layers[161].compute_node[3][3], {compute_graph.Layer_183_l2_wts.out[0], compute_graph.Layer_183_l2_wts.out[1], compute_graph.Layer_183_l2_wts.out[2], compute_graph.Layer_183_l2_wts.out[3], compute_graph.l2_182.out[0], compute_graph.l2_182.out[1], compute_graph.l2_182.out[2], compute_graph.l2_182.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.out[3], compute_graph.l2_183.in[0], compute_graph.l2_183.in[1], compute_graph.l2_183.in[2], compute_graph.l2_183.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(875): mode(0), layer(202): {compute_graph.Layer_184_wts_ddr.out[0], compute_graph.Layer_184_wts_ddr.out[1], compute_graph.Layer_184_l2_wts.in[0], compute_graph.Layer_184_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.Layer_184_l2_wts.out[0], compute_graph.Layer_184_l2_wts.out[1], compute_graph.Layer_184_l2_wts.out[2], compute_graph.Layer_184_l2_wts.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1490), layer(202): compute_graph.flexml_layers[162].compute_node[0][0], compute_graph.flexml_layers[162].compute_node[0][1], compute_graph.flexml_layers[162].compute_node[0][2], compute_graph.flexml_layers[162].compute_node[0][3], compute_graph.flexml_layers[162].compute_node[1][0], compute_graph.flexml_layers[162].compute_node[1][1], compute_graph.flexml_layers[162].compute_node[1][2], compute_graph.flexml_layers[162].compute_node[1][3], compute_graph.flexml_layers[162].compute_node[2][0], compute_graph.flexml_layers[162].compute_node[2][1], compute_graph.flexml_layers[162].compute_node[2][2], compute_graph.flexml_layers[162].compute_node[2][3], compute_graph.flexml_layers[162].compute_node[3][0], compute_graph.flexml_layers[162].compute_node[3][1], compute_graph.flexml_layers[162].compute_node[3][2], compute_graph.flexml_layers[162].compute_node[3][3], {compute_graph.Layer_184_l2_wts.out[0], compute_graph.Layer_184_l2_wts.out[1], compute_graph.Layer_184_l2_wts.out[2], compute_graph.Layer_184_l2_wts.out[3], compute_graph.l2_183.out[0], compute_graph.l2_183.out[1], compute_graph.l2_183.out[2], compute_graph.l2_183.out[3], compute_graph.l2_184.in[0], compute_graph.l2_184.in[1], compute_graph.l2_184.in[2], compute_graph.l2_184.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(947): mode(3), layer(202): {compute_graph.l2_183.out[4], compute_graph.l2_183.out[5], compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(947) is pipelined with KernelLayerNode(1490) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1491), layer(203): compute_graph.flexml_layers[163].compute_node[0][0], compute_graph.flexml_layers[163].compute_node[0][1], compute_graph.flexml_layers[163].compute_node[0][2], compute_graph.flexml_layers[163].compute_node[0][3], compute_graph.flexml_layers[163].compute_node[1][0], compute_graph.flexml_layers[163].compute_node[1][1], compute_graph.flexml_layers[163].compute_node[1][2], compute_graph.flexml_layers[163].compute_node[1][3], compute_graph.flexml_layers[163].compute_node[2][0], compute_graph.flexml_layers[163].compute_node[2][1], compute_graph.flexml_layers[163].compute_node[2][2], compute_graph.flexml_layers[163].compute_node[2][3], compute_graph.flexml_layers[163].compute_node[3][0], compute_graph.flexml_layers[163].compute_node[3][1], compute_graph.flexml_layers[163].compute_node[3][2], compute_graph.flexml_layers[163].compute_node[3][3], {compute_graph.l2_184.out[0], compute_graph.l2_184.out[1], compute_graph.l2_184.out[2], compute_graph.l2_184.out[3], compute_graph.l2_185.in[0], compute_graph.l2_185.in[1], compute_graph.l2_185.in[2], compute_graph.l2_185.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(948): mode(3), layer(203): {compute_graph.l2_184.out[4], compute_graph.l2_184.out[5], compute_graph.l2l3_184_spill.in[0], compute_graph.l2l3_184_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(948) is pipelined with KernelLayerNode(1491) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1492), layer(204): compute_graph.flexml_layers[164].compute_node[0][0], compute_graph.flexml_layers[164].compute_node[0][1], compute_graph.flexml_layers[164].compute_node[0][2], compute_graph.flexml_layers[164].compute_node[0][3], compute_graph.flexml_layers[164].compute_node[1][0], compute_graph.flexml_layers[164].compute_node[1][1], compute_graph.flexml_layers[164].compute_node[1][2], compute_graph.flexml_layers[164].compute_node[1][3], compute_graph.flexml_layers[164].compute_node[2][0], compute_graph.flexml_layers[164].compute_node[2][1], compute_graph.flexml_layers[164].compute_node[2][2], compute_graph.flexml_layers[164].compute_node[2][3], compute_graph.flexml_layers[164].compute_node[3][0], compute_graph.flexml_layers[164].compute_node[3][1], compute_graph.flexml_layers[164].compute_node[3][2], compute_graph.flexml_layers[164].compute_node[3][3], {compute_graph.l2_185.out[0], compute_graph.l2_185.out[1], compute_graph.l2_185.out[2], compute_graph.l2_185.out[3], compute_graph.l2_186.in[0], compute_graph.l2_186.in[1], compute_graph.l2_186.in[2], compute_graph.l2_186.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1493), layer(205): compute_graph.flexml_layers[165].compute_node[0][0], compute_graph.flexml_layers[165].compute_node[0][1], compute_graph.flexml_layers[165].compute_node[0][2], compute_graph.flexml_layers[165].compute_node[0][3], compute_graph.flexml_layers[165].compute_node[1][0], compute_graph.flexml_layers[165].compute_node[1][1], compute_graph.flexml_layers[165].compute_node[1][2], compute_graph.flexml_layers[165].compute_node[1][3], compute_graph.flexml_layers[165].compute_node[2][0], compute_graph.flexml_layers[165].compute_node[2][1], compute_graph.flexml_layers[165].compute_node[2][2], compute_graph.flexml_layers[165].compute_node[2][3], compute_graph.flexml_layers[165].compute_node[3][0], compute_graph.flexml_layers[165].compute_node[3][1], compute_graph.flexml_layers[165].compute_node[3][2], compute_graph.flexml_layers[165].compute_node[3][3], {compute_graph.l2_186.out[0], compute_graph.l2_186.out[1], compute_graph.l2_186.out[2], compute_graph.l2_186.out[3], compute_graph.l2_187.in[0], compute_graph.l2_187.in[1], compute_graph.l2_187.in[2], compute_graph.l2_187.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(949): mode(0), layer(205): {compute_graph.l2_187.out[0], compute_graph.l2_187.out[1], compute_graph.l2l3_187_spill.in[0], compute_graph.l2l3_187_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 950 will not be pipelined because it's in the same layer 206 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1597), layer(206): {compute_graph.l2l3_184_spill.out[0], compute_graph.l2l3_187_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1598), layer(206): {compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1599), layer(206): {compute_graph.l2l3_184_spill.out[1], compute_graph.l2l3_187_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1600), layer(206): {compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1494), layer(206): compute_graph.flexml_layers[166].compute_node[0][0], compute_graph.flexml_layers[166].compute_node[0][1], compute_graph.flexml_layers[166].compute_node[0][2], compute_graph.flexml_layers[166].compute_node[0][3], compute_graph.flexml_layers[166].compute_node[1][0], compute_graph.flexml_layers[166].compute_node[1][1], compute_graph.flexml_layers[166].compute_node[1][2], compute_graph.flexml_layers[166].compute_node[1][3], compute_graph.flexml_layers[166].compute_node[2][0], compute_graph.flexml_layers[166].compute_node[2][1], compute_graph.flexml_layers[166].compute_node[2][2], compute_graph.flexml_layers[166].compute_node[2][3], compute_graph.flexml_layers[166].compute_node[3][0], compute_graph.flexml_layers[166].compute_node[3][1], compute_graph.flexml_layers[166].compute_node[3][2], compute_graph.flexml_layers[166].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.out[3], compute_graph.l2_188.in[0], compute_graph.l2_188.in[1], compute_graph.l2_188.in[2], compute_graph.l2_188.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(876): mode(3), layer(207): {compute_graph.Layer_189_wts_ddr.out[0], compute_graph.Layer_189_wts_ddr.out[1], compute_graph.Layer_189_l2_wts.in[0], compute_graph.Layer_189_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(876) is pipelined with KernelLayerNode(1494) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(950): mode(3), layer(206): {compute_graph.l2_188.out[0], compute_graph.l2_188.out[1], compute_graph.l2l3_188_spill.in[0], compute_graph.l2l3_188_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1111): mode(0), layer(207): {compute_graph.l2l3_188_spill.out[0], compute_graph.l2l3_188_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1495), layer(207): compute_graph.flexml_layers[167].compute_node[0][0], compute_graph.flexml_layers[167].compute_node[0][1], compute_graph.flexml_layers[167].compute_node[0][2], compute_graph.flexml_layers[167].compute_node[0][3], compute_graph.flexml_layers[167].compute_node[1][0], compute_graph.flexml_layers[167].compute_node[1][1], compute_graph.flexml_layers[167].compute_node[1][2], compute_graph.flexml_layers[167].compute_node[1][3], compute_graph.flexml_layers[167].compute_node[2][0], compute_graph.flexml_layers[167].compute_node[2][1], compute_graph.flexml_layers[167].compute_node[2][2], compute_graph.flexml_layers[167].compute_node[2][3], compute_graph.flexml_layers[167].compute_node[3][0], compute_graph.flexml_layers[167].compute_node[3][1], compute_graph.flexml_layers[167].compute_node[3][2], compute_graph.flexml_layers[167].compute_node[3][3], {compute_graph.Layer_189_l2_wts.out[0], compute_graph.Layer_189_l2_wts.out[1], compute_graph.Layer_189_l2_wts.out[2], compute_graph.Layer_189_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.out[3], compute_graph.l2_189.in[0], compute_graph.l2_189.in[1], compute_graph.l2_189.in[2], compute_graph.l2_189.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1496), layer(208): compute_graph.flexml_layers[168].compute_node[0][0], compute_graph.flexml_layers[168].compute_node[0][1], compute_graph.flexml_layers[168].compute_node[0][2], compute_graph.flexml_layers[168].compute_node[0][3], compute_graph.flexml_layers[168].compute_node[1][0], compute_graph.flexml_layers[168].compute_node[1][1], compute_graph.flexml_layers[168].compute_node[1][2], compute_graph.flexml_layers[168].compute_node[1][3], compute_graph.flexml_layers[168].compute_node[2][0], compute_graph.flexml_layers[168].compute_node[2][1], compute_graph.flexml_layers[168].compute_node[2][2], compute_graph.flexml_layers[168].compute_node[2][3], compute_graph.flexml_layers[168].compute_node[3][0], compute_graph.flexml_layers[168].compute_node[3][1], compute_graph.flexml_layers[168].compute_node[3][2], compute_graph.flexml_layers[168].compute_node[3][3], {compute_graph.l2_189.out[0], compute_graph.l2_189.out[1], compute_graph.l2_189.out[2], compute_graph.l2_189.out[3], compute_graph.l2_190.in[0], compute_graph.l2_190.in[1], compute_graph.l2_190.in[2], compute_graph.l2_190.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(951): mode(3), layer(208): {compute_graph.l2_189.out[4], compute_graph.l2_189.out[5], compute_graph.l2l3_189_spill.in[0], compute_graph.l2l3_189_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(951) is pipelined with KernelLayerNode(1496) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1497), layer(209): compute_graph.flexml_layers[169].compute_node[0][0], compute_graph.flexml_layers[169].compute_node[0][1], compute_graph.flexml_layers[169].compute_node[0][2], compute_graph.flexml_layers[169].compute_node[0][3], compute_graph.flexml_layers[169].compute_node[1][0], compute_graph.flexml_layers[169].compute_node[1][1], compute_graph.flexml_layers[169].compute_node[1][2], compute_graph.flexml_layers[169].compute_node[1][3], compute_graph.flexml_layers[169].compute_node[2][0], compute_graph.flexml_layers[169].compute_node[2][1], compute_graph.flexml_layers[169].compute_node[2][2], compute_graph.flexml_layers[169].compute_node[2][3], compute_graph.flexml_layers[169].compute_node[3][0], compute_graph.flexml_layers[169].compute_node[3][1], compute_graph.flexml_layers[169].compute_node[3][2], compute_graph.flexml_layers[169].compute_node[3][3], {compute_graph.l2_190.out[0], compute_graph.l2_190.out[1], compute_graph.l2_190.out[2], compute_graph.l2_190.out[3], compute_graph.l2_191.in[0], compute_graph.l2_191.in[1], compute_graph.l2_191.in[2], compute_graph.l2_191.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1498), layer(210): compute_graph.flexml_layers[170].compute_node[0][0], compute_graph.flexml_layers[170].compute_node[0][1], compute_graph.flexml_layers[170].compute_node[0][2], compute_graph.flexml_layers[170].compute_node[0][3], compute_graph.flexml_layers[170].compute_node[1][0], compute_graph.flexml_layers[170].compute_node[1][1], compute_graph.flexml_layers[170].compute_node[1][2], compute_graph.flexml_layers[170].compute_node[1][3], compute_graph.flexml_layers[170].compute_node[2][0], compute_graph.flexml_layers[170].compute_node[2][1], compute_graph.flexml_layers[170].compute_node[2][2], compute_graph.flexml_layers[170].compute_node[2][3], compute_graph.flexml_layers[170].compute_node[3][0], compute_graph.flexml_layers[170].compute_node[3][1], compute_graph.flexml_layers[170].compute_node[3][2], compute_graph.flexml_layers[170].compute_node[3][3], {compute_graph.l2_191.out[0], compute_graph.l2_191.out[1], compute_graph.l2_191.out[2], compute_graph.l2_191.out[3], compute_graph.l2_192.in[0], compute_graph.l2_192.in[1], compute_graph.l2_192.in[2], compute_graph.l2_192.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(952): mode(0), layer(210): {compute_graph.l2_192.out[0], compute_graph.l2_192.out[1], compute_graph.l2l3_192_spill.in[0], compute_graph.l2l3_192_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 953 will not be pipelined because it's in the same layer 211 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1601), layer(211): {compute_graph.l2l3_189_spill.out[0], compute_graph.l2l3_192_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1602), layer(211): {compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1603), layer(211): {compute_graph.l2l3_189_spill.out[1], compute_graph.l2l3_192_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1604), layer(211): {compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1499), layer(211): compute_graph.flexml_layers[171].compute_node[0][0], compute_graph.flexml_layers[171].compute_node[0][1], compute_graph.flexml_layers[171].compute_node[0][2], compute_graph.flexml_layers[171].compute_node[0][3], compute_graph.flexml_layers[171].compute_node[1][0], compute_graph.flexml_layers[171].compute_node[1][1], compute_graph.flexml_layers[171].compute_node[1][2], compute_graph.flexml_layers[171].compute_node[1][3], compute_graph.flexml_layers[171].compute_node[2][0], compute_graph.flexml_layers[171].compute_node[2][1], compute_graph.flexml_layers[171].compute_node[2][2], compute_graph.flexml_layers[171].compute_node[2][3], compute_graph.flexml_layers[171].compute_node[3][0], compute_graph.flexml_layers[171].compute_node[3][1], compute_graph.flexml_layers[171].compute_node[3][2], compute_graph.flexml_layers[171].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.out[3], compute_graph.l2_193.in[0], compute_graph.l2_193.in[1], compute_graph.l2_193.in[2], compute_graph.l2_193.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(953): mode(3), layer(211): {compute_graph.l2_193.out[0], compute_graph.l2_193.out[1], compute_graph.l2l3_193_spill.in[0], compute_graph.l2l3_193_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1114): mode(0), layer(212): {compute_graph.l2l3_193_spill.out[0], compute_graph.templated_graph_194.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1278): mode(0), layer(212): {compute_graph.templated_graph_194.trans_mt_ifm.out[0], compute_graph.templated_graph_194.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1279): mode(0), layer(212): {compute_graph.templated_graph_194.trans_mt_ofm.out[0], compute_graph.l2l3_194_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1116): mode(0), layer(213): {compute_graph.l2l3_194_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1500), layer(213): compute_graph.flexml_layers[172].compute_node[0][0], compute_graph.flexml_layers[172].compute_node[0][1], compute_graph.flexml_layers[172].compute_node[0][2], compute_graph.flexml_layers[172].compute_node[0][3], compute_graph.flexml_layers[172].compute_node[1][0], compute_graph.flexml_layers[172].compute_node[1][1], compute_graph.flexml_layers[172].compute_node[1][2], compute_graph.flexml_layers[172].compute_node[1][3], compute_graph.flexml_layers[172].compute_node[2][0], compute_graph.flexml_layers[172].compute_node[2][1], compute_graph.flexml_layers[172].compute_node[2][2], compute_graph.flexml_layers[172].compute_node[2][3], compute_graph.flexml_layers[172].compute_node[3][0], compute_graph.flexml_layers[172].compute_node[3][1], compute_graph.flexml_layers[172].compute_node[3][2], compute_graph.flexml_layers[172].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.out[3], compute_graph.l2_195.in[0], compute_graph.l2_195.in[1], compute_graph.l2_195.in[2], compute_graph.l2_195.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(877): mode(3), layer(214): {compute_graph.Layer_196_wts_ddr.out[0], compute_graph.Layer_196_wts_ddr.out[1], compute_graph.Layer_196_l2_wts.in[0], compute_graph.Layer_196_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(877) is pipelined with KernelLayerNode(1500) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1501), layer(214): compute_graph.flexml_layers[173].compute_node[0][0], compute_graph.flexml_layers[173].compute_node[0][1], compute_graph.flexml_layers[173].compute_node[0][2], compute_graph.flexml_layers[173].compute_node[0][3], compute_graph.flexml_layers[173].compute_node[1][0], compute_graph.flexml_layers[173].compute_node[1][1], compute_graph.flexml_layers[173].compute_node[1][2], compute_graph.flexml_layers[173].compute_node[1][3], compute_graph.flexml_layers[173].compute_node[2][0], compute_graph.flexml_layers[173].compute_node[2][1], compute_graph.flexml_layers[173].compute_node[2][2], compute_graph.flexml_layers[173].compute_node[2][3], compute_graph.flexml_layers[173].compute_node[3][0], compute_graph.flexml_layers[173].compute_node[3][1], compute_graph.flexml_layers[173].compute_node[3][2], compute_graph.flexml_layers[173].compute_node[3][3], {compute_graph.Layer_196_l2_wts.out[0], compute_graph.Layer_196_l2_wts.out[1], compute_graph.Layer_196_l2_wts.out[2], compute_graph.Layer_196_l2_wts.out[3], compute_graph.l2_195.out[0], compute_graph.l2_195.out[1], compute_graph.l2_195.out[2], compute_graph.l2_195.out[3], compute_graph.l2_196.in[0], compute_graph.l2_196.in[1], compute_graph.l2_196.in[2], compute_graph.l2_196.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(878): mode(0), layer(215): {compute_graph.Layer_197_wts_ddr.out[0], compute_graph.Layer_197_wts_ddr.out[1], compute_graph.Layer_197_l2_wts.in[0], compute_graph.Layer_197_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1502), layer(215): compute_graph.flexml_layers[174].compute_node[0][0], compute_graph.flexml_layers[174].compute_node[0][1], compute_graph.flexml_layers[174].compute_node[0][2], compute_graph.flexml_layers[174].compute_node[0][3], compute_graph.flexml_layers[174].compute_node[1][0], compute_graph.flexml_layers[174].compute_node[1][1], compute_graph.flexml_layers[174].compute_node[1][2], compute_graph.flexml_layers[174].compute_node[1][3], compute_graph.flexml_layers[174].compute_node[2][0], compute_graph.flexml_layers[174].compute_node[2][1], compute_graph.flexml_layers[174].compute_node[2][2], compute_graph.flexml_layers[174].compute_node[2][3], compute_graph.flexml_layers[174].compute_node[3][0], compute_graph.flexml_layers[174].compute_node[3][1], compute_graph.flexml_layers[174].compute_node[3][2], compute_graph.flexml_layers[174].compute_node[3][3], {compute_graph.Layer_197_l2_wts.out[0], compute_graph.Layer_197_l2_wts.out[1], compute_graph.Layer_197_l2_wts.out[2], compute_graph.Layer_197_l2_wts.out[3], compute_graph.l2_196.out[0], compute_graph.l2_196.out[1], compute_graph.l2_196.out[2], compute_graph.l2_196.out[3], compute_graph.l2_197.in[0], compute_graph.l2_197.in[1], compute_graph.l2_197.in[2], compute_graph.l2_197.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1503), layer(216): compute_graph.flexml_layers[175].compute_node[0][0], compute_graph.flexml_layers[175].compute_node[0][1], compute_graph.flexml_layers[175].compute_node[0][2], compute_graph.flexml_layers[175].compute_node[0][3], compute_graph.flexml_layers[175].compute_node[1][0], compute_graph.flexml_layers[175].compute_node[1][1], compute_graph.flexml_layers[175].compute_node[1][2], compute_graph.flexml_layers[175].compute_node[1][3], compute_graph.flexml_layers[175].compute_node[2][0], compute_graph.flexml_layers[175].compute_node[2][1], compute_graph.flexml_layers[175].compute_node[2][2], compute_graph.flexml_layers[175].compute_node[2][3], compute_graph.flexml_layers[175].compute_node[3][0], compute_graph.flexml_layers[175].compute_node[3][1], compute_graph.flexml_layers[175].compute_node[3][2], compute_graph.flexml_layers[175].compute_node[3][3], {compute_graph.l2_197.out[0], compute_graph.l2_197.out[1], compute_graph.l2_197.out[2], compute_graph.l2_197.out[3], compute_graph.l2_198.in[0], compute_graph.l2_198.in[1], compute_graph.l2_198.in[2], compute_graph.l2_198.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1504), layer(217): compute_graph.flexml_layers[176].compute_node[0][0], compute_graph.flexml_layers[176].compute_node[0][1], compute_graph.flexml_layers[176].compute_node[0][2], compute_graph.flexml_layers[176].compute_node[0][3], compute_graph.flexml_layers[176].compute_node[1][0], compute_graph.flexml_layers[176].compute_node[1][1], compute_graph.flexml_layers[176].compute_node[1][2], compute_graph.flexml_layers[176].compute_node[1][3], compute_graph.flexml_layers[176].compute_node[2][0], compute_graph.flexml_layers[176].compute_node[2][1], compute_graph.flexml_layers[176].compute_node[2][2], compute_graph.flexml_layers[176].compute_node[2][3], compute_graph.flexml_layers[176].compute_node[3][0], compute_graph.flexml_layers[176].compute_node[3][1], compute_graph.flexml_layers[176].compute_node[3][2], compute_graph.flexml_layers[176].compute_node[3][3], {compute_graph.l2_198.out[0], compute_graph.l2_198.out[1], compute_graph.l2_198.out[2], compute_graph.l2_198.out[3], compute_graph.l2_199.in[0], compute_graph.l2_199.in[1], compute_graph.l2_199.in[2], compute_graph.l2_199.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1505), layer(218): compute_graph.flexml_layers[177].compute_node[0][0], compute_graph.flexml_layers[177].compute_node[0][1], compute_graph.flexml_layers[177].compute_node[0][2], compute_graph.flexml_layers[177].compute_node[0][3], compute_graph.flexml_layers[177].compute_node[1][0], compute_graph.flexml_layers[177].compute_node[1][1], compute_graph.flexml_layers[177].compute_node[1][2], compute_graph.flexml_layers[177].compute_node[1][3], compute_graph.flexml_layers[177].compute_node[2][0], compute_graph.flexml_layers[177].compute_node[2][1], compute_graph.flexml_layers[177].compute_node[2][2], compute_graph.flexml_layers[177].compute_node[2][3], compute_graph.flexml_layers[177].compute_node[3][0], compute_graph.flexml_layers[177].compute_node[3][1], compute_graph.flexml_layers[177].compute_node[3][2], compute_graph.flexml_layers[177].compute_node[3][3], {compute_graph.l2_199.out[0], compute_graph.l2_199.out[1], compute_graph.l2_199.out[2], compute_graph.l2_199.out[3], compute_graph.l2_200.in[0], compute_graph.l2_200.in[1], compute_graph.l2_200.in[2], compute_graph.l2_200.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(954): mode(0), layer(218): {compute_graph.l2_200.out[1], compute_graph.l2l3_200_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1117): mode(0), layer(219): {compute_graph.l2l3_200_spill.out[0], compute_graph.templated_graph_201.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1280): mode(0), layer(219): {compute_graph.templated_graph_201.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_201_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1118): mode(0), layer(220): {compute_graph.l2l3_scratch_0_201_spill.out[0], compute_graph.templated_graph_201.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1281): mode(0), layer(220): {compute_graph.templated_graph_201.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_201_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1119): mode(0), layer(221): {compute_graph.l2l3_scratch_1_201_spill.out[0], compute_graph.templated_graph_201.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1282): mode(0), layer(221): {compute_graph.templated_graph_201.g2.ifm_mem.out[0], compute_graph.l2l3_201_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1115): mode(0), layer(222): {compute_graph.l2l3_193_spill.out[1], compute_graph.l2l3_193_spill.out[2], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1120): mode(0), layer(222): {compute_graph.l2l3_201_spill.out[0], compute_graph.l2l3_201_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1506), layer(222): compute_graph.flexml_layers[178].compute_node[0][0], compute_graph.flexml_layers[178].compute_node[0][1], compute_graph.flexml_layers[178].compute_node[0][2], compute_graph.flexml_layers[178].compute_node[0][3], compute_graph.flexml_layers[178].compute_node[1][0], compute_graph.flexml_layers[178].compute_node[1][1], compute_graph.flexml_layers[178].compute_node[1][2], compute_graph.flexml_layers[178].compute_node[1][3], compute_graph.flexml_layers[178].compute_node[2][0], compute_graph.flexml_layers[178].compute_node[2][1], compute_graph.flexml_layers[178].compute_node[2][2], compute_graph.flexml_layers[178].compute_node[2][3], compute_graph.flexml_layers[178].compute_node[3][0], compute_graph.flexml_layers[178].compute_node[3][1], compute_graph.flexml_layers[178].compute_node[3][2], compute_graph.flexml_layers[178].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.out[3], compute_graph.l2_202.in[0], compute_graph.l2_202.in[1], compute_graph.l2_202.in[2], compute_graph.l2_202.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(879): mode(3), layer(223): {compute_graph.Layer_203_wts_ddr.out[0], compute_graph.Layer_203_wts_ddr.out[1], compute_graph.Layer_203_l2_wts.in[0], compute_graph.Layer_203_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(879) is pipelined with KernelLayerNode(1506) +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x0): Memory space used by compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183 overlaps with memory space used by compute_graph.l2_203. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1108): mode(0), layer(223): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[3], compute_graph.l2_202.out[0], compute_graph.l2_202.out[1], compute_graph.l2_202.out[2], compute_graph.l2_202.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1507), layer(223): compute_graph.flexml_layers[179].compute_node[0][0], compute_graph.flexml_layers[179].compute_node[0][1], compute_graph.flexml_layers[179].compute_node[0][2], compute_graph.flexml_layers[179].compute_node[0][3], compute_graph.flexml_layers[179].compute_node[1][0], compute_graph.flexml_layers[179].compute_node[1][1], compute_graph.flexml_layers[179].compute_node[1][2], compute_graph.flexml_layers[179].compute_node[1][3], compute_graph.flexml_layers[179].compute_node[2][0], compute_graph.flexml_layers[179].compute_node[2][1], compute_graph.flexml_layers[179].compute_node[2][2], compute_graph.flexml_layers[179].compute_node[2][3], compute_graph.flexml_layers[179].compute_node[3][0], compute_graph.flexml_layers[179].compute_node[3][1], compute_graph.flexml_layers[179].compute_node[3][2], compute_graph.flexml_layers[179].compute_node[3][3], {compute_graph.Layer_203_l2_wts.out[0], compute_graph.Layer_203_l2_wts.out[1], compute_graph.Layer_203_l2_wts.out[2], compute_graph.Layer_203_l2_wts.out[3], compute_graph.l2_202.out[0], compute_graph.l2_202.out[1], compute_graph.l2_202.out[2], compute_graph.l2_202.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.out[3], compute_graph.l2_203.in[0], compute_graph.l2_203.in[1], compute_graph.l2_203.in[2], compute_graph.l2_203.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(880): mode(0), layer(224): {compute_graph.Layer_204_wts_ddr.out[0], compute_graph.Layer_204_wts_ddr.out[1], compute_graph.Layer_204_l2_wts.in[0], compute_graph.Layer_204_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.Layer_204_l2_wts.out[0], compute_graph.Layer_204_l2_wts.out[1], compute_graph.Layer_204_l2_wts.out[2], compute_graph.Layer_204_l2_wts.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1508), layer(224): compute_graph.flexml_layers[180].compute_node[0][0], compute_graph.flexml_layers[180].compute_node[0][1], compute_graph.flexml_layers[180].compute_node[0][2], compute_graph.flexml_layers[180].compute_node[0][3], compute_graph.flexml_layers[180].compute_node[1][0], compute_graph.flexml_layers[180].compute_node[1][1], compute_graph.flexml_layers[180].compute_node[1][2], compute_graph.flexml_layers[180].compute_node[1][3], compute_graph.flexml_layers[180].compute_node[2][0], compute_graph.flexml_layers[180].compute_node[2][1], compute_graph.flexml_layers[180].compute_node[2][2], compute_graph.flexml_layers[180].compute_node[2][3], compute_graph.flexml_layers[180].compute_node[3][0], compute_graph.flexml_layers[180].compute_node[3][1], compute_graph.flexml_layers[180].compute_node[3][2], compute_graph.flexml_layers[180].compute_node[3][3], {compute_graph.Layer_204_l2_wts.out[0], compute_graph.Layer_204_l2_wts.out[1], compute_graph.Layer_204_l2_wts.out[2], compute_graph.Layer_204_l2_wts.out[3], compute_graph.l2_203.out[0], compute_graph.l2_203.out[1], compute_graph.l2_203.out[2], compute_graph.l2_203.out[3], compute_graph.l2_204.in[0], compute_graph.l2_204.in[1], compute_graph.l2_204.in[2], compute_graph.l2_204.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1509), layer(225): compute_graph.flexml_layers[181].compute_node[0][0], compute_graph.flexml_layers[181].compute_node[0][1], compute_graph.flexml_layers[181].compute_node[0][2], compute_graph.flexml_layers[181].compute_node[0][3], compute_graph.flexml_layers[181].compute_node[1][0], compute_graph.flexml_layers[181].compute_node[1][1], compute_graph.flexml_layers[181].compute_node[1][2], compute_graph.flexml_layers[181].compute_node[1][3], compute_graph.flexml_layers[181].compute_node[2][0], compute_graph.flexml_layers[181].compute_node[2][1], compute_graph.flexml_layers[181].compute_node[2][2], compute_graph.flexml_layers[181].compute_node[2][3], compute_graph.flexml_layers[181].compute_node[3][0], compute_graph.flexml_layers[181].compute_node[3][1], compute_graph.flexml_layers[181].compute_node[3][2], compute_graph.flexml_layers[181].compute_node[3][3], {compute_graph.l2_204.out[0], compute_graph.l2_204.out[1], compute_graph.l2_204.out[2], compute_graph.l2_204.out[3], compute_graph.l2_205.in[0], compute_graph.l2_205.in[1], compute_graph.l2_205.in[2], compute_graph.l2_205.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(955): mode(3), layer(225): {compute_graph.l2_204.out[4], compute_graph.l2_204.out[5], compute_graph.l2l3_204_spill.in[0], compute_graph.l2l3_204_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(955) is pipelined with KernelLayerNode(1509) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1510), layer(226): compute_graph.flexml_layers[182].compute_node[0][0], compute_graph.flexml_layers[182].compute_node[0][1], compute_graph.flexml_layers[182].compute_node[0][2], compute_graph.flexml_layers[182].compute_node[0][3], compute_graph.flexml_layers[182].compute_node[1][0], compute_graph.flexml_layers[182].compute_node[1][1], compute_graph.flexml_layers[182].compute_node[1][2], compute_graph.flexml_layers[182].compute_node[1][3], compute_graph.flexml_layers[182].compute_node[2][0], compute_graph.flexml_layers[182].compute_node[2][1], compute_graph.flexml_layers[182].compute_node[2][2], compute_graph.flexml_layers[182].compute_node[2][3], compute_graph.flexml_layers[182].compute_node[3][0], compute_graph.flexml_layers[182].compute_node[3][1], compute_graph.flexml_layers[182].compute_node[3][2], compute_graph.flexml_layers[182].compute_node[3][3], {compute_graph.l2_205.out[0], compute_graph.l2_205.out[1], compute_graph.l2_205.out[2], compute_graph.l2_205.out[3], compute_graph.l2_206.in[0], compute_graph.l2_206.in[1], compute_graph.l2_206.in[2], compute_graph.l2_206.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1511), layer(227): compute_graph.flexml_layers[183].compute_node[0][0], compute_graph.flexml_layers[183].compute_node[0][1], compute_graph.flexml_layers[183].compute_node[0][2], compute_graph.flexml_layers[183].compute_node[0][3], compute_graph.flexml_layers[183].compute_node[1][0], compute_graph.flexml_layers[183].compute_node[1][1], compute_graph.flexml_layers[183].compute_node[1][2], compute_graph.flexml_layers[183].compute_node[1][3], compute_graph.flexml_layers[183].compute_node[2][0], compute_graph.flexml_layers[183].compute_node[2][1], compute_graph.flexml_layers[183].compute_node[2][2], compute_graph.flexml_layers[183].compute_node[2][3], compute_graph.flexml_layers[183].compute_node[3][0], compute_graph.flexml_layers[183].compute_node[3][1], compute_graph.flexml_layers[183].compute_node[3][2], compute_graph.flexml_layers[183].compute_node[3][3], {compute_graph.l2_206.out[0], compute_graph.l2_206.out[1], compute_graph.l2_206.out[2], compute_graph.l2_206.out[3], compute_graph.l2_207.in[0], compute_graph.l2_207.in[1], compute_graph.l2_207.in[2], compute_graph.l2_207.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(956): mode(0), layer(227): {compute_graph.l2_207.out[0], compute_graph.l2_207.out[1], compute_graph.l2l3_207_spill.in[0], compute_graph.l2l3_207_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 957 will not be pipelined because it's in the same layer 228 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1605), layer(228): {compute_graph.l2l3_204_spill.out[0], compute_graph.l2l3_207_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1606), layer(228): {compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1607), layer(228): {compute_graph.l2l3_204_spill.out[1], compute_graph.l2l3_207_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1608), layer(228): {compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1512), layer(228): compute_graph.flexml_layers[184].compute_node[0][0], compute_graph.flexml_layers[184].compute_node[0][1], compute_graph.flexml_layers[184].compute_node[0][2], compute_graph.flexml_layers[184].compute_node[0][3], compute_graph.flexml_layers[184].compute_node[1][0], compute_graph.flexml_layers[184].compute_node[1][1], compute_graph.flexml_layers[184].compute_node[1][2], compute_graph.flexml_layers[184].compute_node[1][3], compute_graph.flexml_layers[184].compute_node[2][0], compute_graph.flexml_layers[184].compute_node[2][1], compute_graph.flexml_layers[184].compute_node[2][2], compute_graph.flexml_layers[184].compute_node[2][3], compute_graph.flexml_layers[184].compute_node[3][0], compute_graph.flexml_layers[184].compute_node[3][1], compute_graph.flexml_layers[184].compute_node[3][2], compute_graph.flexml_layers[184].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.out[3], compute_graph.l2_208.in[0], compute_graph.l2_208.in[1], compute_graph.l2_208.in[2], compute_graph.l2_208.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(957): mode(3), layer(228): {compute_graph.l2_208.out[0], compute_graph.l2_208.out[1], compute_graph.l2l3_208_spill.in[0], compute_graph.l2l3_208_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1123): mode(0), layer(229): {compute_graph.l2l3_208_spill.out[0], compute_graph.templated_graph_209.trans_mt_ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1283): mode(0), layer(229): {compute_graph.templated_graph_209.trans_mt_ifm.out[0], compute_graph.templated_graph_209.trans_mt_ofm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1284): mode(0), layer(229): {compute_graph.templated_graph_209.trans_mt_ofm.out[0], compute_graph.l2l3_209_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1125): mode(0), layer(230): {compute_graph.l2l3_209_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1513), layer(230): compute_graph.flexml_layers[185].compute_node[0][0], compute_graph.flexml_layers[185].compute_node[0][1], compute_graph.flexml_layers[185].compute_node[0][2], compute_graph.flexml_layers[185].compute_node[0][3], compute_graph.flexml_layers[185].compute_node[1][0], compute_graph.flexml_layers[185].compute_node[1][1], compute_graph.flexml_layers[185].compute_node[1][2], compute_graph.flexml_layers[185].compute_node[1][3], compute_graph.flexml_layers[185].compute_node[2][0], compute_graph.flexml_layers[185].compute_node[2][1], compute_graph.flexml_layers[185].compute_node[2][2], compute_graph.flexml_layers[185].compute_node[2][3], compute_graph.flexml_layers[185].compute_node[3][0], compute_graph.flexml_layers[185].compute_node[3][1], compute_graph.flexml_layers[185].compute_node[3][2], compute_graph.flexml_layers[185].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.out[3], compute_graph.l2_210.in[0], compute_graph.l2_210.in[1], compute_graph.l2_210.in[2], compute_graph.l2_210.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(881): mode(3), layer(231): {compute_graph.Layer_211_wts_ddr.out[0], compute_graph.Layer_211_wts_ddr.out[1], compute_graph.Layer_211_l2_wts.in[0], compute_graph.Layer_211_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(881) is pipelined with KernelLayerNode(1513) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1514), layer(231): compute_graph.flexml_layers[186].compute_node[0][0], compute_graph.flexml_layers[186].compute_node[0][1], compute_graph.flexml_layers[186].compute_node[0][2], compute_graph.flexml_layers[186].compute_node[0][3], compute_graph.flexml_layers[186].compute_node[1][0], compute_graph.flexml_layers[186].compute_node[1][1], compute_graph.flexml_layers[186].compute_node[1][2], compute_graph.flexml_layers[186].compute_node[1][3], compute_graph.flexml_layers[186].compute_node[2][0], compute_graph.flexml_layers[186].compute_node[2][1], compute_graph.flexml_layers[186].compute_node[2][2], compute_graph.flexml_layers[186].compute_node[2][3], compute_graph.flexml_layers[186].compute_node[3][0], compute_graph.flexml_layers[186].compute_node[3][1], compute_graph.flexml_layers[186].compute_node[3][2], compute_graph.flexml_layers[186].compute_node[3][3], {compute_graph.Layer_211_l2_wts.out[0], compute_graph.Layer_211_l2_wts.out[1], compute_graph.Layer_211_l2_wts.out[2], compute_graph.Layer_211_l2_wts.out[3], compute_graph.l2_210.out[0], compute_graph.l2_210.out[1], compute_graph.l2_210.out[2], compute_graph.l2_210.out[3], compute_graph.l2_211.in[0], compute_graph.l2_211.in[1], compute_graph.l2_211.in[2], compute_graph.l2_211.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1515), layer(232): compute_graph.flexml_layers[187].compute_node[0][0], compute_graph.flexml_layers[187].compute_node[0][1], compute_graph.flexml_layers[187].compute_node[0][2], compute_graph.flexml_layers[187].compute_node[0][3], compute_graph.flexml_layers[187].compute_node[1][0], compute_graph.flexml_layers[187].compute_node[1][1], compute_graph.flexml_layers[187].compute_node[1][2], compute_graph.flexml_layers[187].compute_node[1][3], compute_graph.flexml_layers[187].compute_node[2][0], compute_graph.flexml_layers[187].compute_node[2][1], compute_graph.flexml_layers[187].compute_node[2][2], compute_graph.flexml_layers[187].compute_node[2][3], compute_graph.flexml_layers[187].compute_node[3][0], compute_graph.flexml_layers[187].compute_node[3][1], compute_graph.flexml_layers[187].compute_node[3][2], compute_graph.flexml_layers[187].compute_node[3][3], {compute_graph.l2_211.out[0], compute_graph.l2_211.out[1], compute_graph.l2_211.out[2], compute_graph.l2_211.out[3], compute_graph.l2_212.in[0], compute_graph.l2_212.in[1], compute_graph.l2_212.in[2], compute_graph.l2_212.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(958): mode(0), layer(232): {compute_graph.l2_212.out[1], compute_graph.l2l3_212_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1126): mode(0), layer(233): {compute_graph.l2l3_212_spill.out[0], compute_graph.templated_graph_213.g0.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1285): mode(0), layer(233): {compute_graph.templated_graph_213.g0.ifm_mem.out[0], compute_graph.l2l3_scratch_0_213_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1127): mode(0), layer(234): {compute_graph.l2l3_scratch_0_213_spill.out[0], compute_graph.templated_graph_213.g1.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1286): mode(0), layer(234): {compute_graph.templated_graph_213.g1.ifm_mem.out[0], compute_graph.l2l3_scratch_1_213_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1128): mode(0), layer(235): {compute_graph.l2l3_scratch_1_213_spill.out[0], compute_graph.templated_graph_213.g2.ifm_mem.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1287): mode(0), layer(235): {compute_graph.templated_graph_213.g2.ifm_mem.out[0], compute_graph.l2l3_213_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 0, row 0 (Address: 0x70800): Memory space used by compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1 overlaps with memory space used by compute_graph.l2_214. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 0, row 0 +WARNING: [aiecompiler 77-22828] At tile tileType 2, col 1, row 0 (Address: 0x0): Memory space used by compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1 overlaps with memory space used by compute_graph.l2_214. +WARNING: [aiecompiler 77-22836] invalid memRsc request at tile location : tileType 2, col 1, row 0 +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1129): mode(0), layer(236): {compute_graph.l2l3_213_spill.out[0], compute_graph.l2l3_213_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(882): mode(4), layer(236): {compute_graph.Layer_214_wts_ddr.out[0], compute_graph.Layer_214_wts_ddr.out[1], compute_graph.Layer_214_l2_wts.in[0], compute_graph.Layer_214_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(882) is pipelined with BufferToBufferNode(1129) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1124): mode(0), layer(236): {compute_graph.l2l3_208_spill.out[1], compute_graph.l2l3_208_spill.out[2], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1516), layer(236): compute_graph.flexml_layers[188].compute_node[0][0], compute_graph.flexml_layers[188].compute_node[0][1], compute_graph.flexml_layers[188].compute_node[0][2], compute_graph.flexml_layers[188].compute_node[0][3], compute_graph.flexml_layers[188].compute_node[1][0], compute_graph.flexml_layers[188].compute_node[1][1], compute_graph.flexml_layers[188].compute_node[1][2], compute_graph.flexml_layers[188].compute_node[1][3], compute_graph.flexml_layers[188].compute_node[2][0], compute_graph.flexml_layers[188].compute_node[2][1], compute_graph.flexml_layers[188].compute_node[2][2], compute_graph.flexml_layers[188].compute_node[2][3], compute_graph.flexml_layers[188].compute_node[3][0], compute_graph.flexml_layers[188].compute_node[3][1], compute_graph.flexml_layers[188].compute_node[3][2], compute_graph.flexml_layers[188].compute_node[3][3], {compute_graph.Layer_214_l2_wts.out[0], compute_graph.Layer_214_l2_wts.out[1], compute_graph.Layer_214_l2_wts.out[2], compute_graph.Layer_214_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.out[3], compute_graph.l2_214.in[0], compute_graph.l2_214.in[1], compute_graph.l2_214.in[2], compute_graph.l2_214.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(959): mode(0), layer(236): {compute_graph.l2_214.out[0], compute_graph.l2_214.out[1], compute_graph.l2l3_214_spill.in[0], compute_graph.l2l3_214_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1130): mode(0), layer(237): {compute_graph.l2l3_214_spill.out[0], compute_graph.l2l3_214_spill.out[1], compute_graph.templated_graph_215.ifm.in[0], compute_graph.templated_graph_215.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1288): mode(0), layer(237): {compute_graph.templated_graph_215.ifm.out[0], compute_graph.templated_graph_215.ifm.out[1], compute_graph.l2l3_215_spill.in[0], compute_graph.l2l3_215_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1131): mode(0), layer(238): {compute_graph.l2l3_214_spill.out[2], compute_graph.l2l3_214_spill.out[3], compute_graph.templated_graph_216.ifm.in[0], compute_graph.templated_graph_216.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1289): mode(0), layer(238): {compute_graph.templated_graph_216.ifm.out[0], compute_graph.templated_graph_216.ifm.out[1], compute_graph.l2l3_216_spill.in[0], compute_graph.l2l3_216_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1021): mode(0), layer(239): {compute_graph.l2l3_5_spill.out[4], compute_graph.l2l3_5_spill.out[5], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1005): mode(0), layer(239): {compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.out[1], compute_graph.spill_L3_Concat_Buffer_layer_294.in[2], compute_graph.spill_L3_Concat_Buffer_layer_294.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1199): mode(0), layer(240): {compute_graph.ifm_ddr_1.out[0], compute_graph.ifm_ddr_1.out[1], compute_graph.L2_IFM_Buffer_for_input1_0_port1.in[0], compute_graph.L2_IFM_Buffer_for_input1_0_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(992): mode(0), layer(240): {compute_graph.L2_IFM_Buffer_for_input1_0_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_0_port1.out[1], compute_graph.spill_L3_Concat_Buffer_layer_279.in[2], compute_graph.spill_L3_Concat_Buffer_layer_279.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1157): mode(0), layer(241): {compute_graph.ifm_ddr_3.out[0], compute_graph.ifm_ddr_3.out[1], compute_graph.L2_IFM_Buffer_for_input3_0_port1.in[0], compute_graph.L2_IFM_Buffer_for_input3_0_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(976): mode(0), layer(241): {compute_graph.L2_IFM_Buffer_for_input3_0_port1.out[0], compute_graph.L2_IFM_Buffer_for_input3_0_port1.out[1], compute_graph.spill_L3_Concat_Buffer_layer_240.in[2], compute_graph.spill_L3_Concat_Buffer_layer_240.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1132): mode(0), layer(242): {compute_graph.l2l3_215_spill.out[0], compute_graph.l2l3_215_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(968): mode(0), layer(242): {compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_230.in[0], compute_graph.spill_L3_Concat_Buffer_layer_230.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1134): mode(0), layer(243): {compute_graph.l2l3_216_spill.out[2], compute_graph.l2l3_216_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(965): mode(0), layer(243): {compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_225.in[0], compute_graph.spill_L3_Concat_Buffer_layer_225.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1135): mode(0), layer(244): {compute_graph.ifm_ddr_4.out[0], compute_graph.ifm_ddr_4.out[1], compute_graph.L2_IFM_Buffer_for_input4_0_port1.in[0], compute_graph.L2_IFM_Buffer_for_input4_0_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(961): mode(0), layer(244): {compute_graph.L2_IFM_Buffer_for_input4_0_port1.out[0], compute_graph.L2_IFM_Buffer_for_input4_0_port1.out[1], compute_graph.spill_L3_Concat_Buffer_layer_217.in[2], compute_graph.spill_L3_Concat_Buffer_layer_217.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1133): mode(0), layer(245): {compute_graph.l2l3_216_spill.out[0], compute_graph.l2l3_216_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(960): mode(0), layer(245): {compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_217.in[0], compute_graph.spill_L3_Concat_Buffer_layer_217.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1138): mode(0), layer(246): {compute_graph.spill_L3_Concat_Buffer_layer_217.out[0], compute_graph.spill_L3_Concat_Buffer_layer_217.out[1], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(883): mode(4), layer(246): {compute_graph.Layer_218_wts_ddr.out[0], compute_graph.Layer_218_wts_ddr.out[1], compute_graph.Layer_218_l2_wts.in[0], compute_graph.Layer_218_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(883) is pipelined with BufferToBufferNode(1138) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1517), layer(246): compute_graph.flexml_layers[189].compute_node[0][0], compute_graph.flexml_layers[189].compute_node[0][1], compute_graph.flexml_layers[189].compute_node[0][2], compute_graph.flexml_layers[189].compute_node[0][3], compute_graph.flexml_layers[189].compute_node[1][0], compute_graph.flexml_layers[189].compute_node[1][1], compute_graph.flexml_layers[189].compute_node[1][2], compute_graph.flexml_layers[189].compute_node[1][3], compute_graph.flexml_layers[189].compute_node[2][0], compute_graph.flexml_layers[189].compute_node[2][1], compute_graph.flexml_layers[189].compute_node[2][2], compute_graph.flexml_layers[189].compute_node[2][3], compute_graph.flexml_layers[189].compute_node[3][0], compute_graph.flexml_layers[189].compute_node[3][1], compute_graph.flexml_layers[189].compute_node[3][2], compute_graph.flexml_layers[189].compute_node[3][3], {compute_graph.Layer_218_l2_wts.out[0], compute_graph.Layer_218_l2_wts.out[1], compute_graph.Layer_218_l2_wts.out[2], compute_graph.Layer_218_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.out[3], compute_graph.l2_218.in[0], compute_graph.l2_218.in[1], compute_graph.l2_218.in[2], compute_graph.l2_218.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1518), layer(247): compute_graph.flexml_layers[190].compute_node[0][0], compute_graph.flexml_layers[190].compute_node[0][1], compute_graph.flexml_layers[190].compute_node[0][2], compute_graph.flexml_layers[190].compute_node[0][3], compute_graph.flexml_layers[190].compute_node[1][0], compute_graph.flexml_layers[190].compute_node[1][1], compute_graph.flexml_layers[190].compute_node[1][2], compute_graph.flexml_layers[190].compute_node[1][3], compute_graph.flexml_layers[190].compute_node[2][0], compute_graph.flexml_layers[190].compute_node[2][1], compute_graph.flexml_layers[190].compute_node[2][2], compute_graph.flexml_layers[190].compute_node[2][3], compute_graph.flexml_layers[190].compute_node[3][0], compute_graph.flexml_layers[190].compute_node[3][1], compute_graph.flexml_layers[190].compute_node[3][2], compute_graph.flexml_layers[190].compute_node[3][3], {compute_graph.l2_218.out[0], compute_graph.l2_218.out[1], compute_graph.l2_218.out[2], compute_graph.l2_218.out[3], compute_graph.l2_219.in[0], compute_graph.l2_219.in[1], compute_graph.l2_219.in[2], compute_graph.l2_219.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(962): mode(0), layer(247): {compute_graph.l2_219.out[0], compute_graph.l2_219.out[1], compute_graph.l2l3_219_spill.in[0], compute_graph.l2l3_219_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1139): mode(0), layer(248): {compute_graph.l2l3_219_spill.out[0], compute_graph.l2l3_219_spill.out[1], compute_graph.templated_graph_220.ifm.in[0], compute_graph.templated_graph_220.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1290): mode(0), layer(248): {compute_graph.templated_graph_220.ifm.out[0], compute_graph.templated_graph_220.ifm.out[1], compute_graph.l2l3_220_spill.in[0], compute_graph.l2l3_220_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1141): mode(0), layer(249): {compute_graph.l2l3_220_spill.out[0], compute_graph.l2l3_220_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1143): mode(0), layer(249): {compute_graph.const_ifm_ddr_3.out[0], compute_graph.const_ifm_ddr_3.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1519), layer(249): compute_graph.flexml_layers[191].compute_node[0][0], compute_graph.flexml_layers[191].compute_node[0][1], compute_graph.flexml_layers[191].compute_node[0][2], compute_graph.flexml_layers[191].compute_node[0][3], compute_graph.flexml_layers[191].compute_node[1][0], compute_graph.flexml_layers[191].compute_node[1][1], compute_graph.flexml_layers[191].compute_node[1][2], compute_graph.flexml_layers[191].compute_node[1][3], compute_graph.flexml_layers[191].compute_node[2][0], compute_graph.flexml_layers[191].compute_node[2][1], compute_graph.flexml_layers[191].compute_node[2][2], compute_graph.flexml_layers[191].compute_node[2][3], compute_graph.flexml_layers[191].compute_node[3][0], compute_graph.flexml_layers[191].compute_node[3][1], compute_graph.flexml_layers[191].compute_node[3][2], compute_graph.flexml_layers[191].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.out[3], compute_graph.l2_221.in[0], compute_graph.l2_221.in[1], compute_graph.l2_221.in[2], compute_graph.l2_221.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1136): mode(0), layer(250): {compute_graph.ifm_ddr_4.out[2], compute_graph.ifm_ddr_4.out[3], compute_graph.L2_IFM_Buffer_for_input4_1_port1.in[0], compute_graph.L2_IFM_Buffer_for_input4_1_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1520), layer(250): compute_graph.flexml_layers[192].compute_node[0][0], compute_graph.flexml_layers[192].compute_node[0][1], compute_graph.flexml_layers[192].compute_node[0][2], compute_graph.flexml_layers[192].compute_node[0][3], compute_graph.flexml_layers[192].compute_node[1][0], compute_graph.flexml_layers[192].compute_node[1][1], compute_graph.flexml_layers[192].compute_node[1][2], compute_graph.flexml_layers[192].compute_node[1][3], compute_graph.flexml_layers[192].compute_node[2][0], compute_graph.flexml_layers[192].compute_node[2][1], compute_graph.flexml_layers[192].compute_node[2][2], compute_graph.flexml_layers[192].compute_node[2][3], compute_graph.flexml_layers[192].compute_node[3][0], compute_graph.flexml_layers[192].compute_node[3][1], compute_graph.flexml_layers[192].compute_node[3][2], compute_graph.flexml_layers[192].compute_node[3][3], {compute_graph.l2_221.out[0], compute_graph.l2_221.out[1], compute_graph.l2_221.out[2], compute_graph.l2_221.out[3], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input4_1_port1.out[3], compute_graph.l2_222.in[0], compute_graph.l2_222.in[1], compute_graph.l2_222.in[2], compute_graph.l2_222.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(963): mode(0), layer(250): {compute_graph.l2_222.out[0], compute_graph.l2_222.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1140): mode(0), layer(251): {compute_graph.l2l3_219_spill.out[2], compute_graph.l2l3_219_spill.out[3], compute_graph.templated_graph_223.ifm.in[0], compute_graph.templated_graph_223.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1291): mode(0), layer(251): {compute_graph.templated_graph_223.ifm.out[0], compute_graph.templated_graph_223.ifm.out[1], compute_graph.l2l3_223_spill.in[0], compute_graph.l2l3_223_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1145): mode(0), layer(252): {compute_graph.l2l3_223_spill.out[0], compute_graph.l2l3_223_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1137): mode(0), layer(252): {compute_graph.ifm_ddr_4.out[4], compute_graph.ifm_ddr_4.out[5], compute_graph.L2_IFM_Buffer_for_input4_2_port1.in[0], compute_graph.L2_IFM_Buffer_for_input4_2_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1521), layer(252): compute_graph.flexml_layers[193].compute_node[0][0], compute_graph.flexml_layers[193].compute_node[0][1], compute_graph.flexml_layers[193].compute_node[0][2], compute_graph.flexml_layers[193].compute_node[0][3], compute_graph.flexml_layers[193].compute_node[1][0], compute_graph.flexml_layers[193].compute_node[1][1], compute_graph.flexml_layers[193].compute_node[1][2], compute_graph.flexml_layers[193].compute_node[1][3], compute_graph.flexml_layers[193].compute_node[2][0], compute_graph.flexml_layers[193].compute_node[2][1], compute_graph.flexml_layers[193].compute_node[2][2], compute_graph.flexml_layers[193].compute_node[2][3], compute_graph.flexml_layers[193].compute_node[3][0], compute_graph.flexml_layers[193].compute_node[3][1], compute_graph.flexml_layers[193].compute_node[3][2], compute_graph.flexml_layers[193].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[0], compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[1], compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[2], compute_graph.L2_IFM_Buffer_for_input4_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.out[3], compute_graph.l2_224.in[0], compute_graph.l2_224.in[1], compute_graph.l2_224.in[2], compute_graph.l2_224.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(884): mode(3), layer(253): {compute_graph.Layer_226_wts_ddr.out[0], compute_graph.Layer_226_wts_ddr.out[1], compute_graph.Layer_226_l2_wts.in[0], compute_graph.Layer_226_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(884) is pipelined with KernelLayerNode(1521) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(964): mode(0), layer(252): {compute_graph.l2_224.out[0], compute_graph.l2_224.out[1], compute_graph.spill_L3_Concat_Buffer_layer_225.in[2], compute_graph.spill_L3_Concat_Buffer_layer_225.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1146): mode(0), layer(253): {compute_graph.spill_L3_Concat_Buffer_layer_225.out[0], compute_graph.spill_L3_Concat_Buffer_layer_225.out[1], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1522), layer(253): compute_graph.flexml_layers[194].compute_node[0][0], compute_graph.flexml_layers[194].compute_node[0][1], compute_graph.flexml_layers[194].compute_node[0][2], compute_graph.flexml_layers[194].compute_node[0][3], compute_graph.flexml_layers[194].compute_node[1][0], compute_graph.flexml_layers[194].compute_node[1][1], compute_graph.flexml_layers[194].compute_node[1][2], compute_graph.flexml_layers[194].compute_node[1][3], compute_graph.flexml_layers[194].compute_node[2][0], compute_graph.flexml_layers[194].compute_node[2][1], compute_graph.flexml_layers[194].compute_node[2][2], compute_graph.flexml_layers[194].compute_node[2][3], compute_graph.flexml_layers[194].compute_node[3][0], compute_graph.flexml_layers[194].compute_node[3][1], compute_graph.flexml_layers[194].compute_node[3][2], compute_graph.flexml_layers[194].compute_node[3][3], {compute_graph.Layer_226_l2_wts.out[0], compute_graph.Layer_226_l2_wts.out[1], compute_graph.Layer_226_l2_wts.out[2], compute_graph.Layer_226_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.out[3], compute_graph.l2_226.in[0], compute_graph.l2_226.in[1], compute_graph.l2_226.in[2], compute_graph.l2_226.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1523), layer(254): compute_graph.flexml_layers[195].compute_node[0][0], compute_graph.flexml_layers[195].compute_node[0][1], compute_graph.flexml_layers[195].compute_node[0][2], compute_graph.flexml_layers[195].compute_node[0][3], compute_graph.flexml_layers[195].compute_node[1][0], compute_graph.flexml_layers[195].compute_node[1][1], compute_graph.flexml_layers[195].compute_node[1][2], compute_graph.flexml_layers[195].compute_node[1][3], compute_graph.flexml_layers[195].compute_node[2][0], compute_graph.flexml_layers[195].compute_node[2][1], compute_graph.flexml_layers[195].compute_node[2][2], compute_graph.flexml_layers[195].compute_node[2][3], compute_graph.flexml_layers[195].compute_node[3][0], compute_graph.flexml_layers[195].compute_node[3][1], compute_graph.flexml_layers[195].compute_node[3][2], compute_graph.flexml_layers[195].compute_node[3][3], {compute_graph.l2_226.out[0], compute_graph.l2_226.out[1], compute_graph.l2_226.out[2], compute_graph.l2_226.out[3], compute_graph.l2_227.in[0], compute_graph.l2_227.in[1], compute_graph.l2_227.in[2], compute_graph.l2_227.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1142): mode(0), layer(255): {compute_graph.l2l3_220_spill.out[2], compute_graph.l2l3_220_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1524), layer(255): compute_graph.flexml_layers[196].compute_node[0][0], compute_graph.flexml_layers[196].compute_node[0][1], compute_graph.flexml_layers[196].compute_node[0][2], compute_graph.flexml_layers[196].compute_node[0][3], compute_graph.flexml_layers[196].compute_node[1][0], compute_graph.flexml_layers[196].compute_node[1][1], compute_graph.flexml_layers[196].compute_node[1][2], compute_graph.flexml_layers[196].compute_node[1][3], compute_graph.flexml_layers[196].compute_node[2][0], compute_graph.flexml_layers[196].compute_node[2][1], compute_graph.flexml_layers[196].compute_node[2][2], compute_graph.flexml_layers[196].compute_node[2][3], compute_graph.flexml_layers[196].compute_node[3][0], compute_graph.flexml_layers[196].compute_node[3][1], compute_graph.flexml_layers[196].compute_node[3][2], compute_graph.flexml_layers[196].compute_node[3][3], {compute_graph.l2_227.out[0], compute_graph.l2_227.out[1], compute_graph.l2_227.out[2], compute_graph.l2_227.out[3], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.out[3], compute_graph.l2_228.in[0], compute_graph.l2_228.in[1], compute_graph.l2_228.in[2], compute_graph.l2_228.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1144): mode(0), layer(256): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1525), layer(256): compute_graph.flexml_layers[197].compute_node[0][0], compute_graph.flexml_layers[197].compute_node[0][1], compute_graph.flexml_layers[197].compute_node[0][2], compute_graph.flexml_layers[197].compute_node[0][3], compute_graph.flexml_layers[197].compute_node[1][0], compute_graph.flexml_layers[197].compute_node[1][1], compute_graph.flexml_layers[197].compute_node[1][2], compute_graph.flexml_layers[197].compute_node[1][3], compute_graph.flexml_layers[197].compute_node[2][0], compute_graph.flexml_layers[197].compute_node[2][1], compute_graph.flexml_layers[197].compute_node[2][2], compute_graph.flexml_layers[197].compute_node[2][3], compute_graph.flexml_layers[197].compute_node[3][0], compute_graph.flexml_layers[197].compute_node[3][1], compute_graph.flexml_layers[197].compute_node[3][2], compute_graph.flexml_layers[197].compute_node[3][3], {compute_graph.l2_228.out[0], compute_graph.l2_228.out[1], compute_graph.l2_228.out[2], compute_graph.l2_228.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.out[3], compute_graph.l2_229.in[0], compute_graph.l2_229.in[1], compute_graph.l2_229.in[2], compute_graph.l2_229.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(967): mode(0), layer(256): {compute_graph.l2_229.out[2], compute_graph.l2_229.out[3], compute_graph.spill_L3_Concat_Buffer_layer_230.in[2], compute_graph.spill_L3_Concat_Buffer_layer_230.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(966): mode(0), layer(256): {compute_graph.l2_229.out[0], compute_graph.l2_229.out[1], compute_graph.ofm_ddr_0_l2l3_229_spill.in[0], compute_graph.ofm_ddr_0_l2l3_229_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1147): mode(0), layer(257): {compute_graph.spill_L3_Concat_Buffer_layer_230.out[0], compute_graph.spill_L3_Concat_Buffer_layer_230.out[1], compute_graph.templated_graph_231.ifm_mem.in[0], compute_graph.templated_graph_231.ifm_mem.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1293), layer(257): compute_graph.templated_graph_231.mk[0][0], compute_graph.templated_graph_231.mk[0][1], compute_graph.templated_graph_231.mk[0][2], compute_graph.templated_graph_231.mk[0][3], compute_graph.templated_graph_231.mk[1][0], compute_graph.templated_graph_231.mk[1][1], compute_graph.templated_graph_231.mk[1][2], compute_graph.templated_graph_231.mk[1][3], compute_graph.templated_graph_231.mk[2][0], compute_graph.templated_graph_231.mk[2][1], compute_graph.templated_graph_231.mk[2][2], compute_graph.templated_graph_231.mk[2][3], compute_graph.templated_graph_231.mk[3][0], compute_graph.templated_graph_231.mk[3][1], compute_graph.templated_graph_231.mk[3][2], compute_graph.templated_graph_231.mk[3][3], {compute_graph.templated_graph_231.ifm_mem.out[0], compute_graph.templated_graph_231.ifm_mem.out[1], compute_graph.templated_graph_231.ifm_mem.out[2], compute_graph.templated_graph_231.ifm_mem.out[3], compute_graph.templated_graph_231.ofm_mem.in[0], compute_graph.templated_graph_231.ofm_mem.in[1], compute_graph.templated_graph_231.ofm_mem.in[2], compute_graph.templated_graph_231.ofm_mem.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1292): mode(0), layer(257): {compute_graph.templated_graph_231.ofm_mem.out[0], compute_graph.templated_graph_231.ofm_mem.out[1], compute_graph.l2l3_231_spill.in[0], compute_graph.l2l3_231_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1148): mode(0), layer(258): {compute_graph.l2l3_231_spill.out[0], compute_graph.l2l3_231_spill.out[1], compute_graph.templated_graph_232.ifm.in[0], compute_graph.templated_graph_232.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1294): mode(0), layer(258): {compute_graph.templated_graph_232.ifm.out[0], compute_graph.templated_graph_232.ifm.out[1], compute_graph.l2l3_232_spill.in[0], compute_graph.l2l3_232_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1020): mode(0), layer(259): {compute_graph.l2l3_5_spill.out[2], compute_graph.l2l3_5_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1526), layer(259): compute_graph.flexml_layers[198].compute_node[0][0], compute_graph.flexml_layers[198].compute_node[0][1], compute_graph.flexml_layers[198].compute_node[0][2], compute_graph.flexml_layers[198].compute_node[0][3], compute_graph.flexml_layers[198].compute_node[1][0], compute_graph.flexml_layers[198].compute_node[1][1], compute_graph.flexml_layers[198].compute_node[1][2], compute_graph.flexml_layers[198].compute_node[1][3], compute_graph.flexml_layers[198].compute_node[2][0], compute_graph.flexml_layers[198].compute_node[2][1], compute_graph.flexml_layers[198].compute_node[2][2], compute_graph.flexml_layers[198].compute_node[2][3], compute_graph.flexml_layers[198].compute_node[3][0], compute_graph.flexml_layers[198].compute_node[3][1], compute_graph.flexml_layers[198].compute_node[3][2], compute_graph.flexml_layers[198].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.out[3], compute_graph.l2_233.in[0], compute_graph.l2_233.in[1], compute_graph.l2_233.in[2], compute_graph.l2_233.in[3]}, Scheduler computes number of sub-layer phases to be 24. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1609), layer(259): {compute_graph.l2_233.out[0], compute_graph.l2_233.out[2]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1610), layer(259): {compute_graph.l2l3_233_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[4]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1611), layer(259): {compute_graph.l2_233.out[1], compute_graph.l2_233.out[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1612), layer(259): {compute_graph.l2l3_233_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[5]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1150): mode(0), layer(260): {compute_graph.l2l3_233_spill.out[0], compute_graph.l2l3_233_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1527), layer(260): compute_graph.flexml_layers[199].compute_node[0][0], compute_graph.flexml_layers[199].compute_node[0][1], compute_graph.flexml_layers[199].compute_node[0][2], compute_graph.flexml_layers[199].compute_node[0][3], compute_graph.flexml_layers[199].compute_node[1][0], compute_graph.flexml_layers[199].compute_node[1][1], compute_graph.flexml_layers[199].compute_node[1][2], compute_graph.flexml_layers[199].compute_node[1][3], compute_graph.flexml_layers[199].compute_node[2][0], compute_graph.flexml_layers[199].compute_node[2][1], compute_graph.flexml_layers[199].compute_node[2][2], compute_graph.flexml_layers[199].compute_node[2][3], compute_graph.flexml_layers[199].compute_node[3][0], compute_graph.flexml_layers[199].compute_node[3][1], compute_graph.flexml_layers[199].compute_node[3][2], compute_graph.flexml_layers[199].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.out[3], compute_graph.l2_234.in[0], compute_graph.l2_234.in[1], compute_graph.l2_234.in[2], compute_graph.l2_234.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1528), layer(261): compute_graph.flexml_layers[200].compute_node[0][0], compute_graph.flexml_layers[200].compute_node[0][1], compute_graph.flexml_layers[200].compute_node[0][2], compute_graph.flexml_layers[200].compute_node[0][3], compute_graph.flexml_layers[200].compute_node[1][0], compute_graph.flexml_layers[200].compute_node[1][1], compute_graph.flexml_layers[200].compute_node[1][2], compute_graph.flexml_layers[200].compute_node[1][3], compute_graph.flexml_layers[200].compute_node[2][0], compute_graph.flexml_layers[200].compute_node[2][1], compute_graph.flexml_layers[200].compute_node[2][2], compute_graph.flexml_layers[200].compute_node[2][3], compute_graph.flexml_layers[200].compute_node[3][0], compute_graph.flexml_layers[200].compute_node[3][1], compute_graph.flexml_layers[200].compute_node[3][2], compute_graph.flexml_layers[200].compute_node[3][3], {compute_graph.l2_234.out[0], compute_graph.l2_234.out[1], compute_graph.l2_234.out[2], compute_graph.l2_234.out[3], compute_graph.l2_235.in[0], compute_graph.l2_235.in[1], compute_graph.l2_235.in[2], compute_graph.l2_235.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(972): mode(3), layer(261): {compute_graph.l2_234.out[4], compute_graph.l2_234.out[5], compute_graph.spill_L3_Concat_Buffer_layer_256.in[4], compute_graph.spill_L3_Concat_Buffer_layer_256.in[5]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(972) is pipelined with KernelLayerNode(1528) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(973): mode(0), layer(261): {compute_graph.l2_235.out[0], compute_graph.l2_235.out[1], compute_graph.spill_L3_Concat_Buffer_layer_236.in[4], compute_graph.spill_L3_Concat_Buffer_layer_236.in[5]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1149): mode(0), layer(262): {compute_graph.l2l3_232_spill.out[0], compute_graph.l2l3_232_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(969): mode(0), layer(262): {compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_236.in[0], compute_graph.spill_L3_Concat_Buffer_layer_236.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1151): mode(0), layer(263): {compute_graph.spill_L3_Concat_Buffer_layer_236.out[0], compute_graph.spill_L3_Concat_Buffer_layer_236.out[1], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(885): mode(4), layer(263): {compute_graph.Layer_237_wts_ddr.out[0], compute_graph.Layer_237_wts_ddr.out[1], compute_graph.Layer_237_l2_wts.in[0], compute_graph.Layer_237_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(885) is pipelined with BufferToBufferNode(1151) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1529), layer(263): compute_graph.flexml_layers[201].compute_node[0][0], compute_graph.flexml_layers[201].compute_node[0][1], compute_graph.flexml_layers[201].compute_node[0][2], compute_graph.flexml_layers[201].compute_node[0][3], compute_graph.flexml_layers[201].compute_node[1][0], compute_graph.flexml_layers[201].compute_node[1][1], compute_graph.flexml_layers[201].compute_node[1][2], compute_graph.flexml_layers[201].compute_node[1][3], compute_graph.flexml_layers[201].compute_node[2][0], compute_graph.flexml_layers[201].compute_node[2][1], compute_graph.flexml_layers[201].compute_node[2][2], compute_graph.flexml_layers[201].compute_node[2][3], compute_graph.flexml_layers[201].compute_node[3][0], compute_graph.flexml_layers[201].compute_node[3][1], compute_graph.flexml_layers[201].compute_node[3][2], compute_graph.flexml_layers[201].compute_node[3][3], {compute_graph.Layer_237_l2_wts.out[0], compute_graph.Layer_237_l2_wts.out[1], compute_graph.Layer_237_l2_wts.out[2], compute_graph.Layer_237_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.out[3], compute_graph.l2_237.in[0], compute_graph.l2_237.in[1], compute_graph.l2_237.in[2], compute_graph.l2_237.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(974): mode(0), layer(263): {compute_graph.l2_237.out[0], compute_graph.l2_237.out[1], compute_graph.l2l3_237_spill.in[0], compute_graph.l2l3_237_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1152): mode(0), layer(264): {compute_graph.l2l3_237_spill.out[0], compute_graph.l2l3_237_spill.out[1], compute_graph.templated_graph_238.ifm.in[0], compute_graph.templated_graph_238.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1295): mode(0), layer(264): {compute_graph.templated_graph_238.ifm.out[0], compute_graph.templated_graph_238.ifm.out[1], compute_graph.l2l3_238_spill.in[0], compute_graph.l2l3_238_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1153): mode(0), layer(265): {compute_graph.l2l3_237_spill.out[2], compute_graph.l2l3_237_spill.out[3], compute_graph.templated_graph_239.ifm.in[0], compute_graph.templated_graph_239.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1296): mode(0), layer(265): {compute_graph.templated_graph_239.ifm.out[0], compute_graph.templated_graph_239.ifm.out[1], compute_graph.l2l3_239_spill.in[0], compute_graph.l2l3_239_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1156): mode(0), layer(266): {compute_graph.l2l3_239_spill.out[2], compute_graph.l2l3_239_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(980): mode(0), layer(266): {compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_248.in[0], compute_graph.spill_L3_Concat_Buffer_layer_248.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1155): mode(0), layer(267): {compute_graph.l2l3_239_spill.out[0], compute_graph.l2l3_239_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(975): mode(0), layer(267): {compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_240.in[0], compute_graph.spill_L3_Concat_Buffer_layer_240.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1160): mode(0), layer(268): {compute_graph.spill_L3_Concat_Buffer_layer_240.out[0], compute_graph.spill_L3_Concat_Buffer_layer_240.out[1], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(886): mode(4), layer(268): {compute_graph.Layer_241_wts_ddr.out[0], compute_graph.Layer_241_wts_ddr.out[1], compute_graph.Layer_241_l2_wts.in[0], compute_graph.Layer_241_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(886) is pipelined with BufferToBufferNode(1160) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1530), layer(268): compute_graph.flexml_layers[202].compute_node[0][0], compute_graph.flexml_layers[202].compute_node[0][1], compute_graph.flexml_layers[202].compute_node[0][2], compute_graph.flexml_layers[202].compute_node[0][3], compute_graph.flexml_layers[202].compute_node[1][0], compute_graph.flexml_layers[202].compute_node[1][1], compute_graph.flexml_layers[202].compute_node[1][2], compute_graph.flexml_layers[202].compute_node[1][3], compute_graph.flexml_layers[202].compute_node[2][0], compute_graph.flexml_layers[202].compute_node[2][1], compute_graph.flexml_layers[202].compute_node[2][2], compute_graph.flexml_layers[202].compute_node[2][3], compute_graph.flexml_layers[202].compute_node[3][0], compute_graph.flexml_layers[202].compute_node[3][1], compute_graph.flexml_layers[202].compute_node[3][2], compute_graph.flexml_layers[202].compute_node[3][3], {compute_graph.Layer_241_l2_wts.out[0], compute_graph.Layer_241_l2_wts.out[1], compute_graph.Layer_241_l2_wts.out[2], compute_graph.Layer_241_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.out[3], compute_graph.l2_241.in[0], compute_graph.l2_241.in[1], compute_graph.l2_241.in[2], compute_graph.l2_241.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1531), layer(269): compute_graph.flexml_layers[203].compute_node[0][0], compute_graph.flexml_layers[203].compute_node[0][1], compute_graph.flexml_layers[203].compute_node[0][2], compute_graph.flexml_layers[203].compute_node[0][3], compute_graph.flexml_layers[203].compute_node[1][0], compute_graph.flexml_layers[203].compute_node[1][1], compute_graph.flexml_layers[203].compute_node[1][2], compute_graph.flexml_layers[203].compute_node[1][3], compute_graph.flexml_layers[203].compute_node[2][0], compute_graph.flexml_layers[203].compute_node[2][1], compute_graph.flexml_layers[203].compute_node[2][2], compute_graph.flexml_layers[203].compute_node[2][3], compute_graph.flexml_layers[203].compute_node[3][0], compute_graph.flexml_layers[203].compute_node[3][1], compute_graph.flexml_layers[203].compute_node[3][2], compute_graph.flexml_layers[203].compute_node[3][3], {compute_graph.l2_241.out[0], compute_graph.l2_241.out[1], compute_graph.l2_241.out[2], compute_graph.l2_241.out[3], compute_graph.l2_242.in[0], compute_graph.l2_242.in[1], compute_graph.l2_242.in[2], compute_graph.l2_242.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(977): mode(0), layer(269): {compute_graph.l2_242.out[0], compute_graph.l2_242.out[1], compute_graph.l2l3_242_spill.in[0], compute_graph.l2l3_242_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1161): mode(0), layer(270): {compute_graph.l2l3_242_spill.out[0], compute_graph.l2l3_242_spill.out[1], compute_graph.templated_graph_243.ifm.in[0], compute_graph.templated_graph_243.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1297): mode(0), layer(270): {compute_graph.templated_graph_243.ifm.out[0], compute_graph.templated_graph_243.ifm.out[1], compute_graph.l2l3_243_spill.in[0], compute_graph.l2l3_243_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1163): mode(0), layer(271): {compute_graph.l2l3_243_spill.out[0], compute_graph.l2l3_243_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1165): mode(0), layer(271): {compute_graph.const_ifm_ddr_2.out[0], compute_graph.const_ifm_ddr_2.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1532), layer(271): compute_graph.flexml_layers[204].compute_node[0][0], compute_graph.flexml_layers[204].compute_node[0][1], compute_graph.flexml_layers[204].compute_node[0][2], compute_graph.flexml_layers[204].compute_node[0][3], compute_graph.flexml_layers[204].compute_node[1][0], compute_graph.flexml_layers[204].compute_node[1][1], compute_graph.flexml_layers[204].compute_node[1][2], compute_graph.flexml_layers[204].compute_node[1][3], compute_graph.flexml_layers[204].compute_node[2][0], compute_graph.flexml_layers[204].compute_node[2][1], compute_graph.flexml_layers[204].compute_node[2][2], compute_graph.flexml_layers[204].compute_node[2][3], compute_graph.flexml_layers[204].compute_node[3][0], compute_graph.flexml_layers[204].compute_node[3][1], compute_graph.flexml_layers[204].compute_node[3][2], compute_graph.flexml_layers[204].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.out[3], compute_graph.l2_244.in[0], compute_graph.l2_244.in[1], compute_graph.l2_244.in[2], compute_graph.l2_244.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1158): mode(0), layer(272): {compute_graph.ifm_ddr_3.out[2], compute_graph.ifm_ddr_3.out[3], compute_graph.L2_IFM_Buffer_for_input3_1_port1.in[0], compute_graph.L2_IFM_Buffer_for_input3_1_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1533), layer(272): compute_graph.flexml_layers[205].compute_node[0][0], compute_graph.flexml_layers[205].compute_node[0][1], compute_graph.flexml_layers[205].compute_node[0][2], compute_graph.flexml_layers[205].compute_node[0][3], compute_graph.flexml_layers[205].compute_node[1][0], compute_graph.flexml_layers[205].compute_node[1][1], compute_graph.flexml_layers[205].compute_node[1][2], compute_graph.flexml_layers[205].compute_node[1][3], compute_graph.flexml_layers[205].compute_node[2][0], compute_graph.flexml_layers[205].compute_node[2][1], compute_graph.flexml_layers[205].compute_node[2][2], compute_graph.flexml_layers[205].compute_node[2][3], compute_graph.flexml_layers[205].compute_node[3][0], compute_graph.flexml_layers[205].compute_node[3][1], compute_graph.flexml_layers[205].compute_node[3][2], compute_graph.flexml_layers[205].compute_node[3][3], {compute_graph.l2_244.out[0], compute_graph.l2_244.out[1], compute_graph.l2_244.out[2], compute_graph.l2_244.out[3], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input3_1_port1.out[3], compute_graph.l2_245.in[0], compute_graph.l2_245.in[1], compute_graph.l2_245.in[2], compute_graph.l2_245.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(978): mode(0), layer(272): {compute_graph.l2_245.out[0], compute_graph.l2_245.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1162): mode(0), layer(273): {compute_graph.l2l3_242_spill.out[2], compute_graph.l2l3_242_spill.out[3], compute_graph.templated_graph_246.ifm.in[0], compute_graph.templated_graph_246.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1298): mode(0), layer(273): {compute_graph.templated_graph_246.ifm.out[0], compute_graph.templated_graph_246.ifm.out[1], compute_graph.l2l3_246_spill.in[0], compute_graph.l2l3_246_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1167): mode(0), layer(274): {compute_graph.l2l3_246_spill.out[0], compute_graph.l2l3_246_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1159): mode(0), layer(274): {compute_graph.ifm_ddr_3.out[4], compute_graph.ifm_ddr_3.out[5], compute_graph.L2_IFM_Buffer_for_input3_2_port1.in[0], compute_graph.L2_IFM_Buffer_for_input3_2_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1534), layer(274): compute_graph.flexml_layers[206].compute_node[0][0], compute_graph.flexml_layers[206].compute_node[0][1], compute_graph.flexml_layers[206].compute_node[0][2], compute_graph.flexml_layers[206].compute_node[0][3], compute_graph.flexml_layers[206].compute_node[1][0], compute_graph.flexml_layers[206].compute_node[1][1], compute_graph.flexml_layers[206].compute_node[1][2], compute_graph.flexml_layers[206].compute_node[1][3], compute_graph.flexml_layers[206].compute_node[2][0], compute_graph.flexml_layers[206].compute_node[2][1], compute_graph.flexml_layers[206].compute_node[2][2], compute_graph.flexml_layers[206].compute_node[2][3], compute_graph.flexml_layers[206].compute_node[3][0], compute_graph.flexml_layers[206].compute_node[3][1], compute_graph.flexml_layers[206].compute_node[3][2], compute_graph.flexml_layers[206].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[0], compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[1], compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[2], compute_graph.L2_IFM_Buffer_for_input3_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.out[3], compute_graph.l2_247.in[0], compute_graph.l2_247.in[1], compute_graph.l2_247.in[2], compute_graph.l2_247.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(887): mode(3), layer(275): {compute_graph.Layer_249_wts_ddr.out[0], compute_graph.Layer_249_wts_ddr.out[1], compute_graph.Layer_249_l2_wts.in[0], compute_graph.Layer_249_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(887) is pipelined with KernelLayerNode(1534) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(979): mode(0), layer(274): {compute_graph.l2_247.out[0], compute_graph.l2_247.out[1], compute_graph.spill_L3_Concat_Buffer_layer_248.in[2], compute_graph.spill_L3_Concat_Buffer_layer_248.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1168): mode(0), layer(275): {compute_graph.spill_L3_Concat_Buffer_layer_248.out[0], compute_graph.spill_L3_Concat_Buffer_layer_248.out[1], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1535), layer(275): compute_graph.flexml_layers[207].compute_node[0][0], compute_graph.flexml_layers[207].compute_node[0][1], compute_graph.flexml_layers[207].compute_node[0][2], compute_graph.flexml_layers[207].compute_node[0][3], compute_graph.flexml_layers[207].compute_node[1][0], compute_graph.flexml_layers[207].compute_node[1][1], compute_graph.flexml_layers[207].compute_node[1][2], compute_graph.flexml_layers[207].compute_node[1][3], compute_graph.flexml_layers[207].compute_node[2][0], compute_graph.flexml_layers[207].compute_node[2][1], compute_graph.flexml_layers[207].compute_node[2][2], compute_graph.flexml_layers[207].compute_node[2][3], compute_graph.flexml_layers[207].compute_node[3][0], compute_graph.flexml_layers[207].compute_node[3][1], compute_graph.flexml_layers[207].compute_node[3][2], compute_graph.flexml_layers[207].compute_node[3][3], {compute_graph.Layer_249_l2_wts.out[0], compute_graph.Layer_249_l2_wts.out[1], compute_graph.Layer_249_l2_wts.out[2], compute_graph.Layer_249_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.out[3], compute_graph.l2_249.in[0], compute_graph.l2_249.in[1], compute_graph.l2_249.in[2], compute_graph.l2_249.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1536), layer(276): compute_graph.flexml_layers[208].compute_node[0][0], compute_graph.flexml_layers[208].compute_node[0][1], compute_graph.flexml_layers[208].compute_node[0][2], compute_graph.flexml_layers[208].compute_node[0][3], compute_graph.flexml_layers[208].compute_node[1][0], compute_graph.flexml_layers[208].compute_node[1][1], compute_graph.flexml_layers[208].compute_node[1][2], compute_graph.flexml_layers[208].compute_node[1][3], compute_graph.flexml_layers[208].compute_node[2][0], compute_graph.flexml_layers[208].compute_node[2][1], compute_graph.flexml_layers[208].compute_node[2][2], compute_graph.flexml_layers[208].compute_node[2][3], compute_graph.flexml_layers[208].compute_node[3][0], compute_graph.flexml_layers[208].compute_node[3][1], compute_graph.flexml_layers[208].compute_node[3][2], compute_graph.flexml_layers[208].compute_node[3][3], {compute_graph.l2_249.out[0], compute_graph.l2_249.out[1], compute_graph.l2_249.out[2], compute_graph.l2_249.out[3], compute_graph.l2_250.in[0], compute_graph.l2_250.in[1], compute_graph.l2_250.in[2], compute_graph.l2_250.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1164): mode(0), layer(277): {compute_graph.l2l3_243_spill.out[2], compute_graph.l2l3_243_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1537), layer(277): compute_graph.flexml_layers[209].compute_node[0][0], compute_graph.flexml_layers[209].compute_node[0][1], compute_graph.flexml_layers[209].compute_node[0][2], compute_graph.flexml_layers[209].compute_node[0][3], compute_graph.flexml_layers[209].compute_node[1][0], compute_graph.flexml_layers[209].compute_node[1][1], compute_graph.flexml_layers[209].compute_node[1][2], compute_graph.flexml_layers[209].compute_node[1][3], compute_graph.flexml_layers[209].compute_node[2][0], compute_graph.flexml_layers[209].compute_node[2][1], compute_graph.flexml_layers[209].compute_node[2][2], compute_graph.flexml_layers[209].compute_node[2][3], compute_graph.flexml_layers[209].compute_node[3][0], compute_graph.flexml_layers[209].compute_node[3][1], compute_graph.flexml_layers[209].compute_node[3][2], compute_graph.flexml_layers[209].compute_node[3][3], {compute_graph.l2_250.out[0], compute_graph.l2_250.out[1], compute_graph.l2_250.out[2], compute_graph.l2_250.out[3], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.out[3], compute_graph.l2_251.in[0], compute_graph.l2_251.in[1], compute_graph.l2_251.in[2], compute_graph.l2_251.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1166): mode(0), layer(278): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1538), layer(278): compute_graph.flexml_layers[210].compute_node[0][0], compute_graph.flexml_layers[210].compute_node[0][1], compute_graph.flexml_layers[210].compute_node[0][2], compute_graph.flexml_layers[210].compute_node[0][3], compute_graph.flexml_layers[210].compute_node[1][0], compute_graph.flexml_layers[210].compute_node[1][1], compute_graph.flexml_layers[210].compute_node[1][2], compute_graph.flexml_layers[210].compute_node[1][3], compute_graph.flexml_layers[210].compute_node[2][0], compute_graph.flexml_layers[210].compute_node[2][1], compute_graph.flexml_layers[210].compute_node[2][2], compute_graph.flexml_layers[210].compute_node[2][3], compute_graph.flexml_layers[210].compute_node[3][0], compute_graph.flexml_layers[210].compute_node[3][1], compute_graph.flexml_layers[210].compute_node[3][2], compute_graph.flexml_layers[210].compute_node[3][3], {compute_graph.l2_251.out[0], compute_graph.l2_251.out[1], compute_graph.l2_251.out[2], compute_graph.l2_251.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.out[3], compute_graph.l2_252.in[0], compute_graph.l2_252.in[1], compute_graph.l2_252.in[2], compute_graph.l2_252.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(981): mode(0), layer(278): {compute_graph.l2_252.out[0], compute_graph.l2_252.out[1], compute_graph.ofm_ddr_1_l2l3_252_spill.in[0], compute_graph.ofm_ddr_1_l2l3_252_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(982): mode(0), layer(278): {compute_graph.l2_252.out[2], compute_graph.l2_252.out[3], compute_graph.spill_L3_Concat_Buffer_layer_253.in[2], compute_graph.spill_L3_Concat_Buffer_layer_253.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1154): mode(0), layer(279): {compute_graph.l2l3_238_spill.out[0], compute_graph.l2l3_238_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(983): mode(0), layer(279): {compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_253.in[0], compute_graph.spill_L3_Concat_Buffer_layer_253.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1169): mode(0), layer(280): {compute_graph.spill_L3_Concat_Buffer_layer_253.out[0], compute_graph.spill_L3_Concat_Buffer_layer_253.out[1], compute_graph.templated_graph_254.ifm_mem.in[0], compute_graph.templated_graph_254.ifm_mem.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.templated_graph_254.ofm_mem.in[0], compute_graph.templated_graph_254.ofm_mem.in[1], compute_graph.templated_graph_254.ofm_mem.in[2], compute_graph.templated_graph_254.ofm_mem.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1300), layer(280): compute_graph.templated_graph_254.mk[0][0], compute_graph.templated_graph_254.mk[0][1], compute_graph.templated_graph_254.mk[0][2], compute_graph.templated_graph_254.mk[0][3], compute_graph.templated_graph_254.mk[1][0], compute_graph.templated_graph_254.mk[1][1], compute_graph.templated_graph_254.mk[1][2], compute_graph.templated_graph_254.mk[1][3], compute_graph.templated_graph_254.mk[2][0], compute_graph.templated_graph_254.mk[2][1], compute_graph.templated_graph_254.mk[2][2], compute_graph.templated_graph_254.mk[2][3], compute_graph.templated_graph_254.mk[3][0], compute_graph.templated_graph_254.mk[3][1], compute_graph.templated_graph_254.mk[3][2], compute_graph.templated_graph_254.mk[3][3], {compute_graph.templated_graph_254.ifm_mem.out[0], compute_graph.templated_graph_254.ifm_mem.out[1], compute_graph.templated_graph_254.ifm_mem.out[2], compute_graph.templated_graph_254.ifm_mem.out[3], compute_graph.templated_graph_254.ofm_mem.in[0], compute_graph.templated_graph_254.ofm_mem.in[1], compute_graph.templated_graph_254.ofm_mem.in[2], compute_graph.templated_graph_254.ofm_mem.in[3]}, Scheduler computes number of sub-layer phases to be 2. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1299): mode(3), layer(280): {compute_graph.templated_graph_254.ofm_mem.out[0], compute_graph.templated_graph_254.ofm_mem.out[1], compute_graph.l2l3_254_spill.in[0], compute_graph.l2l3_254_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1299) is pipelined with KernelLayerNode(1300) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1170): mode(0), layer(281): {compute_graph.l2l3_254_spill.out[0], compute_graph.l2l3_254_spill.out[1], compute_graph.templated_graph_255.ifm.in[0], compute_graph.templated_graph_255.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1301): mode(0), layer(281): {compute_graph.templated_graph_255.ifm.out[0], compute_graph.templated_graph_255.ifm.out[1], compute_graph.l2l3_255_spill.in[0], compute_graph.l2l3_255_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1171): mode(0), layer(282): {compute_graph.l2l3_255_spill.out[0], compute_graph.l2l3_255_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(984): mode(0), layer(282): {compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_256.in[0], compute_graph.spill_L3_Concat_Buffer_layer_256.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1172): mode(0), layer(283): {compute_graph.spill_L3_Concat_Buffer_layer_256.out[0], compute_graph.spill_L3_Concat_Buffer_layer_256.out[1], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(888): mode(4), layer(283): {compute_graph.Layer_257_wts_ddr.out[0], compute_graph.Layer_257_wts_ddr.out[1], compute_graph.Layer_257_l2_wts.in[0], compute_graph.Layer_257_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(888) is pipelined with BufferToBufferNode(1172) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1539), layer(283): compute_graph.flexml_layers[211].compute_node[0][0], compute_graph.flexml_layers[211].compute_node[0][1], compute_graph.flexml_layers[211].compute_node[0][2], compute_graph.flexml_layers[211].compute_node[0][3], compute_graph.flexml_layers[211].compute_node[1][0], compute_graph.flexml_layers[211].compute_node[1][1], compute_graph.flexml_layers[211].compute_node[1][2], compute_graph.flexml_layers[211].compute_node[1][3], compute_graph.flexml_layers[211].compute_node[2][0], compute_graph.flexml_layers[211].compute_node[2][1], compute_graph.flexml_layers[211].compute_node[2][2], compute_graph.flexml_layers[211].compute_node[2][3], compute_graph.flexml_layers[211].compute_node[3][0], compute_graph.flexml_layers[211].compute_node[3][1], compute_graph.flexml_layers[211].compute_node[3][2], compute_graph.flexml_layers[211].compute_node[3][3], {compute_graph.Layer_257_l2_wts.out[0], compute_graph.Layer_257_l2_wts.out[1], compute_graph.Layer_257_l2_wts.out[2], compute_graph.Layer_257_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.out[3], compute_graph.l2_257.in[0], compute_graph.l2_257.in[1], compute_graph.l2_257.in[2], compute_graph.l2_257.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(985): mode(0), layer(283): {compute_graph.l2_257.out[0], compute_graph.l2_257.out[1], compute_graph.l2l3_257_spill.in[0], compute_graph.l2l3_257_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1173): mode(0), layer(284): {compute_graph.l2l3_257_spill.out[0], compute_graph.l2l3_257_spill.out[1], compute_graph.templated_graph_258.ifm.in[0], compute_graph.templated_graph_258.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1303), layer(284): compute_graph.templated_graph_258.compute_node[0][0], compute_graph.templated_graph_258.compute_node[0][1], compute_graph.templated_graph_258.compute_node[0][2], compute_graph.templated_graph_258.compute_node[0][3], compute_graph.templated_graph_258.compute_node[1][0], compute_graph.templated_graph_258.compute_node[1][1], compute_graph.templated_graph_258.compute_node[1][2], compute_graph.templated_graph_258.compute_node[1][3], compute_graph.templated_graph_258.compute_node[2][0], compute_graph.templated_graph_258.compute_node[2][1], compute_graph.templated_graph_258.compute_node[2][2], compute_graph.templated_graph_258.compute_node[2][3], compute_graph.templated_graph_258.compute_node[3][0], compute_graph.templated_graph_258.compute_node[3][1], compute_graph.templated_graph_258.compute_node[3][2], compute_graph.templated_graph_258.compute_node[3][3], {compute_graph.templated_graph_258.ifm.out[0], compute_graph.templated_graph_258.ifm.out[1], compute_graph.templated_graph_258.ifm.out[2], compute_graph.templated_graph_258.ifm.out[3], compute_graph.templated_graph_258.ofm.in[0], compute_graph.templated_graph_258.ofm.in[1], compute_graph.templated_graph_258.ofm.in[2], compute_graph.templated_graph_258.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1302): mode(3), layer(284): {compute_graph.templated_graph_258.ofm.out[0], compute_graph.templated_graph_258.ofm.out[1], compute_graph.l2l3_258_spill.in[0], compute_graph.l2l3_258_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1302) is pipelined with KernelLayerNode(1303) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1174): mode(0), layer(285): {compute_graph.l2l3_257_spill.out[2], compute_graph.l2l3_257_spill.out[3], compute_graph.templated_graph_259.ifm.in[0], compute_graph.templated_graph_259.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1305), layer(285): compute_graph.templated_graph_259.compute_node[0][0], compute_graph.templated_graph_259.compute_node[0][1], compute_graph.templated_graph_259.compute_node[0][2], compute_graph.templated_graph_259.compute_node[0][3], compute_graph.templated_graph_259.compute_node[1][0], compute_graph.templated_graph_259.compute_node[1][1], compute_graph.templated_graph_259.compute_node[1][2], compute_graph.templated_graph_259.compute_node[1][3], compute_graph.templated_graph_259.compute_node[2][0], compute_graph.templated_graph_259.compute_node[2][1], compute_graph.templated_graph_259.compute_node[2][2], compute_graph.templated_graph_259.compute_node[2][3], compute_graph.templated_graph_259.compute_node[3][0], compute_graph.templated_graph_259.compute_node[3][1], compute_graph.templated_graph_259.compute_node[3][2], compute_graph.templated_graph_259.compute_node[3][3], {compute_graph.templated_graph_259.ifm.out[0], compute_graph.templated_graph_259.ifm.out[1], compute_graph.templated_graph_259.ifm.out[2], compute_graph.templated_graph_259.ifm.out[3], compute_graph.templated_graph_259.ofm.in[0], compute_graph.templated_graph_259.ofm.in[1], compute_graph.templated_graph_259.ofm.in[2], compute_graph.templated_graph_259.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1304): mode(3), layer(285): {compute_graph.templated_graph_259.ofm.out[0], compute_graph.templated_graph_259.ofm.out[1], compute_graph.l2l3_259_spill.in[0], compute_graph.l2l3_259_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1304) is pipelined with KernelLayerNode(1305) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1178): mode(3), layer(286): {compute_graph.ifm_ddr_2.out[0], compute_graph.templated_graph_260.wts.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1178) is pipelined with KernelLayerNode(1305) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1176): mode(0), layer(286): {compute_graph.l2l3_259_spill.out[0], compute_graph.templated_graph_260.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1307), layer(286): compute_graph.templated_graph_260.mk[0][0], compute_graph.templated_graph_260.mk[0][1], compute_graph.templated_graph_260.mk[0][2], compute_graph.templated_graph_260.mk[0][3], compute_graph.templated_graph_260.mk[1][0], compute_graph.templated_graph_260.mk[1][1], compute_graph.templated_graph_260.mk[1][2], compute_graph.templated_graph_260.mk[1][3], compute_graph.templated_graph_260.mk[2][0], compute_graph.templated_graph_260.mk[2][1], compute_graph.templated_graph_260.mk[2][2], compute_graph.templated_graph_260.mk[2][3], compute_graph.templated_graph_260.mk[3][0], compute_graph.templated_graph_260.mk[3][1], compute_graph.templated_graph_260.mk[3][2], compute_graph.templated_graph_260.mk[3][3], {compute_graph.templated_graph_260.ifm.out[0], compute_graph.templated_graph_260.ifm.out[1], compute_graph.templated_graph_260.ifm.out[2], compute_graph.templated_graph_260.ifm.out[3], compute_graph.templated_graph_260.wts.out[0], compute_graph.templated_graph_260.wts.out[1], compute_graph.templated_graph_260.wts.out[2], compute_graph.templated_graph_260.wts.out[3], compute_graph.templated_graph_260.ofm.in[0], compute_graph.templated_graph_260.ofm.in[1], compute_graph.templated_graph_260.ofm.in[2], compute_graph.templated_graph_260.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1306): mode(3), layer(286): {compute_graph.templated_graph_260.ofm.out[0], compute_graph.l2l3_260_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1306) is pipelined with KernelLayerNode(1307) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(889): mode(3), layer(287): {compute_graph.Layer_261_wts_ddr.out[0], compute_graph.Layer_261_wts_ddr.out[1], compute_graph.Layer_261_l2_wts.in[0], compute_graph.Layer_261_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(889) is pipelined with KernelLayerNode(1307) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1181): mode(0), layer(287): {compute_graph.l2l3_260_spill.out[0], compute_graph.l2l3_260_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1540), layer(287): compute_graph.flexml_layers[212].compute_node[0][0], compute_graph.flexml_layers[212].compute_node[0][1], compute_graph.flexml_layers[212].compute_node[0][2], compute_graph.flexml_layers[212].compute_node[0][3], compute_graph.flexml_layers[212].compute_node[1][0], compute_graph.flexml_layers[212].compute_node[1][1], compute_graph.flexml_layers[212].compute_node[1][2], compute_graph.flexml_layers[212].compute_node[1][3], compute_graph.flexml_layers[212].compute_node[2][0], compute_graph.flexml_layers[212].compute_node[2][1], compute_graph.flexml_layers[212].compute_node[2][2], compute_graph.flexml_layers[212].compute_node[2][3], compute_graph.flexml_layers[212].compute_node[3][0], compute_graph.flexml_layers[212].compute_node[3][1], compute_graph.flexml_layers[212].compute_node[3][2], compute_graph.flexml_layers[212].compute_node[3][3], {compute_graph.Layer_261_l2_wts.out[0], compute_graph.Layer_261_l2_wts.out[1], compute_graph.Layer_261_l2_wts.out[2], compute_graph.Layer_261_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.out[3], compute_graph.l2_261.in[0], compute_graph.l2_261.in[1], compute_graph.l2_261.in[2], compute_graph.l2_261.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2_261.out[0], compute_graph.l2_261.out[1], compute_graph.l2_261.out[2], compute_graph.l2_261.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1541), layer(288): compute_graph.flexml_layers[213].compute_node[0][0], compute_graph.flexml_layers[213].compute_node[0][1], compute_graph.flexml_layers[213].compute_node[0][2], compute_graph.flexml_layers[213].compute_node[0][3], compute_graph.flexml_layers[213].compute_node[1][0], compute_graph.flexml_layers[213].compute_node[1][1], compute_graph.flexml_layers[213].compute_node[1][2], compute_graph.flexml_layers[213].compute_node[1][3], compute_graph.flexml_layers[213].compute_node[2][0], compute_graph.flexml_layers[213].compute_node[2][1], compute_graph.flexml_layers[213].compute_node[2][2], compute_graph.flexml_layers[213].compute_node[2][3], compute_graph.flexml_layers[213].compute_node[3][0], compute_graph.flexml_layers[213].compute_node[3][1], compute_graph.flexml_layers[213].compute_node[3][2], compute_graph.flexml_layers[213].compute_node[3][3], {compute_graph.l2_261.out[0], compute_graph.l2_261.out[1], compute_graph.l2_261.out[2], compute_graph.l2_261.out[3], compute_graph.l2_262.in[0], compute_graph.l2_262.in[1], compute_graph.l2_262.in[2], compute_graph.l2_262.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(986): mode(0), layer(288): {compute_graph.l2_262.out[0], compute_graph.l2_262.out[1], compute_graph.l2l3_262_spill.in[0], compute_graph.l2l3_262_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1182): mode(0), layer(289): {compute_graph.l2l3_262_spill.out[0], compute_graph.l2l3_262_spill.out[1], compute_graph.templated_graph_263.ifm.in[0], compute_graph.templated_graph_263.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1309), layer(289): compute_graph.templated_graph_263.compute_node[0][0], compute_graph.templated_graph_263.compute_node[0][1], compute_graph.templated_graph_263.compute_node[0][2], compute_graph.templated_graph_263.compute_node[0][3], compute_graph.templated_graph_263.compute_node[1][0], compute_graph.templated_graph_263.compute_node[1][1], compute_graph.templated_graph_263.compute_node[1][2], compute_graph.templated_graph_263.compute_node[1][3], compute_graph.templated_graph_263.compute_node[2][0], compute_graph.templated_graph_263.compute_node[2][1], compute_graph.templated_graph_263.compute_node[2][2], compute_graph.templated_graph_263.compute_node[2][3], compute_graph.templated_graph_263.compute_node[3][0], compute_graph.templated_graph_263.compute_node[3][1], compute_graph.templated_graph_263.compute_node[3][2], compute_graph.templated_graph_263.compute_node[3][3], {compute_graph.templated_graph_263.ifm.out[0], compute_graph.templated_graph_263.ifm.out[1], compute_graph.templated_graph_263.ifm.out[2], compute_graph.templated_graph_263.ifm.out[3], compute_graph.templated_graph_263.ofm.in[0], compute_graph.templated_graph_263.ofm.in[1], compute_graph.templated_graph_263.ofm.in[2], compute_graph.templated_graph_263.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1308): mode(3), layer(289): {compute_graph.templated_graph_263.ofm.out[0], compute_graph.templated_graph_263.ofm.out[1], compute_graph.l2l3_263_spill.in[0], compute_graph.l2l3_263_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1308) is pipelined with KernelLayerNode(1309) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1184): mode(0), layer(290): {compute_graph.l2l3_263_spill.out[0], compute_graph.l2l3_263_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1186): mode(0), layer(290): {compute_graph.const_ifm_ddr_1.out[0], compute_graph.const_ifm_ddr_1.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.in[0], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1542), layer(290): compute_graph.flexml_layers[214].compute_node[0][0], compute_graph.flexml_layers[214].compute_node[0][1], compute_graph.flexml_layers[214].compute_node[0][2], compute_graph.flexml_layers[214].compute_node[0][3], compute_graph.flexml_layers[214].compute_node[1][0], compute_graph.flexml_layers[214].compute_node[1][1], compute_graph.flexml_layers[214].compute_node[1][2], compute_graph.flexml_layers[214].compute_node[1][3], compute_graph.flexml_layers[214].compute_node[2][0], compute_graph.flexml_layers[214].compute_node[2][1], compute_graph.flexml_layers[214].compute_node[2][2], compute_graph.flexml_layers[214].compute_node[2][3], compute_graph.flexml_layers[214].compute_node[3][0], compute_graph.flexml_layers[214].compute_node[3][1], compute_graph.flexml_layers[214].compute_node[3][2], compute_graph.flexml_layers[214].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.out[3], compute_graph.l2_264.in[0], compute_graph.l2_264.in[1], compute_graph.l2_264.in[2], compute_graph.l2_264.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1179): mode(0), layer(291): {compute_graph.ifm_ddr_2.out[1], compute_graph.ifm_ddr_2.out[2], compute_graph.L2_IFM_Buffer_for_input2_1_port1.in[0], compute_graph.L2_IFM_Buffer_for_input2_1_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1543), layer(291): compute_graph.flexml_layers[215].compute_node[0][0], compute_graph.flexml_layers[215].compute_node[0][1], compute_graph.flexml_layers[215].compute_node[0][2], compute_graph.flexml_layers[215].compute_node[0][3], compute_graph.flexml_layers[215].compute_node[1][0], compute_graph.flexml_layers[215].compute_node[1][1], compute_graph.flexml_layers[215].compute_node[1][2], compute_graph.flexml_layers[215].compute_node[1][3], compute_graph.flexml_layers[215].compute_node[2][0], compute_graph.flexml_layers[215].compute_node[2][1], compute_graph.flexml_layers[215].compute_node[2][2], compute_graph.flexml_layers[215].compute_node[2][3], compute_graph.flexml_layers[215].compute_node[3][0], compute_graph.flexml_layers[215].compute_node[3][1], compute_graph.flexml_layers[215].compute_node[3][2], compute_graph.flexml_layers[215].compute_node[3][3], {compute_graph.l2_264.out[0], compute_graph.l2_264.out[1], compute_graph.l2_264.out[2], compute_graph.l2_264.out[3], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input2_1_port1.out[3], compute_graph.l2_265.in[0], compute_graph.l2_265.in[1], compute_graph.l2_265.in[2], compute_graph.l2_265.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(987): mode(0), layer(291): {compute_graph.l2_265.out[0], compute_graph.l2_265.out[1], compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.in[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1183): mode(0), layer(292): {compute_graph.l2l3_262_spill.out[2], compute_graph.l2l3_262_spill.out[3], compute_graph.templated_graph_266.ifm.in[0], compute_graph.templated_graph_266.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1311), layer(292): compute_graph.templated_graph_266.compute_node[0][0], compute_graph.templated_graph_266.compute_node[0][1], compute_graph.templated_graph_266.compute_node[0][2], compute_graph.templated_graph_266.compute_node[0][3], compute_graph.templated_graph_266.compute_node[1][0], compute_graph.templated_graph_266.compute_node[1][1], compute_graph.templated_graph_266.compute_node[1][2], compute_graph.templated_graph_266.compute_node[1][3], compute_graph.templated_graph_266.compute_node[2][0], compute_graph.templated_graph_266.compute_node[2][1], compute_graph.templated_graph_266.compute_node[2][2], compute_graph.templated_graph_266.compute_node[2][3], compute_graph.templated_graph_266.compute_node[3][0], compute_graph.templated_graph_266.compute_node[3][1], compute_graph.templated_graph_266.compute_node[3][2], compute_graph.templated_graph_266.compute_node[3][3], {compute_graph.templated_graph_266.ifm.out[0], compute_graph.templated_graph_266.ifm.out[1], compute_graph.templated_graph_266.ifm.out[2], compute_graph.templated_graph_266.ifm.out[3], compute_graph.templated_graph_266.ofm.in[0], compute_graph.templated_graph_266.ofm.in[1], compute_graph.templated_graph_266.ofm.in[2], compute_graph.templated_graph_266.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1310): mode(3), layer(292): {compute_graph.templated_graph_266.ofm.out[0], compute_graph.templated_graph_266.ofm.out[1], compute_graph.l2l3_266_spill.in[0], compute_graph.l2l3_266_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1310) is pipelined with KernelLayerNode(1311) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1188): mode(0), layer(293): {compute_graph.l2l3_266_spill.out[0], compute_graph.l2l3_266_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1180): mode(0), layer(293): {compute_graph.ifm_ddr_2.out[3], compute_graph.ifm_ddr_2.out[4], compute_graph.L2_IFM_Buffer_for_input2_2_port1.in[0], compute_graph.L2_IFM_Buffer_for_input2_2_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1544), layer(293): compute_graph.flexml_layers[216].compute_node[0][0], compute_graph.flexml_layers[216].compute_node[0][1], compute_graph.flexml_layers[216].compute_node[0][2], compute_graph.flexml_layers[216].compute_node[0][3], compute_graph.flexml_layers[216].compute_node[1][0], compute_graph.flexml_layers[216].compute_node[1][1], compute_graph.flexml_layers[216].compute_node[1][2], compute_graph.flexml_layers[216].compute_node[1][3], compute_graph.flexml_layers[216].compute_node[2][0], compute_graph.flexml_layers[216].compute_node[2][1], compute_graph.flexml_layers[216].compute_node[2][2], compute_graph.flexml_layers[216].compute_node[2][3], compute_graph.flexml_layers[216].compute_node[3][0], compute_graph.flexml_layers[216].compute_node[3][1], compute_graph.flexml_layers[216].compute_node[3][2], compute_graph.flexml_layers[216].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[0], compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[1], compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[2], compute_graph.L2_IFM_Buffer_for_input2_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.out[3], compute_graph.l2_267.in[0], compute_graph.l2_267.in[1], compute_graph.l2_267.in[2], compute_graph.l2_267.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(988): mode(0), layer(293): {compute_graph.l2_267.out[0], compute_graph.l2_267.out[1], compute_graph.l2l3_267_spill.in[0], compute_graph.l2l3_267_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1177): mode(0), layer(294): {compute_graph.l2l3_259_spill.out[1], compute_graph.templated_graph_268.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1189): mode(4), layer(294): {compute_graph.l2l3_267_spill.out[0], compute_graph.templated_graph_268.wts.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(1189) is pipelined with BufferToBufferNode(1177) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1313), layer(294): compute_graph.templated_graph_268.mk[0][0], compute_graph.templated_graph_268.mk[0][1], compute_graph.templated_graph_268.mk[0][2], compute_graph.templated_graph_268.mk[0][3], compute_graph.templated_graph_268.mk[1][0], compute_graph.templated_graph_268.mk[1][1], compute_graph.templated_graph_268.mk[1][2], compute_graph.templated_graph_268.mk[1][3], compute_graph.templated_graph_268.mk[2][0], compute_graph.templated_graph_268.mk[2][1], compute_graph.templated_graph_268.mk[2][2], compute_graph.templated_graph_268.mk[2][3], compute_graph.templated_graph_268.mk[3][0], compute_graph.templated_graph_268.mk[3][1], compute_graph.templated_graph_268.mk[3][2], compute_graph.templated_graph_268.mk[3][3], {compute_graph.templated_graph_268.ifm.out[0], compute_graph.templated_graph_268.ifm.out[1], compute_graph.templated_graph_268.ifm.out[2], compute_graph.templated_graph_268.ifm.out[3], compute_graph.templated_graph_268.wts.out[0], compute_graph.templated_graph_268.wts.out[1], compute_graph.templated_graph_268.wts.out[2], compute_graph.templated_graph_268.wts.out[3], compute_graph.templated_graph_268.ofm.in[0], compute_graph.templated_graph_268.ofm.in[1], compute_graph.templated_graph_268.ofm.in[2], compute_graph.templated_graph_268.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1312): mode(3), layer(294): {compute_graph.templated_graph_268.ofm.out[0], compute_graph.l2l3_268_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1312) is pipelined with KernelLayerNode(1313) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1190): mode(0), layer(295): {compute_graph.l2l3_268_spill.out[0], compute_graph.l2l3_268_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(890): mode(4), layer(295): {compute_graph.Layer_269_wts_ddr.out[0], compute_graph.Layer_269_wts_ddr.out[1], compute_graph.Layer_269_l2_wts.in[0], compute_graph.Layer_269_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(890) is pipelined with BufferToBufferNode(1190) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1545), layer(295): compute_graph.flexml_layers[217].compute_node[0][0], compute_graph.flexml_layers[217].compute_node[0][1], compute_graph.flexml_layers[217].compute_node[0][2], compute_graph.flexml_layers[217].compute_node[0][3], compute_graph.flexml_layers[217].compute_node[1][0], compute_graph.flexml_layers[217].compute_node[1][1], compute_graph.flexml_layers[217].compute_node[1][2], compute_graph.flexml_layers[217].compute_node[1][3], compute_graph.flexml_layers[217].compute_node[2][0], compute_graph.flexml_layers[217].compute_node[2][1], compute_graph.flexml_layers[217].compute_node[2][2], compute_graph.flexml_layers[217].compute_node[2][3], compute_graph.flexml_layers[217].compute_node[3][0], compute_graph.flexml_layers[217].compute_node[3][1], compute_graph.flexml_layers[217].compute_node[3][2], compute_graph.flexml_layers[217].compute_node[3][3], {compute_graph.Layer_269_l2_wts.out[0], compute_graph.Layer_269_l2_wts.out[1], compute_graph.Layer_269_l2_wts.out[2], compute_graph.Layer_269_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.out[3], compute_graph.l2_269.in[0], compute_graph.l2_269.in[1], compute_graph.l2_269.in[2], compute_graph.l2_269.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1546), layer(296): compute_graph.flexml_layers[218].compute_node[0][0], compute_graph.flexml_layers[218].compute_node[0][1], compute_graph.flexml_layers[218].compute_node[0][2], compute_graph.flexml_layers[218].compute_node[0][3], compute_graph.flexml_layers[218].compute_node[1][0], compute_graph.flexml_layers[218].compute_node[1][1], compute_graph.flexml_layers[218].compute_node[1][2], compute_graph.flexml_layers[218].compute_node[1][3], compute_graph.flexml_layers[218].compute_node[2][0], compute_graph.flexml_layers[218].compute_node[2][1], compute_graph.flexml_layers[218].compute_node[2][2], compute_graph.flexml_layers[218].compute_node[2][3], compute_graph.flexml_layers[218].compute_node[3][0], compute_graph.flexml_layers[218].compute_node[3][1], compute_graph.flexml_layers[218].compute_node[3][2], compute_graph.flexml_layers[218].compute_node[3][3], {compute_graph.l2_269.out[0], compute_graph.l2_269.out[1], compute_graph.l2_269.out[2], compute_graph.l2_269.out[3], compute_graph.l2_270.in[0], compute_graph.l2_270.in[1], compute_graph.l2_270.in[2], compute_graph.l2_270.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1185): mode(0), layer(297): {compute_graph.l2l3_263_spill.out[2], compute_graph.l2l3_263_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1547), layer(297): compute_graph.flexml_layers[219].compute_node[0][0], compute_graph.flexml_layers[219].compute_node[0][1], compute_graph.flexml_layers[219].compute_node[0][2], compute_graph.flexml_layers[219].compute_node[0][3], compute_graph.flexml_layers[219].compute_node[1][0], compute_graph.flexml_layers[219].compute_node[1][1], compute_graph.flexml_layers[219].compute_node[1][2], compute_graph.flexml_layers[219].compute_node[1][3], compute_graph.flexml_layers[219].compute_node[2][0], compute_graph.flexml_layers[219].compute_node[2][1], compute_graph.flexml_layers[219].compute_node[2][2], compute_graph.flexml_layers[219].compute_node[2][3], compute_graph.flexml_layers[219].compute_node[3][0], compute_graph.flexml_layers[219].compute_node[3][1], compute_graph.flexml_layers[219].compute_node[3][2], compute_graph.flexml_layers[219].compute_node[3][3], {compute_graph.l2_270.out[0], compute_graph.l2_270.out[1], compute_graph.l2_270.out[2], compute_graph.l2_270.out[3], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.out[3], compute_graph.l2_271.in[0], compute_graph.l2_271.in[1], compute_graph.l2_271.in[2], compute_graph.l2_271.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1187): mode(0), layer(298): {compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.out[0], compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.in[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1548), layer(298): compute_graph.flexml_layers[220].compute_node[0][0], compute_graph.flexml_layers[220].compute_node[0][1], compute_graph.flexml_layers[220].compute_node[0][2], compute_graph.flexml_layers[220].compute_node[0][3], compute_graph.flexml_layers[220].compute_node[1][0], compute_graph.flexml_layers[220].compute_node[1][1], compute_graph.flexml_layers[220].compute_node[1][2], compute_graph.flexml_layers[220].compute_node[1][3], compute_graph.flexml_layers[220].compute_node[2][0], compute_graph.flexml_layers[220].compute_node[2][1], compute_graph.flexml_layers[220].compute_node[2][2], compute_graph.flexml_layers[220].compute_node[2][3], compute_graph.flexml_layers[220].compute_node[3][0], compute_graph.flexml_layers[220].compute_node[3][1], compute_graph.flexml_layers[220].compute_node[3][2], compute_graph.flexml_layers[220].compute_node[3][3], {compute_graph.l2_271.out[0], compute_graph.l2_271.out[1], compute_graph.l2_271.out[2], compute_graph.l2_271.out[3], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[0], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[1], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[2], compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.out[3], compute_graph.l2_272.in[0], compute_graph.l2_272.in[1], compute_graph.l2_272.in[2], compute_graph.l2_272.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(989): mode(0), layer(298): {compute_graph.l2_272.out[0], compute_graph.l2_272.out[1], compute_graph.ofm_ddr_2_l2l3_272_spill.in[0], compute_graph.ofm_ddr_2_l2l3_272_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1175): mode(0), layer(299): {compute_graph.l2l3_258_spill.out[0], compute_graph.templated_graph_273.ifm.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1191): mode(4), layer(299): {compute_graph.ofm_ddr_2_l2l3_272_spill.out[0], compute_graph.templated_graph_273.wts.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(1191) is pipelined with BufferToBufferNode(1175) +INFO: [aiecompiler 77-6295] For KernelLayerNode(1315), layer(299): compute_graph.templated_graph_273.mk[0][0], compute_graph.templated_graph_273.mk[0][1], compute_graph.templated_graph_273.mk[0][2], compute_graph.templated_graph_273.mk[0][3], compute_graph.templated_graph_273.mk[1][0], compute_graph.templated_graph_273.mk[1][1], compute_graph.templated_graph_273.mk[1][2], compute_graph.templated_graph_273.mk[1][3], compute_graph.templated_graph_273.mk[2][0], compute_graph.templated_graph_273.mk[2][1], compute_graph.templated_graph_273.mk[2][2], compute_graph.templated_graph_273.mk[2][3], compute_graph.templated_graph_273.mk[3][0], compute_graph.templated_graph_273.mk[3][1], compute_graph.templated_graph_273.mk[3][2], compute_graph.templated_graph_273.mk[3][3], {compute_graph.templated_graph_273.ifm.out[0], compute_graph.templated_graph_273.ifm.out[1], compute_graph.templated_graph_273.ifm.out[2], compute_graph.templated_graph_273.ifm.out[3], compute_graph.templated_graph_273.wts.out[0], compute_graph.templated_graph_273.wts.out[1], compute_graph.templated_graph_273.wts.out[2], compute_graph.templated_graph_273.wts.out[3], compute_graph.templated_graph_273.ofm.in[0], compute_graph.templated_graph_273.ofm.in[1], compute_graph.templated_graph_273.ofm.in[2], compute_graph.templated_graph_273.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1314): mode(3), layer(299): {compute_graph.templated_graph_273.ofm.out[0], compute_graph.l2l3_273_spill.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1314) is pipelined with KernelLayerNode(1315) +INFO: [aiecompiler 77-23129] BufferToBufferNode 1316 will not be pipelined because it's in the same layer 300 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2l3_273_spill.out[0], compute_graph.l2l3_273_spill.out[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1192): mode(0), layer(300): {compute_graph.l2l3_273_spill.out[0], compute_graph.l2l3_273_spill.out[1], compute_graph.templated_graph_274.ifm_mem.in[0], compute_graph.templated_graph_274.ifm_mem.in[1]}, Scheduler computes number of sub-layer phases to be 12. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1317), layer(300): compute_graph.templated_graph_274.mk[0][0], compute_graph.templated_graph_274.mk[0][1], compute_graph.templated_graph_274.mk[0][2], compute_graph.templated_graph_274.mk[0][3], compute_graph.templated_graph_274.mk[1][0], compute_graph.templated_graph_274.mk[1][1], compute_graph.templated_graph_274.mk[1][2], compute_graph.templated_graph_274.mk[1][3], compute_graph.templated_graph_274.mk[2][0], compute_graph.templated_graph_274.mk[2][1], compute_graph.templated_graph_274.mk[2][2], compute_graph.templated_graph_274.mk[2][3], compute_graph.templated_graph_274.mk[3][0], compute_graph.templated_graph_274.mk[3][1], compute_graph.templated_graph_274.mk[3][2], compute_graph.templated_graph_274.mk[3][3], {compute_graph.templated_graph_274.ifm_mem.out[0], compute_graph.templated_graph_274.ifm_mem.out[1], compute_graph.templated_graph_274.ifm_mem.out[2], compute_graph.templated_graph_274.ifm_mem.out[3], compute_graph.templated_graph_274.ofm_mem.in[0], compute_graph.templated_graph_274.ofm_mem.in[1], compute_graph.templated_graph_274.ofm_mem.in[2], compute_graph.templated_graph_274.ofm_mem.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(891): mode(3), layer(301): {compute_graph.Layer_276_wts_ddr.out[0], compute_graph.Layer_276_wts_ddr.out[1], compute_graph.Layer_276_l2_wts.in[0], compute_graph.Layer_276_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(891) is pipelined with KernelLayerNode(1317) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.templated_graph_274.ofm_mem.out[0], compute_graph.templated_graph_274.ofm_mem.out[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1316): mode(3), layer(300): {compute_graph.templated_graph_274.ofm_mem.out[0], compute_graph.templated_graph_274.ofm_mem.out[1], compute_graph.spill_L3_Concat_Buffer_layer_275.in[0], compute_graph.spill_L3_Concat_Buffer_layer_275.in[1]}, Scheduler computes number of sub-layer phases to be 12. +INFO: [aiecompiler 77-23129] BufferToBufferNode 990 will not be pipelined because it's in the same layer 301 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.spill_L3_Concat_Buffer_layer_275.out[0], compute_graph.spill_L3_Concat_Buffer_layer_275.out[1], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1193): mode(0), layer(301): {compute_graph.spill_L3_Concat_Buffer_layer_275.out[0], compute_graph.spill_L3_Concat_Buffer_layer_275.out[1], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[1]}, Scheduler computes number of sub-layer phases to be 5. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1549), layer(301): compute_graph.flexml_layers[221].compute_node[0][0], compute_graph.flexml_layers[221].compute_node[0][1], compute_graph.flexml_layers[221].compute_node[0][2], compute_graph.flexml_layers[221].compute_node[0][3], compute_graph.flexml_layers[221].compute_node[1][0], compute_graph.flexml_layers[221].compute_node[1][1], compute_graph.flexml_layers[221].compute_node[1][2], compute_graph.flexml_layers[221].compute_node[1][3], compute_graph.flexml_layers[221].compute_node[2][0], compute_graph.flexml_layers[221].compute_node[2][1], compute_graph.flexml_layers[221].compute_node[2][2], compute_graph.flexml_layers[221].compute_node[2][3], compute_graph.flexml_layers[221].compute_node[3][0], compute_graph.flexml_layers[221].compute_node[3][1], compute_graph.flexml_layers[221].compute_node[3][2], compute_graph.flexml_layers[221].compute_node[3][3], {compute_graph.Layer_276_l2_wts.out[0], compute_graph.Layer_276_l2_wts.out[1], compute_graph.Layer_276_l2_wts.out[2], compute_graph.Layer_276_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.out[3], compute_graph.l2_276.in[0], compute_graph.l2_276.in[1], compute_graph.l2_276.in[2], compute_graph.l2_276.in[3]}, Scheduler computes number of sub-layer phases to be 5. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(990): mode(3), layer(301): {compute_graph.l2_276.out[0], compute_graph.l2_276.out[1], compute_graph.l2l3_276_spill.in[0], compute_graph.l2l3_276_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1194): mode(0), layer(302): {compute_graph.l2l3_276_spill.out[0], compute_graph.l2l3_276_spill.out[1], compute_graph.templated_graph_277.ifm.in[0], compute_graph.templated_graph_277.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1318): mode(0), layer(302): {compute_graph.templated_graph_277.ifm.out[0], compute_graph.templated_graph_277.ifm.out[1], compute_graph.l2l3_277_spill.in[0], compute_graph.l2l3_277_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1195): mode(0), layer(303): {compute_graph.l2l3_276_spill.out[2], compute_graph.l2l3_276_spill.out[3], compute_graph.templated_graph_278.ifm.in[0], compute_graph.templated_graph_278.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1319): mode(0), layer(303): {compute_graph.templated_graph_278.ifm.out[0], compute_graph.templated_graph_278.ifm.out[1], compute_graph.l2l3_278_spill.in[0], compute_graph.l2l3_278_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1198): mode(0), layer(304): {compute_graph.l2l3_278_spill.out[2], compute_graph.l2l3_278_spill.out[3], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(998): mode(0), layer(304): {compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_287.in[0], compute_graph.spill_L3_Concat_Buffer_layer_287.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1197): mode(0), layer(305): {compute_graph.l2l3_278_spill.out[0], compute_graph.l2l3_278_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(991): mode(0), layer(305): {compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_279.in[0], compute_graph.spill_L3_Concat_Buffer_layer_279.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1202): mode(0), layer(306): {compute_graph.spill_L3_Concat_Buffer_layer_279.out[0], compute_graph.spill_L3_Concat_Buffer_layer_279.out[1], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(892): mode(4), layer(306): {compute_graph.Layer_280_wts_ddr.out[0], compute_graph.Layer_280_wts_ddr.out[1], compute_graph.Layer_280_l2_wts.in[0], compute_graph.Layer_280_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22491] BufferToBufferNode(892) is pipelined with BufferToBufferNode(1202) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[3], compute_graph.l2_280.in[0], compute_graph.l2_280.in[1], compute_graph.l2_280.in[2], compute_graph.l2_280.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1550), layer(306): compute_graph.flexml_layers[222].compute_node[0][0], compute_graph.flexml_layers[222].compute_node[0][1], compute_graph.flexml_layers[222].compute_node[0][2], compute_graph.flexml_layers[222].compute_node[0][3], compute_graph.flexml_layers[222].compute_node[1][0], compute_graph.flexml_layers[222].compute_node[1][1], compute_graph.flexml_layers[222].compute_node[1][2], compute_graph.flexml_layers[222].compute_node[1][3], compute_graph.flexml_layers[222].compute_node[2][0], compute_graph.flexml_layers[222].compute_node[2][1], compute_graph.flexml_layers[222].compute_node[2][2], compute_graph.flexml_layers[222].compute_node[2][3], compute_graph.flexml_layers[222].compute_node[3][0], compute_graph.flexml_layers[222].compute_node[3][1], compute_graph.flexml_layers[222].compute_node[3][2], compute_graph.flexml_layers[222].compute_node[3][3], {compute_graph.Layer_280_l2_wts.out[0], compute_graph.Layer_280_l2_wts.out[1], compute_graph.Layer_280_l2_wts.out[2], compute_graph.Layer_280_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.out[3], compute_graph.l2_280.in[0], compute_graph.l2_280.in[1], compute_graph.l2_280.in[2], compute_graph.l2_280.in[3]}, Scheduler computes number of sub-layer phases to be 6. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(993): mode(3), layer(306): {compute_graph.l2_280.out[0], compute_graph.l2_280.out[1], compute_graph.l2l3_280_spill.in[0], compute_graph.l2l3_280_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(993) is pipelined with KernelLayerNode(1550) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1203): mode(0), layer(307): {compute_graph.l2l3_280_spill.out[0], compute_graph.l2l3_280_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1551), layer(307): compute_graph.flexml_layers[223].compute_node[0][0], compute_graph.flexml_layers[223].compute_node[0][1], compute_graph.flexml_layers[223].compute_node[0][2], compute_graph.flexml_layers[223].compute_node[0][3], compute_graph.flexml_layers[223].compute_node[1][0], compute_graph.flexml_layers[223].compute_node[1][1], compute_graph.flexml_layers[223].compute_node[1][2], compute_graph.flexml_layers[223].compute_node[1][3], compute_graph.flexml_layers[223].compute_node[2][0], compute_graph.flexml_layers[223].compute_node[2][1], compute_graph.flexml_layers[223].compute_node[2][2], compute_graph.flexml_layers[223].compute_node[2][3], compute_graph.flexml_layers[223].compute_node[3][0], compute_graph.flexml_layers[223].compute_node[3][1], compute_graph.flexml_layers[223].compute_node[3][2], compute_graph.flexml_layers[223].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.out[3], compute_graph.l2_281.in[0], compute_graph.l2_281.in[1], compute_graph.l2_281.in[2], compute_graph.l2_281.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(994): mode(3), layer(307): {compute_graph.l2_281.out[0], compute_graph.l2_281.out[1], compute_graph.l2l3_281_spill.in[0], compute_graph.l2l3_281_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(994) is pipelined with KernelLayerNode(1551) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1204): mode(0), layer(308): {compute_graph.l2l3_281_spill.out[0], compute_graph.l2l3_281_spill.out[1], compute_graph.templated_graph_282.ifm.in[0], compute_graph.templated_graph_282.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1320): mode(0), layer(308): {compute_graph.templated_graph_282.ifm.out[0], compute_graph.templated_graph_282.ifm.out[1], compute_graph.l2l3_282_spill.in[0], compute_graph.l2l3_282_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 995 will not be pipelined because it's in the same layer 309 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1613), layer(309): {compute_graph.l2l3_282_spill.out[0], compute_graph.const_ifm_ddr.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1614), layer(309): {compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1615), layer(309): {compute_graph.l2l3_282_spill.out[1], compute_graph.const_ifm_ddr.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1616), layer(309): {compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1552), layer(309): compute_graph.flexml_layers[224].compute_node[0][0], compute_graph.flexml_layers[224].compute_node[0][1], compute_graph.flexml_layers[224].compute_node[0][2], compute_graph.flexml_layers[224].compute_node[0][3], compute_graph.flexml_layers[224].compute_node[1][0], compute_graph.flexml_layers[224].compute_node[1][1], compute_graph.flexml_layers[224].compute_node[1][2], compute_graph.flexml_layers[224].compute_node[1][3], compute_graph.flexml_layers[224].compute_node[2][0], compute_graph.flexml_layers[224].compute_node[2][1], compute_graph.flexml_layers[224].compute_node[2][2], compute_graph.flexml_layers[224].compute_node[2][3], compute_graph.flexml_layers[224].compute_node[3][0], compute_graph.flexml_layers[224].compute_node[3][1], compute_graph.flexml_layers[224].compute_node[3][2], compute_graph.flexml_layers[224].compute_node[3][3], {compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[0], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[1], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[2], compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.out[3], compute_graph.l2_283.in[0], compute_graph.l2_283.in[1], compute_graph.l2_283.in[2], compute_graph.l2_283.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(995): mode(3), layer(309): {compute_graph.l2_283.out[0], compute_graph.l2_283.out[1], compute_graph.l2l3_283_spill.in[0], compute_graph.l2l3_283_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 996 will not be pipelined because it's in the same layer 310 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1617), layer(310): {compute_graph.ifm_ddr_1.out[2], compute_graph.l2l3_283_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1618), layer(310): {compute_graph.L2_IFM_Buffer_for_input1_1_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1619), layer(310): {compute_graph.ifm_ddr_1.out[3], compute_graph.l2l3_283_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1620), layer(310): {compute_graph.L2_IFM_Buffer_for_input1_1_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1553), layer(310): compute_graph.flexml_layers[225].compute_node[0][0], compute_graph.flexml_layers[225].compute_node[0][1], compute_graph.flexml_layers[225].compute_node[0][2], compute_graph.flexml_layers[225].compute_node[0][3], compute_graph.flexml_layers[225].compute_node[1][0], compute_graph.flexml_layers[225].compute_node[1][1], compute_graph.flexml_layers[225].compute_node[1][2], compute_graph.flexml_layers[225].compute_node[1][3], compute_graph.flexml_layers[225].compute_node[2][0], compute_graph.flexml_layers[225].compute_node[2][1], compute_graph.flexml_layers[225].compute_node[2][2], compute_graph.flexml_layers[225].compute_node[2][3], compute_graph.flexml_layers[225].compute_node[3][0], compute_graph.flexml_layers[225].compute_node[3][1], compute_graph.flexml_layers[225].compute_node[3][2], compute_graph.flexml_layers[225].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[1], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[2], compute_graph.L2_IFM_Buffer_for_input1_1_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.out[3], compute_graph.l2_284.in[0], compute_graph.l2_284.in[1], compute_graph.l2_284.in[2], compute_graph.l2_284.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(996): mode(3), layer(310): {compute_graph.l2_284.out[0], compute_graph.l2_284.out[1], compute_graph.l2l3_284_spill.in[0], compute_graph.l2l3_284_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1205): mode(0), layer(311): {compute_graph.l2l3_281_spill.out[2], compute_graph.l2l3_281_spill.out[3], compute_graph.templated_graph_285.ifm.in[0], compute_graph.templated_graph_285.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1321): mode(0), layer(311): {compute_graph.templated_graph_285.ifm.out[0], compute_graph.templated_graph_285.ifm.out[1], compute_graph.l2l3_285_spill.in[0], compute_graph.l2l3_285_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 997 will not be pipelined because it's in the same layer 312 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1621), layer(312): {compute_graph.ifm_ddr_1.out[4], compute_graph.l2l3_285_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1622), layer(312): {compute_graph.L2_IFM_Buffer_for_input1_2_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1623), layer(312): {compute_graph.ifm_ddr_1.out[5], compute_graph.l2l3_285_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1624), layer(312): {compute_graph.L2_IFM_Buffer_for_input1_2_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[1], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[2], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1554), layer(312): compute_graph.flexml_layers[226].compute_node[0][0], compute_graph.flexml_layers[226].compute_node[0][1], compute_graph.flexml_layers[226].compute_node[0][2], compute_graph.flexml_layers[226].compute_node[0][3], compute_graph.flexml_layers[226].compute_node[1][0], compute_graph.flexml_layers[226].compute_node[1][1], compute_graph.flexml_layers[226].compute_node[1][2], compute_graph.flexml_layers[226].compute_node[1][3], compute_graph.flexml_layers[226].compute_node[2][0], compute_graph.flexml_layers[226].compute_node[2][1], compute_graph.flexml_layers[226].compute_node[2][2], compute_graph.flexml_layers[226].compute_node[2][3], compute_graph.flexml_layers[226].compute_node[3][0], compute_graph.flexml_layers[226].compute_node[3][1], compute_graph.flexml_layers[226].compute_node[3][2], compute_graph.flexml_layers[226].compute_node[3][3], {compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[0], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[1], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[2], compute_graph.L2_IFM_Buffer_for_input1_2_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.out[3], compute_graph.l2_286.in[0], compute_graph.l2_286.in[1], compute_graph.l2_286.in[2], compute_graph.l2_286.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(893): mode(3), layer(313): {compute_graph.Layer_288_wts_ddr.out[0], compute_graph.Layer_288_wts_ddr.out[1], compute_graph.Layer_288_l2_wts.in[0], compute_graph.Layer_288_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(893) is pipelined with KernelLayerNode(1554) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(997): mode(3), layer(312): {compute_graph.l2_286.out[0], compute_graph.l2_286.out[1], compute_graph.spill_L3_Concat_Buffer_layer_287.in[2], compute_graph.spill_L3_Concat_Buffer_layer_287.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1212): mode(0), layer(313): {compute_graph.spill_L3_Concat_Buffer_layer_287.out[0], compute_graph.spill_L3_Concat_Buffer_layer_287.out[1], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[3], compute_graph.l2_288.in[0], compute_graph.l2_288.in[1], compute_graph.l2_288.in[2], compute_graph.l2_288.in[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1555), layer(313): compute_graph.flexml_layers[227].compute_node[0][0], compute_graph.flexml_layers[227].compute_node[0][1], compute_graph.flexml_layers[227].compute_node[0][2], compute_graph.flexml_layers[227].compute_node[0][3], compute_graph.flexml_layers[227].compute_node[1][0], compute_graph.flexml_layers[227].compute_node[1][1], compute_graph.flexml_layers[227].compute_node[1][2], compute_graph.flexml_layers[227].compute_node[1][3], compute_graph.flexml_layers[227].compute_node[2][0], compute_graph.flexml_layers[227].compute_node[2][1], compute_graph.flexml_layers[227].compute_node[2][2], compute_graph.flexml_layers[227].compute_node[2][3], compute_graph.flexml_layers[227].compute_node[3][0], compute_graph.flexml_layers[227].compute_node[3][1], compute_graph.flexml_layers[227].compute_node[3][2], compute_graph.flexml_layers[227].compute_node[3][3], {compute_graph.Layer_288_l2_wts.out[0], compute_graph.Layer_288_l2_wts.out[1], compute_graph.Layer_288_l2_wts.out[2], compute_graph.Layer_288_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.out[3], compute_graph.l2_288.in[0], compute_graph.l2_288.in[1], compute_graph.l2_288.in[2], compute_graph.l2_288.in[3]}, Scheduler computes number of sub-layer phases to be 6. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(999): mode(3), layer(313): {compute_graph.l2_288.out[0], compute_graph.l2_288.out[1], compute_graph.l2l3_288_spill.in[0], compute_graph.l2l3_288_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(999) is pipelined with KernelLayerNode(1555) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1213): mode(0), layer(314): {compute_graph.l2l3_288_spill.out[0], compute_graph.l2l3_288_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1556), layer(314): compute_graph.flexml_layers[228].compute_node[0][0], compute_graph.flexml_layers[228].compute_node[0][1], compute_graph.flexml_layers[228].compute_node[0][2], compute_graph.flexml_layers[228].compute_node[0][3], compute_graph.flexml_layers[228].compute_node[1][0], compute_graph.flexml_layers[228].compute_node[1][1], compute_graph.flexml_layers[228].compute_node[1][2], compute_graph.flexml_layers[228].compute_node[1][3], compute_graph.flexml_layers[228].compute_node[2][0], compute_graph.flexml_layers[228].compute_node[2][1], compute_graph.flexml_layers[228].compute_node[2][2], compute_graph.flexml_layers[228].compute_node[2][3], compute_graph.flexml_layers[228].compute_node[3][0], compute_graph.flexml_layers[228].compute_node[3][1], compute_graph.flexml_layers[228].compute_node[3][2], compute_graph.flexml_layers[228].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.out[3], compute_graph.l2_289.in[0], compute_graph.l2_289.in[1], compute_graph.l2_289.in[2], compute_graph.l2_289.in[3]}, Scheduler computes number of sub-layer phases to be 4. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1000): mode(0), layer(314): {compute_graph.l2_289.out[0], compute_graph.l2_289.out[1], compute_graph.l2l3_289_spill.in[0], compute_graph.l2l3_289_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 1001 will not be pipelined because it's in the same layer 315 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1625), layer(315): {compute_graph.l2l3_282_spill.out[2], compute_graph.l2l3_289_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1626), layer(315): {compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1627), layer(315): {compute_graph.l2l3_282_spill.out[3], compute_graph.l2l3_289_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1628), layer(315): {compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1557), layer(315): compute_graph.flexml_layers[229].compute_node[0][0], compute_graph.flexml_layers[229].compute_node[0][1], compute_graph.flexml_layers[229].compute_node[0][2], compute_graph.flexml_layers[229].compute_node[0][3], compute_graph.flexml_layers[229].compute_node[1][0], compute_graph.flexml_layers[229].compute_node[1][1], compute_graph.flexml_layers[229].compute_node[1][2], compute_graph.flexml_layers[229].compute_node[1][3], compute_graph.flexml_layers[229].compute_node[2][0], compute_graph.flexml_layers[229].compute_node[2][1], compute_graph.flexml_layers[229].compute_node[2][2], compute_graph.flexml_layers[229].compute_node[2][3], compute_graph.flexml_layers[229].compute_node[3][0], compute_graph.flexml_layers[229].compute_node[3][1], compute_graph.flexml_layers[229].compute_node[3][2], compute_graph.flexml_layers[229].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.out[3], compute_graph.l2_290.in[0], compute_graph.l2_290.in[1], compute_graph.l2_290.in[2], compute_graph.l2_290.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1001): mode(3), layer(315): {compute_graph.l2_290.out[0], compute_graph.l2_290.out[1], compute_graph.l2l3_290_spill.in[0], compute_graph.l2l3_290_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1629), layer(316): {compute_graph.l2l3_284_spill.out[0], compute_graph.l2l3_290_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1630), layer(316): {compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1631), layer(316): {compute_graph.l2l3_284_spill.out[1], compute_graph.l2l3_290_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1632), layer(316): {compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.in[1], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1558), layer(316): compute_graph.flexml_layers[230].compute_node[0][0], compute_graph.flexml_layers[230].compute_node[0][1], compute_graph.flexml_layers[230].compute_node[0][2], compute_graph.flexml_layers[230].compute_node[0][3], compute_graph.flexml_layers[230].compute_node[1][0], compute_graph.flexml_layers[230].compute_node[1][1], compute_graph.flexml_layers[230].compute_node[1][2], compute_graph.flexml_layers[230].compute_node[1][3], compute_graph.flexml_layers[230].compute_node[2][0], compute_graph.flexml_layers[230].compute_node[2][1], compute_graph.flexml_layers[230].compute_node[2][2], compute_graph.flexml_layers[230].compute_node[2][3], compute_graph.flexml_layers[230].compute_node[3][0], compute_graph.flexml_layers[230].compute_node[3][1], compute_graph.flexml_layers[230].compute_node[3][2], compute_graph.flexml_layers[230].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.out[3], compute_graph.l2_291.in[0], compute_graph.l2_291.in[1], compute_graph.l2_291.in[2], compute_graph.l2_291.in[3]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1633), layer(316): {compute_graph.l2_291.out[0], compute_graph.l2_291.out[2]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1634), layer(316): {compute_graph.ofm_ddr_3_l2l3_291_spill.in[0], compute_graph.spill_L3_Concat_Buffer_layer_292.in[2]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1635), layer(316): {compute_graph.l2_291.out[1], compute_graph.l2_291.out[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1636), layer(316): {compute_graph.ofm_ddr_3_l2l3_291_spill.in[1], compute_graph.spill_L3_Concat_Buffer_layer_292.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1196): mode(0), layer(317): {compute_graph.l2l3_277_spill.out[0], compute_graph.l2l3_277_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1004): mode(0), layer(317): {compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.out[1], compute_graph.spill_L3_Concat_Buffer_layer_292.in[0], compute_graph.spill_L3_Concat_Buffer_layer_292.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 1322 will not be pipelined because it's in the same layer 318 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.spill_L3_Concat_Buffer_layer_292.out[0], compute_graph.spill_L3_Concat_Buffer_layer_292.out[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1216): mode(0), layer(318): {compute_graph.spill_L3_Concat_Buffer_layer_292.out[0], compute_graph.spill_L3_Concat_Buffer_layer_292.out[1], compute_graph.templated_graph_293.ifm_mem.in[0], compute_graph.templated_graph_293.ifm_mem.in[1]}, Scheduler computes number of sub-layer phases to be 23. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1323), layer(318): compute_graph.templated_graph_293.mk[0][0], compute_graph.templated_graph_293.mk[0][1], compute_graph.templated_graph_293.mk[0][2], compute_graph.templated_graph_293.mk[0][3], compute_graph.templated_graph_293.mk[1][0], compute_graph.templated_graph_293.mk[1][1], compute_graph.templated_graph_293.mk[1][2], compute_graph.templated_graph_293.mk[1][3], compute_graph.templated_graph_293.mk[2][0], compute_graph.templated_graph_293.mk[2][1], compute_graph.templated_graph_293.mk[2][2], compute_graph.templated_graph_293.mk[2][3], compute_graph.templated_graph_293.mk[3][0], compute_graph.templated_graph_293.mk[3][1], compute_graph.templated_graph_293.mk[3][2], compute_graph.templated_graph_293.mk[3][3], {compute_graph.templated_graph_293.ifm_mem.out[0], compute_graph.templated_graph_293.ifm_mem.out[1], compute_graph.templated_graph_293.ifm_mem.out[2], compute_graph.templated_graph_293.ifm_mem.out[3], compute_graph.templated_graph_293.ofm_mem.in[0], compute_graph.templated_graph_293.ofm_mem.in[1], compute_graph.templated_graph_293.ofm_mem.in[2], compute_graph.templated_graph_293.ofm_mem.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(894): mode(3), layer(319): {compute_graph.Layer_295_wts_ddr.out[0], compute_graph.Layer_295_wts_ddr.out[1], compute_graph.Layer_295_l2_wts.in[0], compute_graph.Layer_295_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(894) is pipelined with KernelLayerNode(1323) +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.templated_graph_293.ofm_mem.out[0], compute_graph.templated_graph_293.ofm_mem.out[1], compute_graph.spill_L3_Concat_Buffer_layer_294.in[0], compute_graph.spill_L3_Concat_Buffer_layer_294.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1322): mode(3), layer(318): {compute_graph.templated_graph_293.ofm_mem.out[0], compute_graph.templated_graph_293.ofm_mem.out[1], compute_graph.spill_L3_Concat_Buffer_layer_294.in[0], compute_graph.spill_L3_Concat_Buffer_layer_294.in[1]}, Scheduler computes number of sub-layer phases to be 23. +INFO: [aiecompiler 77-23129] BufferToBufferNode 1006 will not be pipelined because it's in the same layer 319 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.spill_L3_Concat_Buffer_layer_294.out[0], compute_graph.spill_L3_Concat_Buffer_layer_294.out[1], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1217): mode(0), layer(319): {compute_graph.spill_L3_Concat_Buffer_layer_294.out[0], compute_graph.spill_L3_Concat_Buffer_layer_294.out[1], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[1]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1559), layer(319): compute_graph.flexml_layers[231].compute_node[0][0], compute_graph.flexml_layers[231].compute_node[0][1], compute_graph.flexml_layers[231].compute_node[0][2], compute_graph.flexml_layers[231].compute_node[0][3], compute_graph.flexml_layers[231].compute_node[1][0], compute_graph.flexml_layers[231].compute_node[1][1], compute_graph.flexml_layers[231].compute_node[1][2], compute_graph.flexml_layers[231].compute_node[1][3], compute_graph.flexml_layers[231].compute_node[2][0], compute_graph.flexml_layers[231].compute_node[2][1], compute_graph.flexml_layers[231].compute_node[2][2], compute_graph.flexml_layers[231].compute_node[2][3], compute_graph.flexml_layers[231].compute_node[3][0], compute_graph.flexml_layers[231].compute_node[3][1], compute_graph.flexml_layers[231].compute_node[3][2], compute_graph.flexml_layers[231].compute_node[3][3], {compute_graph.Layer_295_l2_wts.out[0], compute_graph.Layer_295_l2_wts.out[1], compute_graph.Layer_295_l2_wts.out[2], compute_graph.Layer_295_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.out[3], compute_graph.l2_295.in[0], compute_graph.l2_295.in[1], compute_graph.l2_295.in[2], compute_graph.l2_295.in[3]}, Scheduler computes number of sub-layer phases to be 6. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(895): mode(3), layer(320): {compute_graph.Layer_296_wts_ddr.out[0], compute_graph.Layer_296_wts_ddr.out[1], compute_graph.Layer_296_l2_wts.in[0], compute_graph.Layer_296_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(895) is pipelined with KernelLayerNode(1559) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1006): mode(3), layer(319): {compute_graph.l2_295.out[0], compute_graph.l2_295.out[1], compute_graph.l2l3_295_spill.in[0], compute_graph.l2l3_295_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-23129] BufferToBufferNode 1007 will not be pipelined because it's in the same layer 320 in interleaving mode +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.l2l3_295_spill.out[0], compute_graph.l2l3_295_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[1], +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1218): mode(0), layer(320): {compute_graph.l2l3_295_spill.out[0], compute_graph.l2l3_295_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[1]}, Scheduler computes number of sub-layer phases to be 3. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1560), layer(320): compute_graph.flexml_layers[232].compute_node[0][0], compute_graph.flexml_layers[232].compute_node[0][1], compute_graph.flexml_layers[232].compute_node[0][2], compute_graph.flexml_layers[232].compute_node[0][3], compute_graph.flexml_layers[232].compute_node[1][0], compute_graph.flexml_layers[232].compute_node[1][1], compute_graph.flexml_layers[232].compute_node[1][2], compute_graph.flexml_layers[232].compute_node[1][3], compute_graph.flexml_layers[232].compute_node[2][0], compute_graph.flexml_layers[232].compute_node[2][1], compute_graph.flexml_layers[232].compute_node[2][2], compute_graph.flexml_layers[232].compute_node[2][3], compute_graph.flexml_layers[232].compute_node[3][0], compute_graph.flexml_layers[232].compute_node[3][1], compute_graph.flexml_layers[232].compute_node[3][2], compute_graph.flexml_layers[232].compute_node[3][3], {compute_graph.Layer_296_l2_wts.out[0], compute_graph.Layer_296_l2_wts.out[1], compute_graph.Layer_296_l2_wts.out[2], compute_graph.Layer_296_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.out[3], compute_graph.l2_296.in[0], compute_graph.l2_296.in[1], compute_graph.l2_296.in[2], compute_graph.l2_296.in[3]}, Scheduler computes number of sub-layer phases to be 9. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(896): mode(3), layer(321): {compute_graph.Layer_297_wts_ddr.out[0], compute_graph.Layer_297_wts_ddr.out[1], compute_graph.Layer_297_l2_wts.in[0], compute_graph.Layer_297_l2_wts.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(896) is pipelined with KernelLayerNode(1560) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1007): mode(3), layer(320): {compute_graph.l2_296.out[0], compute_graph.l2_296.out[1], compute_graph.l2l3_296_spill.in[0], compute_graph.l2l3_296_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1219): mode(0), layer(321): {compute_graph.l2l3_296_spill.out[0], compute_graph.l2l3_296_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1561), layer(321): compute_graph.flexml_layers[233].compute_node[0][0], compute_graph.flexml_layers[233].compute_node[0][1], compute_graph.flexml_layers[233].compute_node[0][2], compute_graph.flexml_layers[233].compute_node[0][3], compute_graph.flexml_layers[233].compute_node[1][0], compute_graph.flexml_layers[233].compute_node[1][1], compute_graph.flexml_layers[233].compute_node[1][2], compute_graph.flexml_layers[233].compute_node[1][3], compute_graph.flexml_layers[233].compute_node[2][0], compute_graph.flexml_layers[233].compute_node[2][1], compute_graph.flexml_layers[233].compute_node[2][2], compute_graph.flexml_layers[233].compute_node[2][3], compute_graph.flexml_layers[233].compute_node[3][0], compute_graph.flexml_layers[233].compute_node[3][1], compute_graph.flexml_layers[233].compute_node[3][2], compute_graph.flexml_layers[233].compute_node[3][3], {compute_graph.Layer_297_l2_wts.out[0], compute_graph.Layer_297_l2_wts.out[1], compute_graph.Layer_297_l2_wts.out[2], compute_graph.Layer_297_l2_wts.out[3], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.out[3], compute_graph.l2_297.in[0], compute_graph.l2_297.in[1], compute_graph.l2_297.in[2], compute_graph.l2_297.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1008): mode(3), layer(321): {compute_graph.l2_297.out[0], compute_graph.l2_297.out[1], compute_graph.l2l3_297_spill.in[0], compute_graph.l2l3_297_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1008) is pipelined with KernelLayerNode(1561) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1220): mode(0), layer(322): {compute_graph.l2l3_297_spill.out[0], compute_graph.l2l3_297_spill.out[1], compute_graph.templated_graph_298.ifm.in[0], compute_graph.templated_graph_298.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1325), layer(322): compute_graph.templated_graph_298.compute_node[0][0], compute_graph.templated_graph_298.compute_node[0][1], compute_graph.templated_graph_298.compute_node[0][2], compute_graph.templated_graph_298.compute_node[0][3], compute_graph.templated_graph_298.compute_node[1][0], compute_graph.templated_graph_298.compute_node[1][1], compute_graph.templated_graph_298.compute_node[1][2], compute_graph.templated_graph_298.compute_node[1][3], compute_graph.templated_graph_298.compute_node[2][0], compute_graph.templated_graph_298.compute_node[2][1], compute_graph.templated_graph_298.compute_node[2][2], compute_graph.templated_graph_298.compute_node[2][3], compute_graph.templated_graph_298.compute_node[3][0], compute_graph.templated_graph_298.compute_node[3][1], compute_graph.templated_graph_298.compute_node[3][2], compute_graph.templated_graph_298.compute_node[3][3], {compute_graph.templated_graph_298.ifm.out[0], compute_graph.templated_graph_298.ifm.out[1], compute_graph.templated_graph_298.ifm.out[2], compute_graph.templated_graph_298.ifm.out[3], compute_graph.templated_graph_298.ofm.in[0], compute_graph.templated_graph_298.ofm.in[1], compute_graph.templated_graph_298.ofm.in[2], compute_graph.templated_graph_298.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1324): mode(3), layer(322): {compute_graph.templated_graph_298.ofm.out[0], compute_graph.templated_graph_298.ofm.out[1], compute_graph.l2l3_298_spill.in[0], compute_graph.l2l3_298_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1324) is pipelined with KernelLayerNode(1325) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1222): mode(0), layer(323): {compute_graph.l2l3_298_spill.out[0], compute_graph.l2l3_298_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1562), layer(323): compute_graph.flexml_layers[234].compute_node[0][0], compute_graph.flexml_layers[234].compute_node[0][1], compute_graph.flexml_layers[234].compute_node[0][2], compute_graph.flexml_layers[234].compute_node[0][3], compute_graph.flexml_layers[234].compute_node[1][0], compute_graph.flexml_layers[234].compute_node[1][1], compute_graph.flexml_layers[234].compute_node[1][2], compute_graph.flexml_layers[234].compute_node[1][3], compute_graph.flexml_layers[234].compute_node[2][0], compute_graph.flexml_layers[234].compute_node[2][1], compute_graph.flexml_layers[234].compute_node[2][2], compute_graph.flexml_layers[234].compute_node[2][3], compute_graph.flexml_layers[234].compute_node[3][0], compute_graph.flexml_layers[234].compute_node[3][1], compute_graph.flexml_layers[234].compute_node[3][2], compute_graph.flexml_layers[234].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.out[3], compute_graph.L2_OFM_Buffer_layer_299.in[0], compute_graph.L2_OFM_Buffer_layer_299.in[1], compute_graph.L2_OFM_Buffer_layer_299.in[2], compute_graph.L2_OFM_Buffer_layer_299.in[3]}, Scheduler computes number of sub-layer phases to be 18. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1009): mode(3), layer(323): {compute_graph.L2_OFM_Buffer_layer_299.out[0], compute_graph.L2_OFM_Buffer_layer_299.out[1], compute_graph.ofm_ddr_4.in[0], compute_graph.ofm_ddr_4.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1009) is pipelined with KernelLayerNode(1562) +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1221): mode(0), layer(324): {compute_graph.l2l3_297_spill.out[2], compute_graph.l2l3_297_spill.out[3], compute_graph.templated_graph_300.ifm.in[0], compute_graph.templated_graph_300.ifm.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For KernelLayerNode(1327), layer(324): compute_graph.templated_graph_300.compute_node[0][0], compute_graph.templated_graph_300.compute_node[0][1], compute_graph.templated_graph_300.compute_node[0][2], compute_graph.templated_graph_300.compute_node[0][3], compute_graph.templated_graph_300.compute_node[1][0], compute_graph.templated_graph_300.compute_node[1][1], compute_graph.templated_graph_300.compute_node[1][2], compute_graph.templated_graph_300.compute_node[1][3], compute_graph.templated_graph_300.compute_node[2][0], compute_graph.templated_graph_300.compute_node[2][1], compute_graph.templated_graph_300.compute_node[2][2], compute_graph.templated_graph_300.compute_node[2][3], compute_graph.templated_graph_300.compute_node[3][0], compute_graph.templated_graph_300.compute_node[3][1], compute_graph.templated_graph_300.compute_node[3][2], compute_graph.templated_graph_300.compute_node[3][3], {compute_graph.templated_graph_300.ifm.out[0], compute_graph.templated_graph_300.ifm.out[1], compute_graph.templated_graph_300.ifm.out[2], compute_graph.templated_graph_300.ifm.out[3], compute_graph.templated_graph_300.ofm.in[0], compute_graph.templated_graph_300.ofm.in[1], compute_graph.templated_graph_300.ofm.in[2], compute_graph.templated_graph_300.ofm.in[3]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1326): mode(3), layer(324): {compute_graph.templated_graph_300.ofm.out[0], compute_graph.templated_graph_300.ofm.out[1], compute_graph.l2l3_300_spill.in[0], compute_graph.l2l3_300_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1326) is pipelined with KernelLayerNode(1327) +INFO: [aiecompiler 77-23129] BufferToBufferNode 1010 will not be pipelined because it's in the same layer 325 in interleaving mode +INFO: [aiecompiler 77-6295] For BufferSenderNode(1637), layer(325): {compute_graph.l2l3_5_spill.out[6], compute_graph.l2l3_300_spill.out[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1638), layer(325): {compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.in[0], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.in[0]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferSenderNode(1639), layer(325): {compute_graph.l2l3_5_spill.out[7], compute_graph.l2l3_300_spill.out[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferReceiverNode(1640), layer(325): {compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.in[1], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[3], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1563), layer(325): compute_graph.flexml_layers[235].compute_node[0][0], compute_graph.flexml_layers[235].compute_node[0][1], compute_graph.flexml_layers[235].compute_node[0][2], compute_graph.flexml_layers[235].compute_node[0][3], compute_graph.flexml_layers[235].compute_node[1][0], compute_graph.flexml_layers[235].compute_node[1][1], compute_graph.flexml_layers[235].compute_node[1][2], compute_graph.flexml_layers[235].compute_node[1][3], compute_graph.flexml_layers[235].compute_node[2][0], compute_graph.flexml_layers[235].compute_node[2][1], compute_graph.flexml_layers[235].compute_node[2][2], compute_graph.flexml_layers[235].compute_node[2][3], compute_graph.flexml_layers[235].compute_node[3][0], compute_graph.flexml_layers[235].compute_node[3][1], compute_graph.flexml_layers[235].compute_node[3][2], compute_graph.flexml_layers[235].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[0], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[1], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[2], compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.out[3], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.out[3], compute_graph.l2_301.in[0], compute_graph.l2_301.in[1], compute_graph.l2_301.in[2], compute_graph.l2_301.in[3]}, Scheduler computes number of sub-layer phases to be 9. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1010): mode(3), layer(325): {compute_graph.l2_301.out[0], compute_graph.l2_301.out[1], compute_graph.l2l3_301_spill.in[0], compute_graph.l2l3_301_spill.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1224): mode(0), layer(326): {compute_graph.l2l3_301_spill.out[0], compute_graph.l2l3_301_spill.out[1], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.in[0], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22166] Auto-phase process starts for compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[3], +INFO: [aiecompiler 77-6295] For KernelLayerNode(1564), layer(326): compute_graph.flexml_layers[236].compute_node[0][0], compute_graph.flexml_layers[236].compute_node[0][1], compute_graph.flexml_layers[236].compute_node[0][2], compute_graph.flexml_layers[236].compute_node[0][3], compute_graph.flexml_layers[236].compute_node[1][0], compute_graph.flexml_layers[236].compute_node[1][1], compute_graph.flexml_layers[236].compute_node[1][2], compute_graph.flexml_layers[236].compute_node[1][3], compute_graph.flexml_layers[236].compute_node[2][0], compute_graph.flexml_layers[236].compute_node[2][1], compute_graph.flexml_layers[236].compute_node[2][2], compute_graph.flexml_layers[236].compute_node[2][3], compute_graph.flexml_layers[236].compute_node[3][0], compute_graph.flexml_layers[236].compute_node[3][1], compute_graph.flexml_layers[236].compute_node[3][2], compute_graph.flexml_layers[236].compute_node[3][3], {compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[0], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[1], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[2], compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.out[3], compute_graph.L2_OFM_Buffer_layer_302.in[0], compute_graph.L2_OFM_Buffer_layer_302.in[1], compute_graph.L2_OFM_Buffer_layer_302.in[2], compute_graph.L2_OFM_Buffer_layer_302.in[3]}, Scheduler computes number of sub-layer phases to be 18. +INFO: [aiecompiler 77-6295] For BufferToBufferNode(1011): mode(3), layer(326): {compute_graph.L2_OFM_Buffer_layer_302.out[0], compute_graph.L2_OFM_Buffer_layer_302.out[1], compute_graph.ofm_ddr_5.in[0], compute_graph.ofm_ddr_5.in[1]}, Scheduler computes number of sub-layer phases to be 1. +INFO: [aiecompiler 77-22492] BufferToBufferNode(1011) is pipelined with KernelLayerNode(1564) +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-22492' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[8].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[18].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[19].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[28].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[29].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[38].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[39].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[100].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[101].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[118].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[119].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[136].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[137].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[154].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[155].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[172].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[173].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[185].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[186].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.templated_graph_231.mk[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[203].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.templated_graph_254.mk[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.flexml_layers[213].compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.templated_graph_274.mk[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 0) automatic PM reload boundary inserted at: compute_graph.templated_graph_298.compute_node[0][0] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[8].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[18].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[19].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[28].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[29].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[38].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[39].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[100].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[101].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[118].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[119].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[136].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[137].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[154].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[155].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[172].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[173].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[185].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[186].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.templated_graph_231.mk[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[203].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.templated_graph_254.mk[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.flexml_layers[213].compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.templated_graph_274.mk[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 1) automatic PM reload boundary inserted at: compute_graph.templated_graph_298.compute_node[0][1] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[8].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[18].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[19].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[28].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[29].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[38].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[39].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[100].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[101].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[118].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[119].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[136].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[137].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[154].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[155].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[172].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[173].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[185].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[186].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.templated_graph_231.mk[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[203].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.templated_graph_254.mk[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.flexml_layers[213].compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.templated_graph_274.mk[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 2) automatic PM reload boundary inserted at: compute_graph.templated_graph_298.compute_node[0][2] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[8].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[18].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[19].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[28].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[29].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[38].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[39].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[100].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[101].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[118].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[119].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[136].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[137].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[154].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[155].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[172].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[173].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[185].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[186].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.templated_graph_231.mk[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[203].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.templated_graph_254.mk[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.flexml_layers[213].compute_node[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.templated_graph_274.mk[0][3] +INFO: [aiecompiler 77-22549] For tile:(0, 3) automatic PM reload boundary inserted at: compute_graph.templated_graph_298.compute_node[0][3] +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-22549' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-6272] Completing Scheduler pass +INFO: [aiecompiler 77-23111] Writing layer control parameter JSON file Work/aie/layer_control_parameters.json +INFO: [aiecompiler 77-6593] Max LCP size in 32-bit words: 225 +INFO: [aiecompiler 77-6497] runtime-opt stats: Avoided compilation of 418 cores out of 432 cores. +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0/scripts/0_0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1/scripts/0_1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2/scripts/0_2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3/scripts/0_3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (1,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/1_0/scripts/1_0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/1_0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (1,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/1_1/scripts/1_1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/1_1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (1,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/1_2/scripts/1_2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/1_2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (1,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/1_3/scripts/1_3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/1_3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (2,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/2_0/scripts/2_0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/2_0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (2,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/2_1/scripts/2_1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/2_1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (2,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/2_2/scripts/2_2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/2_2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (2,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/2_3/scripts/2_3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/2_3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (3,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/3_0/scripts/3_0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/3_0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (3,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/3_1/scripts/3_1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/3_1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (3,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/3_2/scripts/3_2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/3_2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (3,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/3_3/scripts/3_3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/3_3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable0/scripts/0_0_reloadable0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable1/scripts/0_0_reloadable1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable2/scripts/0_0_reloadable2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable3/scripts/0_0_reloadable3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable4/scripts/0_0_reloadable4.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable4/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable5/scripts/0_0_reloadable5.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable5/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable6/scripts/0_0_reloadable6.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable6/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable7/scripts/0_0_reloadable7.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable7/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable8/scripts/0_0_reloadable8.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable8/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable9/scripts/0_0_reloadable9.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable9/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable10/scripts/0_0_reloadable10.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable10/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable11/scripts/0_0_reloadable11.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable11/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable12/scripts/0_0_reloadable12.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable12/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable13/scripts/0_0_reloadable13.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable13/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable14/scripts/0_0_reloadable14.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable14/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable15/scripts/0_0_reloadable15.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable15/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable16/scripts/0_0_reloadable16.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable16/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable17/scripts/0_0_reloadable17.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable17/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable18/scripts/0_0_reloadable18.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable18/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable19/scripts/0_0_reloadable19.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable19/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable20/scripts/0_0_reloadable20.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable20/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable21/scripts/0_0_reloadable21.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable21/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable22/scripts/0_0_reloadable22.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable22/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable23/scripts/0_0_reloadable23.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable23/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable24/scripts/0_0_reloadable24.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable24/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,0) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_0_reloadable25/scripts/0_0_reloadable25.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_0_reloadable25/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable0/scripts/0_1_reloadable0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable1/scripts/0_1_reloadable1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable2/scripts/0_1_reloadable2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable3/scripts/0_1_reloadable3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable4/scripts/0_1_reloadable4.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable4/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable5/scripts/0_1_reloadable5.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable5/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable6/scripts/0_1_reloadable6.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable6/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable7/scripts/0_1_reloadable7.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable7/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable8/scripts/0_1_reloadable8.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable8/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable9/scripts/0_1_reloadable9.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable9/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable10/scripts/0_1_reloadable10.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable10/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable11/scripts/0_1_reloadable11.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable11/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable12/scripts/0_1_reloadable12.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable12/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable13/scripts/0_1_reloadable13.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable13/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable14/scripts/0_1_reloadable14.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable14/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable15/scripts/0_1_reloadable15.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable15/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable16/scripts/0_1_reloadable16.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable16/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable17/scripts/0_1_reloadable17.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable17/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable18/scripts/0_1_reloadable18.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable18/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable19/scripts/0_1_reloadable19.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable19/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable20/scripts/0_1_reloadable20.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable20/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable21/scripts/0_1_reloadable21.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable21/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable22/scripts/0_1_reloadable22.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable22/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable23/scripts/0_1_reloadable23.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable23/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable24/scripts/0_1_reloadable24.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable24/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,1) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_1_reloadable25/scripts/0_1_reloadable25.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_1_reloadable25/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable0/scripts/0_2_reloadable0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable1/scripts/0_2_reloadable1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable2/scripts/0_2_reloadable2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable3/scripts/0_2_reloadable3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable4/scripts/0_2_reloadable4.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable4/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable5/scripts/0_2_reloadable5.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable5/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable6/scripts/0_2_reloadable6.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable6/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable7/scripts/0_2_reloadable7.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable7/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable8/scripts/0_2_reloadable8.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable8/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable9/scripts/0_2_reloadable9.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable9/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable10/scripts/0_2_reloadable10.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable10/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable11/scripts/0_2_reloadable11.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable11/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable12/scripts/0_2_reloadable12.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable12/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable13/scripts/0_2_reloadable13.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable13/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable14/scripts/0_2_reloadable14.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable14/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable15/scripts/0_2_reloadable15.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable15/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable16/scripts/0_2_reloadable16.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable16/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable17/scripts/0_2_reloadable17.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable17/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable18/scripts/0_2_reloadable18.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable18/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable19/scripts/0_2_reloadable19.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable19/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable20/scripts/0_2_reloadable20.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable20/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable21/scripts/0_2_reloadable21.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable21/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable22/scripts/0_2_reloadable22.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable22/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable23/scripts/0_2_reloadable23.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable23/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable24/scripts/0_2_reloadable24.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable24/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,2) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_2_reloadable25/scripts/0_2_reloadable25.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_2_reloadable25/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable0/scripts/0_3_reloadable0.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3_reloadable0/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable1/scripts/0_3_reloadable1.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3_reloadable1/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable2/scripts/0_3_reloadable2.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3_reloadable2/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable3/scripts/0_3_reloadable3.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3_reloadable3/timestamped_log +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable4/scripts/0_3_reloadable4.bcf +INFO: [aiecompiler 77-21796] ### Created directory with path = Work/aie/0_3_reloadable4/timestamped_log +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-21796' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-6298] Generating CHESS Project File for processor at (0,3) +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-6298' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-6299] Generating Linker script in Work/aie/0_3_reloadable5/scripts/0_3_reloadable5.bcf +INFO: [aiecompiler 17-14] Message 'aiecompiler 77-6299' appears 100 times and further instances of the messages will be disabled. Use the Tcl command set_msg_config to change the current settings. +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/src/0_0.cc -o ir/0_0_orig.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0/src/0_0.cc -oir/0_0_orig.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_orig.ll -o ir/0_0.ll 2> 0_0/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0.ll -o ir/0_0.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable0.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable0/src/0_0_reloadable0.cc -o ir/0_0_reloadable0_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable0.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable0/src/0_0_reloadable0.cc -oir/0_0_reloadable0_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable0/src/0_0_reloadable0.cc:10: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/transpose4d_adf_wrapper.cpp:68:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 68 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +1 warning generated. +Compilation finished successfully (1 errors, 1 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable0_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable0_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable0_orig.ll -o ir/0_0_reloadable0.ll 2> 0_0_reloadable0/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable0.ll -o ir/0_0_reloadable0.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable1.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable1/src/0_0_reloadable1.cc -o ir/0_0_reloadable1_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable1.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable1/src/0_0_reloadable1.cc -oir/0_0_reloadable1_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable1_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable1_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable1_orig.ll -o ir/0_0_reloadable1.ll 2> 0_0_reloadable1/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable1.ll -o ir/0_0_reloadable1.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable2.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable2/src/0_0_reloadable2.cc -o ir/0_0_reloadable2_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable2.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable2/src/0_0_reloadable2.cc -oir/0_0_reloadable2_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable2_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable2_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable2_orig.ll -o ir/0_0_reloadable2.ll 2> 0_0_reloadable2/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable2.ll -o ir/0_0_reloadable2.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable3.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable3/src/0_0_reloadable3.cc -o ir/0_0_reloadable3_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable3.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable3/src/0_0_reloadable3.cc -oir/0_0_reloadable3_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable3_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable3_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable3_orig.ll -o ir/0_0_reloadable3.ll 2> 0_0_reloadable3/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable3.ll -o ir/0_0_reloadable3.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable5.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable5/src/0_0_reloadable5.cc -o ir/0_0_reloadable5_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable5.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable5/src/0_0_reloadable5.cc -oir/0_0_reloadable5_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable5_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable5_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable5_orig.ll -o ir/0_0_reloadable5.ll 2> 0_0_reloadable5/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable5.ll -o ir/0_0_reloadable5.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable17.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable17/src/0_0_reloadable17.cc -o ir/0_0_reloadable17_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable17.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable17/src/0_0_reloadable17.cc -oir/0_0_reloadable17_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable17_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable17_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable17_orig.ll -o ir/0_0_reloadable17.ll 2> 0_0_reloadable17/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable17.ll -o ir/0_0_reloadable17.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable19.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable19/src/0_0_reloadable19.cc -o ir/0_0_reloadable19_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable19.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable19/src/0_0_reloadable19.cc -oir/0_0_reloadable19_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable19_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable19_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable19_orig.ll -o ir/0_0_reloadable19.ll 2> 0_0_reloadable19/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable19.ll -o ir/0_0_reloadable19.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable20.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable20/src/0_0_reloadable20.cc -o ir/0_0_reloadable20_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable20.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable20/src/0_0_reloadable20.cc -oir/0_0_reloadable20_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable20/src/0_0_reloadable20.cc:9: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/resize_adf_wrapper.cpp:70:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 70 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +1 warning generated. +Compilation finished successfully (1 errors, 1 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable20_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable20_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable20_orig.ll -o ir/0_0_reloadable20.ll 2> 0_0_reloadable20/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable20.ll -o ir/0_0_reloadable20.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable21.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable21/src/0_0_reloadable21.cc -o ir/0_0_reloadable21_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable21.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable21/src/0_0_reloadable21.cc -oir/0_0_reloadable21_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable21_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable21_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable21_orig.ll -o ir/0_0_reloadable21.ll 2> 0_0_reloadable21/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable21.ll -o ir/0_0_reloadable21.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable22.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc -o ir/0_0_reloadable22_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable22.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc -oir/0_0_reloadable22_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:9: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/resize_adf_wrapper.cpp:70:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 70 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:66: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d.h:66: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:111:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 111 | #define LAYER_PARAM_BYTES 0 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw_params.h:70:9: note: previous definition is here + 70 | #define LAYER_PARAM_BYTES 64 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:66: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d.h:66: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:1200:9: warning: 'WT_DM_BANK' macro redefined [-Wmacro-redefined] + 1200 | #define WT_DM_BANK __aie_dm_resource_b + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:76:9: note: previous definition is here + 76 | #define WT_DM_BANK __aie_dm_resource_a + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:69: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw.h:98:13: warning: 'LOAD_64' macro redefined [-Wmacro-redefined] + 98 | #define LOAD_64(ifm,ibuff,ifm_incr) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:98:13: note: previous definition is here + 98 | #define LOAD_64(ifm,ibuff) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:71: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16.h:227:13: warning: 'LOAD_32' macro redefined [-Wmacro-redefined] + 227 | #define LOAD_32(ifm,ibuff) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:365:13: note: previous definition is here + 365 | #define LOAD_32(ifm,ibuff) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:71: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16.h:231:13: warning: 'VMUL_W4C8' macro redefined [-Wmacro-redefined] + 231 | #define VMUL_W4C8(ifm,wts,acc) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:370:13: note: previous definition is here + 370 | #define VMUL_W4C8(ifm,wts,acc) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:73: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_transpose.h:67: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_transpose_params.h:76:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 76 | #define LAYER_PARAM_BYTES 64 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16_params.h:69:9: note: previous definition is here + 69 | #define LAYER_PARAM_BYTES 0 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:746:12: warning: 'clr64' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 746 | return clr64(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331425:7: note: 'clr64' has been explicitly marked deprecated here + 331425 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64acc32 clr64(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:749:12: warning: 'clr32' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 749 | return clr32(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331508:7: note: 'clr32' has been explicitly marked deprecated here + 331508 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v32acc64 clr32(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:752:40: warning: 'clr64f' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 752 | v64accfloat chess_storage(dm4) n = clr64f(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331757:7: note: 'clr64f' has been explicitly marked deprecated here + 331757 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64accfloat clr64f(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:757:12: warning: 'clr64f' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 757 | return clr64f(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331757:7: note: 'clr64f' has been explicitly marked deprecated here + 331757 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64accfloat clr64f(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:167:9: warning: 'CHESS_STORAGE' macro redefined [-Wmacro-redefined] + 167 | #define CHESS_STORAGE( T, accs ) { \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:1020:9: note: previous definition is here + 1020 | #define CHESS_STORAGE(T, accs) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:481:13: warning: 'LOAD_MAT_B_DATA_2_8x8' macro redefined [-Wmacro-redefined] + 481 | #define LOAD_MAT_B_DATA_2_8x8(p_mat_b, iterator_mat_b, param, Y0, Y1)\ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:414:13: note: previous definition is here + 414 | #define LOAD_MAT_B_DATA_2_8x8(p_mat_b, iterator_mat_b, param, Y0, Y1, Y2, Y3)\ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:600:13: warning: 'STORE_ACC_8_4x8_FULL_DATA' macro redefined [-Wmacro-redefined] + 600 | #define STORE_ACC_8_4x8_FULL_DATA(p_mat_c, iterator_mat_c, param, C0, C1, C2, C3, C4, C5, C6, C7, p_bias_in) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:474:13: note: previous definition is here + 474 | #define STORE_ACC_8_4x8_FULL_DATA(p_mat_c, C0, C1, C2, C3) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:637:13: warning: 'MMUL_MAC_8_4x8x8' macro redefined [-Wmacro-redefined] + 637 | #define MMUL_MAC_8_4x8x8(C0, C1, C2, C3, C4, C5, C6, C7, X0, X1, X2, X3, Y0, Y1)\ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:504:13: note: previous definition is here + 504 | #define MMUL_MAC_8_4x8x8(C0, C1, C2, C3, X0, Y0, Y1, Y2, Y3)\ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:110: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/mul2d.h:58: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/mul2d_params.h:59:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 59 | #define LAYER_PARAM_BYTES 64 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_params.h:79:9: note: previous definition is here + 79 | #define LAYER_PARAM_BYTES 0 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:148: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/gelu1d.h:56: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/gelu1d_impl.h:197:59: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 197 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:151: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/tanh.h:54: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/tanh_impl.h:280:59: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 280 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:218: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/floor.h:57: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/floor_impl.h:121:55: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 121 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:236: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops.h:51: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:155:17: warning: anonymous non-C-compatible type given name for linkage purposes by typedef declaration; add a tag name here [-Wnon-c-typedef-for-linkage] + 155 | typedef struct alignas(v32uint16) + | ^ + | params_t +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:161:5: note: type is not C-compatible due to this default member initializer + 161 | int8_t ifm1_shift=0;// not support any shift + | ^~~~~~~~~~~~~~~~~ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:190:5: note: type is given name 'params_t' for linkage purposes by this typedef declaration + 190 | } params_t; + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:238: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/ceil.h:54: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/ceil_impl.h:118:55: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 118 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:332: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_asym_qdq/conv2d_asym_qdq.h:3: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq.h:7: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq_helpers.hpp:171: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq_kernel_helpers.h:782:1: warning: '/*' within block comment [-Wcomment] + 782 | /* + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:348: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm_bfp16.h:87:21: warning: extra tokens at end of #ifdef directive [-Wextra-tokens] + 87 | #ifdef __AIE_ARCH__ == 20 + | ^ + | // +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable22/src/0_0_reloadable22.cc:12: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:70:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 70 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +24 warnings generated. +Compilation finished successfully (1 errors, 24 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable22_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable22_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable22_orig.ll -o ir/0_0_reloadable22.ll 2> 0_0_reloadable22/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable22.ll -o ir/0_0_reloadable22.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable23.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc -o ir/0_0_reloadable23_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable23.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc -oir/0_0_reloadable23_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:66: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d.h:66: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:111:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 111 | #define LAYER_PARAM_BYTES 0 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw_params.h:70:9: note: previous definition is here + 70 | #define LAYER_PARAM_BYTES 64 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:66: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d.h:66: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:1200:9: warning: 'WT_DM_BANK' macro redefined [-Wmacro-redefined] + 1200 | #define WT_DM_BANK __aie_dm_resource_b + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_params.h:76:9: note: previous definition is here + 76 | #define WT_DM_BANK __aie_dm_resource_a + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:69: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw.h:98:13: warning: 'LOAD_64' macro redefined [-Wmacro-redefined] + 98 | #define LOAD_64(ifm,ibuff,ifm_incr) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:98:13: note: previous definition is here + 98 | #define LOAD_64(ifm,ibuff) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:71: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16.h:227:13: warning: 'LOAD_32' macro redefined [-Wmacro-redefined] + 227 | #define LOAD_32(ifm,ibuff) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:365:13: note: previous definition is here + 365 | #define LOAD_32(ifm,ibuff) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:71: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16.h:231:13: warning: 'VMUL_W4C8' macro redefined [-Wmacro-redefined] + 231 | #define VMUL_W4C8(ifm,wts,acc) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv1d_dw.h:370:13: note: previous definition is here + 370 | #define VMUL_W4C8(ifm,wts,acc) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:73: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_transpose.h:67: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_transpose_params.h:76:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 76 | #define LAYER_PARAM_BYTES 64 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_dw_bf16_params.h:69:9: note: previous definition is here + 69 | #define LAYER_PARAM_BYTES 0 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:746:12: warning: 'clr64' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 746 | return clr64(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331425:7: note: 'clr64' has been explicitly marked deprecated here + 331425 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64acc32 clr64(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:749:12: warning: 'clr32' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 749 | return clr32(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331508:7: note: 'clr32' has been explicitly marked deprecated here + 331508 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v32acc64 clr32(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:752:40: warning: 'clr64f' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 752 | v64accfloat chess_storage(dm4) n = clr64f(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331757:7: note: 'clr64f' has been explicitly marked deprecated here + 331757 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64accfloat clr64f(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:757:12: warning: 'clr64f' is deprecated: Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead. [-Wdeprecated-declarations] + 757 | return clr64f(); + | ^ +/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.h:331757:7: note: 'clr64f' has been explicitly marked deprecated here + 331757 | [[deprecated("Function 'clr' is deprecated. Please use the 'broadcast_zero_to' variant instead.")]] inline __attribute__((always_inline)) v64accfloat clr64f(void) __attribute__((overloadable)) + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:77: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h:61: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_helper.h:167:9: warning: 'CHESS_STORAGE' macro redefined [-Wmacro-redefined] + 167 | #define CHESS_STORAGE( T, accs ) { \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/kernel_helpers.h:1020:9: note: previous definition is here + 1020 | #define CHESS_STORAGE(T, accs) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:481:13: warning: 'LOAD_MAT_B_DATA_2_8x8' macro redefined [-Wmacro-redefined] + 481 | #define LOAD_MAT_B_DATA_2_8x8(p_mat_b, iterator_mat_b, param, Y0, Y1)\ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:414:13: note: previous definition is here + 414 | #define LOAD_MAT_B_DATA_2_8x8(p_mat_b, iterator_mat_b, param, Y0, Y1, Y2, Y3)\ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:600:13: warning: 'STORE_ACC_8_4x8_FULL_DATA' macro redefined [-Wmacro-redefined] + 600 | #define STORE_ACC_8_4x8_FULL_DATA(p_mat_c, iterator_mat_c, param, C0, C1, C2, C3, C4, C5, C6, C7, p_bias_in) \ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:474:13: note: previous definition is here + 474 | #define STORE_ACC_8_4x8_FULL_DATA(p_mat_c, C0, C1, C2, C3) \ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:107: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm.h:637:13: warning: 'MMUL_MAC_8_4x8x8' macro redefined [-Wmacro-redefined] + 637 | #define MMUL_MAC_8_4x8x8(C0, C1, C2, C3, C4, C5, C6, C7, X0, X1, X2, X3, Y0, Y1)\ + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/fullyconnect.h:504:13: note: previous definition is here + 504 | #define MMUL_MAC_8_4x8x8(C0, C1, C2, C3, X0, Y0, Y1, Y2, Y3)\ + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:110: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/mul2d.h:58: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/mul2d_params.h:59:9: warning: 'LAYER_PARAM_BYTES' macro redefined [-Wmacro-redefined] + 59 | #define LAYER_PARAM_BYTES 64 + | ^ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16_params.h:79:9: note: previous definition is here + 79 | #define LAYER_PARAM_BYTES 0 + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:148: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/gelu1d.h:56: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/gelu1d_impl.h:197:59: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 197 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:151: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/tanh.h:54: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/tanh_impl.h:280:59: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 280 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:218: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/floor.h:57: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/floor_impl.h:121:55: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 121 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:236: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops.h:51: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:155:17: warning: anonymous non-C-compatible type given name for linkage purposes by typedef declaration; add a tag name here [-Wnon-c-typedef-for-linkage] + 155 | typedef struct alignas(v32uint16) + | ^ + | params_t +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:161:5: note: type is not C-compatible due to this default member initializer + 161 | int8_t ifm1_shift=0;// not support any shift + | ^~~~~~~~~~~~~~~~~ +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/compare_ops_impl.h:190:5: note: type is given name 'params_t' for linkage purposes by this typedef declaration + 190 | } params_t; + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:238: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/ceil.h:54: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/ceil_impl.h:118:55: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 118 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:332: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_asym_qdq/conv2d_asym_qdq.h:3: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq.h:7: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq_helpers.hpp:171: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/qdq/qdq_kernel_helpers.h:782:1: warning: '/*' within block comment [-Wcomment] + 782 | /* + | ^ +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:54: +In file included from /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/mllib_kernels.h:348: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/gemm/gemm_bfp16.h:87:21: warning: extra tokens at end of #ifdef directive [-Wextra-tokens] + 87 | #ifdef __AIE_ARCH__ == 20 + | ^ + | // +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable23/src/0_0_reloadable23.cc:13: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp:70:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 70 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +23 warnings generated. +Compilation finished successfully (1 errors, 23 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable23_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable23_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable23_orig.ll -o ir/0_0_reloadable23.ll 2> 0_0_reloadable23/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable23.ll -o ir/0_0_reloadable23.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable24.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable24/src/0_0_reloadable24.cc -o ir/0_0_reloadable24_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable24.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable24/src/0_0_reloadable24.cc -oir/0_0_reloadable24_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +In file included from /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable24/src/0_0_reloadable24.cc:9: +/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/resize_adf_wrapper.cpp:70:53: warning: 'truncate' is deprecated: Use saturate instead [-Wdeprecated-declarations] + 70 | tile::current().set_saturation(saturation_mode::truncate); + | ^ +/usr/local/lib/python3.10/dist-packages/include/aie_api/detail/aie2/../../aie_types.hpp:32:17: note: 'truncate' has been explicitly marked deprecated here + 32 | truncate [[deprecated("Use saturate instead")]] = 1, //!< Deprecated: use saturate instead. + | ^ +1 warning generated. +Compilation finished successfully (1 errors, 1 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable24_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable24_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable24_orig.ll -o ir/0_0_reloadable24.ll 2> 0_0_reloadable24/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable24.ll -o ir/0_0_reloadable24.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable25.llgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chesscc +f +s -p me -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 +Wllvm,-O2,-fno-jump-tables,-fno-discard-value-names,-Xclang,-chess-only-info-critical-passes,-g -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I /usr/local/lib/python3.10/dist-packages/include -I /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/dist-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable25/src/0_0_reloadable25.cc -o ir/0_0_reloadable25_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable25.cc" +chess-clang -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -nostdlibinc -D__chess__ -D__tct_release__=2406 -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/dist-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -xc++ -O2 -include aie_core.h -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -O2 -fno-jump-tables -fno-discard-value-names -Xclang -chess-only-info-critical-passes -g /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie/0_0_reloadable25/src/0_0_reloadable25.cc -oir/0_0_reloadable25_main.ll -emit-llvm --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (1 errors, 0 warnings) +/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chess-llvm-link --chess-config /usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg/me_chess_llvm.cfg -S -o ir/0_0_reloadable25_orig.ll ir/i1100_superkernels.ll ir/0_0_reloadable25_main.ll +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S -load-pass-plugin=/usr/local/lib/python3.10/dist-packages/lib/lnx64.o/libLLVMXLOpt.so -passes=xlopt ir/0_0_reloadable25_orig.ll -o ir/0_0_reloadable25.ll 2> 0_0_reloadable25/xlopt.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +/usr/local/lib/python3.10/dist-packages/lnx64.o/tools/clang/bin/opt -S ir/0_0_reloadable25.ll -o ir/0_0_reloadable25.ll 2>/dev/null +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0/scripts/0_0.prx) 2>&1 |& tee -a 0_0/0_0.log 0_0/timestamped_log/0_0.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0.ll -o../Release/chesswork848/0_0.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork848 ../Release/chesswork848/0_0.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0 -L +chess-backend 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0 -L +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/include/adf/aie/tile_control.h", line 292, column 8: in "/usr/local/lib/python3.10/dist-packages/include/adf/aie/tile_control.h", line 292: (loop #8) + loop software pipelining (to 2 cycles) is feasible for a minimum loop count of 5, + but requires the creation of a post-amble, for which the loop was not prepared + ... consider annotating the loop with `chess_prepare_for_pipelining', as well as + increasing the current `chess_loop_range(1,)` annotation to `chess_loop_range(5,)', or remove it. + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0 -L --common 0_0-F_ZN3adf11block_writeEPKNS_7reg_valEj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "0_0/src/0_0.cc", line 37, column 12: in "0_0/src/0_0.cc", line 37: (loop #13) + loop software pipelining (to 8 cycles) is feasible but requires the creation of a post-amble, + for which the loop was not prepared + ... consider annotating the loop with `chess_prepare_for_pipelining' + +Warning: in "0_0/src/0_0.cc", line 12: (loop #3) + Non leaf loop was prepared for pipelining. But the pipelined solutions have not been selected. + Consider removing the chess_prepare_for_pipelining directive as it may improve results +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0 -L --common 0_0-main_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0.objlist -o../0_0.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0" +bridge -o../Release/0_0 ../Release/0_0.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork848 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (19 errors, 3 warnings) +(readelf --debug-dump=decodedline 0_0/Release/0_0 >> 0_0/Release/0_0.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 0_1/Release/0_1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 0_1/Release/0_1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 0_1/Release/0_1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 0_1/Release/0_1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 0_1/Release/0_1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 0_1/Release/0_1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 0_1/Release/0_1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 0_1/Release/0_1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 0_1/Release/0_1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 0_2/Release/0_2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 0_2/Release/0_2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 0_2/Release/0_2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 0_2/Release/0_2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 0_2/Release/0_2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 0_2/Release/0_2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 0_2/Release/0_2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 0_2/Release/0_2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 0_2/Release/0_2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 0_3/Release/0_3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 0_3/Release/0_3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 0_3/Release/0_3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 0_3/Release/0_3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 0_3/Release/0_3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 0_3/Release/0_3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 0_3/Release/0_3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 0_3/Release/0_3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 0_3/Release/0_3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 1_0/Release/1_0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 1_0/Release/1_0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 1_0/Release/1_0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 1_0/Release/1_0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 1_0/Release/1_0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 1_0/Release/1_0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 1_0/Release/1_0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 1_0/Release/1_0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 1_0/Release/1_0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 1_1/Release/1_1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 1_1/Release/1_1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 1_1/Release/1_1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 1_1/Release/1_1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 1_1/Release/1_1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 1_1/Release/1_1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 1_1/Release/1_1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 1_1/Release/1_1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 1_1/Release/1_1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 1_2/Release/1_2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 1_2/Release/1_2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 1_2/Release/1_2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 1_2/Release/1_2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 1_2/Release/1_2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 1_2/Release/1_2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 1_2/Release/1_2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 1_2/Release/1_2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 1_2/Release/1_2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 1_3/Release/1_3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 1_3/Release/1_3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 1_3/Release/1_3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 1_3/Release/1_3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 1_3/Release/1_3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 1_3/Release/1_3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 1_3/Release/1_3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 1_3/Release/1_3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 1_3/Release/1_3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 2_0/Release/2_0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 2_0/Release/2_0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 2_0/Release/2_0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 2_0/Release/2_0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 2_0/Release/2_0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 2_0/Release/2_0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 2_0/Release/2_0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 2_0/Release/2_0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 2_0/Release/2_0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 2_1/Release/2_1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 2_1/Release/2_1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 2_1/Release/2_1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 2_1/Release/2_1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 2_1/Release/2_1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 2_1/Release/2_1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 2_1/Release/2_1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 2_1/Release/2_1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 2_1/Release/2_1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 2_2/Release/2_2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 2_2/Release/2_2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 2_2/Release/2_2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 2_2/Release/2_2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 2_2/Release/2_2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 2_2/Release/2_2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 2_2/Release/2_2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 2_2/Release/2_2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 2_2/Release/2_2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 2_3/Release/2_3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 2_3/Release/2_3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 2_3/Release/2_3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 2_3/Release/2_3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 2_3/Release/2_3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 2_3/Release/2_3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 2_3/Release/2_3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 2_3/Release/2_3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 2_3/Release/2_3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 3_0/Release/3_0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 3_0/Release/3_0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 3_0/Release/3_0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 3_0/Release/3_0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 3_0/Release/3_0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 3_0/Release/3_0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 3_0/Release/3_0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 3_0/Release/3_0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 3_0/Release/3_0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 3_1/Release/3_1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 3_1/Release/3_1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 3_1/Release/3_1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 3_1/Release/3_1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 3_1/Release/3_1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 3_1/Release/3_1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 3_1/Release/3_1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 3_1/Release/3_1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 3_1/Release/3_1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 3_2/Release/3_2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 3_2/Release/3_2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 3_2/Release/3_2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 3_2/Release/3_2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 3_2/Release/3_2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 3_2/Release/3_2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 3_2/Release/3_2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 3_2/Release/3_2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 3_2/Release/3_2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0 3_3/Release/3_3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.map 3_3/Release/3_3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.lst 3_3/Release/3_3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.calltree 3_3/Release/3_3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.sdr 3_3/Release/3_3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.srv 3_3/Release/3_3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.txt 3_3/Release/3_3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmic2 3_3/Release/3_3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0/Release/0_0.cmico 3_3/Release/3_3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable0.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable0/scripts/0_0_reloadable0.prx) 2>&1 |& tee -a 0_0_reloadable0/0_0_reloadable0.log 0_0_reloadable0/timestamped_log/0_0_reloadable0.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable0.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable0.ll -o../Release/chesswork1033/0_0_reloadable0.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork1033 ../Release/chesswork1033/0_0_reloadable0.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +chess-backend 0_0_reloadable0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +chess-backend 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +chess-backend 0_0_reloadable0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z20transpose4d_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z29setup_transposeshuffle_paramsI8bfloat16EvR23transposeshuffle_paramsRA5_Kj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z16transposeshuffleI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R23transposeshuffle_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable0-F_Z13_b719_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z13_b719_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN12mllib_graphs23transpose4d_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z13_b719_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z13_b719_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z13_b719_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z13_b719_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z13_b719_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable0-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable0-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable0-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable0-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable0-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable0-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable0-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable0-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable0-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable0 -L --common 0_0_reloadable0-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable0.objlist -o../0_0_reloadable0.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable0.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable0" +bridge -o../Release/0_0_reloadable0 ../Release/0_0_reloadable0.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable0.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork1033 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable0 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (195 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable0/Release/0_0_reloadable0 >> 0_0_reloadable0/Release/0_0_reloadable0.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 0_1_reloadable0/Release/0_1_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 0_1_reloadable0/Release/0_1_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 0_1_reloadable0/Release/0_1_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 0_1_reloadable0/Release/0_1_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 0_1_reloadable0/Release/0_1_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 0_1_reloadable0/Release/0_1_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 0_1_reloadable0/Release/0_1_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 0_1_reloadable0/Release/0_1_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 0_1_reloadable0/Release/0_1_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 0_2_reloadable0/Release/0_2_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 0_2_reloadable0/Release/0_2_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 0_2_reloadable0/Release/0_2_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 0_2_reloadable0/Release/0_2_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 0_2_reloadable0/Release/0_2_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 0_2_reloadable0/Release/0_2_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 0_2_reloadable0/Release/0_2_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 0_2_reloadable0/Release/0_2_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 0_2_reloadable0/Release/0_2_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 0_3_reloadable0/Release/0_3_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 0_3_reloadable0/Release/0_3_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 0_3_reloadable0/Release/0_3_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 0_3_reloadable0/Release/0_3_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 0_3_reloadable0/Release/0_3_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 0_3_reloadable0/Release/0_3_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 0_3_reloadable0/Release/0_3_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 0_3_reloadable0/Release/0_3_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 0_3_reloadable0/Release/0_3_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 1_0_reloadable0/Release/1_0_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 1_0_reloadable0/Release/1_0_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 1_0_reloadable0/Release/1_0_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 1_0_reloadable0/Release/1_0_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 1_0_reloadable0/Release/1_0_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 1_0_reloadable0/Release/1_0_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 1_0_reloadable0/Release/1_0_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 1_0_reloadable0/Release/1_0_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 1_0_reloadable0/Release/1_0_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 1_1_reloadable0/Release/1_1_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 1_1_reloadable0/Release/1_1_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 1_1_reloadable0/Release/1_1_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 1_1_reloadable0/Release/1_1_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 1_1_reloadable0/Release/1_1_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 1_1_reloadable0/Release/1_1_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 1_1_reloadable0/Release/1_1_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 1_1_reloadable0/Release/1_1_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 1_1_reloadable0/Release/1_1_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 1_2_reloadable0/Release/1_2_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 1_2_reloadable0/Release/1_2_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 1_2_reloadable0/Release/1_2_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 1_2_reloadable0/Release/1_2_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 1_2_reloadable0/Release/1_2_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 1_2_reloadable0/Release/1_2_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 1_2_reloadable0/Release/1_2_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 1_2_reloadable0/Release/1_2_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 1_2_reloadable0/Release/1_2_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 1_3_reloadable0/Release/1_3_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 1_3_reloadable0/Release/1_3_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 1_3_reloadable0/Release/1_3_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 1_3_reloadable0/Release/1_3_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 1_3_reloadable0/Release/1_3_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 1_3_reloadable0/Release/1_3_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 1_3_reloadable0/Release/1_3_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 1_3_reloadable0/Release/1_3_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 1_3_reloadable0/Release/1_3_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 2_0_reloadable0/Release/2_0_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 2_0_reloadable0/Release/2_0_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 2_0_reloadable0/Release/2_0_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 2_0_reloadable0/Release/2_0_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 2_0_reloadable0/Release/2_0_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 2_0_reloadable0/Release/2_0_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 2_0_reloadable0/Release/2_0_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 2_0_reloadable0/Release/2_0_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 2_0_reloadable0/Release/2_0_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 2_1_reloadable0/Release/2_1_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 2_1_reloadable0/Release/2_1_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 2_1_reloadable0/Release/2_1_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 2_1_reloadable0/Release/2_1_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 2_1_reloadable0/Release/2_1_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 2_1_reloadable0/Release/2_1_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 2_1_reloadable0/Release/2_1_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 2_1_reloadable0/Release/2_1_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 2_1_reloadable0/Release/2_1_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 2_2_reloadable0/Release/2_2_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 2_2_reloadable0/Release/2_2_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 2_2_reloadable0/Release/2_2_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 2_2_reloadable0/Release/2_2_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 2_2_reloadable0/Release/2_2_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 2_2_reloadable0/Release/2_2_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 2_2_reloadable0/Release/2_2_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 2_2_reloadable0/Release/2_2_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 2_2_reloadable0/Release/2_2_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 2_3_reloadable0/Release/2_3_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 2_3_reloadable0/Release/2_3_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 2_3_reloadable0/Release/2_3_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 2_3_reloadable0/Release/2_3_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 2_3_reloadable0/Release/2_3_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 2_3_reloadable0/Release/2_3_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 2_3_reloadable0/Release/2_3_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 2_3_reloadable0/Release/2_3_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 2_3_reloadable0/Release/2_3_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 3_0_reloadable0/Release/3_0_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 3_0_reloadable0/Release/3_0_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 3_0_reloadable0/Release/3_0_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 3_0_reloadable0/Release/3_0_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 3_0_reloadable0/Release/3_0_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 3_0_reloadable0/Release/3_0_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 3_0_reloadable0/Release/3_0_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 3_0_reloadable0/Release/3_0_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 3_0_reloadable0/Release/3_0_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 3_1_reloadable0/Release/3_1_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 3_1_reloadable0/Release/3_1_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 3_1_reloadable0/Release/3_1_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 3_1_reloadable0/Release/3_1_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 3_1_reloadable0/Release/3_1_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 3_1_reloadable0/Release/3_1_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 3_1_reloadable0/Release/3_1_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 3_1_reloadable0/Release/3_1_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 3_1_reloadable0/Release/3_1_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 3_2_reloadable0/Release/3_2_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 3_2_reloadable0/Release/3_2_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 3_2_reloadable0/Release/3_2_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 3_2_reloadable0/Release/3_2_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 3_2_reloadable0/Release/3_2_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 3_2_reloadable0/Release/3_2_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 3_2_reloadable0/Release/3_2_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 3_2_reloadable0/Release/3_2_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 3_2_reloadable0/Release/3_2_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable0/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0 3_3_reloadable0/Release/3_3_reloadable0 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.map 3_3_reloadable0/Release/3_3_reloadable0.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.lst 3_3_reloadable0/Release/3_3_reloadable0.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.calltree 3_3_reloadable0/Release/3_3_reloadable0.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.sdr 3_3_reloadable0/Release/3_3_reloadable0.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.srv 3_3_reloadable0/Release/3_3_reloadable0.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.txt 3_3_reloadable0/Release/3_3_reloadable0.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmic2 3_3_reloadable0/Release/3_3_reloadable0.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable0/Release/0_0_reloadable0.cmico 3_3_reloadable0/Release/3_3_reloadable0.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable1.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable1/scripts/0_0_reloadable1.prx) 2>&1 |& tee -a 0_0_reloadable1/0_0_reloadable1.log 0_0_reloadable1/timestamped_log/0_0_reloadable1.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable1.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable1.ll -o../Release/chesswork1394/0_0_reloadable1.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork1394 ../Release/chesswork1394/0_0_reloadable1.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable1-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +chess-backend 0_0_reloadable1-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +chess-backend 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +chess-backend 0_0_reloadable1-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable1-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable1-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable1-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable1-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable1-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable1-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable1-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable1-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable1-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable1-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable1-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable1 -L --common 0_0_reloadable1-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable1.objlist -o../0_0_reloadable1.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable1.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable1" +bridge -o../Release/0_0_reloadable1 ../Release/0_0_reloadable1.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable1.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork1394 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable1 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (171 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable1/Release/0_0_reloadable1 >> 0_0_reloadable1/Release/0_0_reloadable1.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 0_1_reloadable1/Release/0_1_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 0_1_reloadable1/Release/0_1_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 0_1_reloadable1/Release/0_1_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 0_1_reloadable1/Release/0_1_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 0_1_reloadable1/Release/0_1_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 0_1_reloadable1/Release/0_1_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 0_1_reloadable1/Release/0_1_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 0_1_reloadable1/Release/0_1_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 0_1_reloadable1/Release/0_1_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 0_2_reloadable1/Release/0_2_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 0_2_reloadable1/Release/0_2_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 0_2_reloadable1/Release/0_2_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 0_2_reloadable1/Release/0_2_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 0_2_reloadable1/Release/0_2_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 0_2_reloadable1/Release/0_2_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 0_2_reloadable1/Release/0_2_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 0_2_reloadable1/Release/0_2_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 0_2_reloadable1/Release/0_2_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 0_3_reloadable1/Release/0_3_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 0_3_reloadable1/Release/0_3_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 0_3_reloadable1/Release/0_3_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 0_3_reloadable1/Release/0_3_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 0_3_reloadable1/Release/0_3_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 0_3_reloadable1/Release/0_3_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 0_3_reloadable1/Release/0_3_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 0_3_reloadable1/Release/0_3_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 0_3_reloadable1/Release/0_3_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 1_0_reloadable1/Release/1_0_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 1_0_reloadable1/Release/1_0_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 1_0_reloadable1/Release/1_0_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 1_0_reloadable1/Release/1_0_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 1_0_reloadable1/Release/1_0_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 1_0_reloadable1/Release/1_0_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 1_0_reloadable1/Release/1_0_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 1_0_reloadable1/Release/1_0_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 1_0_reloadable1/Release/1_0_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 1_1_reloadable1/Release/1_1_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 1_1_reloadable1/Release/1_1_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 1_1_reloadable1/Release/1_1_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 1_1_reloadable1/Release/1_1_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 1_1_reloadable1/Release/1_1_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 1_1_reloadable1/Release/1_1_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 1_1_reloadable1/Release/1_1_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 1_1_reloadable1/Release/1_1_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 1_1_reloadable1/Release/1_1_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 1_2_reloadable1/Release/1_2_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 1_2_reloadable1/Release/1_2_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 1_2_reloadable1/Release/1_2_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 1_2_reloadable1/Release/1_2_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 1_2_reloadable1/Release/1_2_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 1_2_reloadable1/Release/1_2_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 1_2_reloadable1/Release/1_2_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 1_2_reloadable1/Release/1_2_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 1_2_reloadable1/Release/1_2_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 1_3_reloadable1/Release/1_3_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 1_3_reloadable1/Release/1_3_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 1_3_reloadable1/Release/1_3_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 1_3_reloadable1/Release/1_3_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 1_3_reloadable1/Release/1_3_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 1_3_reloadable1/Release/1_3_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 1_3_reloadable1/Release/1_3_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 1_3_reloadable1/Release/1_3_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 1_3_reloadable1/Release/1_3_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 2_0_reloadable1/Release/2_0_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 2_0_reloadable1/Release/2_0_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 2_0_reloadable1/Release/2_0_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 2_0_reloadable1/Release/2_0_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 2_0_reloadable1/Release/2_0_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 2_0_reloadable1/Release/2_0_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 2_0_reloadable1/Release/2_0_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 2_0_reloadable1/Release/2_0_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 2_0_reloadable1/Release/2_0_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 2_1_reloadable1/Release/2_1_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 2_1_reloadable1/Release/2_1_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 2_1_reloadable1/Release/2_1_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 2_1_reloadable1/Release/2_1_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 2_1_reloadable1/Release/2_1_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 2_1_reloadable1/Release/2_1_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 2_1_reloadable1/Release/2_1_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 2_1_reloadable1/Release/2_1_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 2_1_reloadable1/Release/2_1_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 2_2_reloadable1/Release/2_2_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 2_2_reloadable1/Release/2_2_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 2_2_reloadable1/Release/2_2_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 2_2_reloadable1/Release/2_2_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 2_2_reloadable1/Release/2_2_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 2_2_reloadable1/Release/2_2_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 2_2_reloadable1/Release/2_2_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 2_2_reloadable1/Release/2_2_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 2_2_reloadable1/Release/2_2_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 2_3_reloadable1/Release/2_3_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 2_3_reloadable1/Release/2_3_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 2_3_reloadable1/Release/2_3_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 2_3_reloadable1/Release/2_3_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 2_3_reloadable1/Release/2_3_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 2_3_reloadable1/Release/2_3_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 2_3_reloadable1/Release/2_3_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 2_3_reloadable1/Release/2_3_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 2_3_reloadable1/Release/2_3_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 3_0_reloadable1/Release/3_0_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 3_0_reloadable1/Release/3_0_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 3_0_reloadable1/Release/3_0_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 3_0_reloadable1/Release/3_0_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 3_0_reloadable1/Release/3_0_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 3_0_reloadable1/Release/3_0_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 3_0_reloadable1/Release/3_0_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 3_0_reloadable1/Release/3_0_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 3_0_reloadable1/Release/3_0_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 3_1_reloadable1/Release/3_1_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 3_1_reloadable1/Release/3_1_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 3_1_reloadable1/Release/3_1_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 3_1_reloadable1/Release/3_1_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 3_1_reloadable1/Release/3_1_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 3_1_reloadable1/Release/3_1_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 3_1_reloadable1/Release/3_1_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 3_1_reloadable1/Release/3_1_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 3_1_reloadable1/Release/3_1_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 3_2_reloadable1/Release/3_2_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 3_2_reloadable1/Release/3_2_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 3_2_reloadable1/Release/3_2_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 3_2_reloadable1/Release/3_2_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 3_2_reloadable1/Release/3_2_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 3_2_reloadable1/Release/3_2_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 3_2_reloadable1/Release/3_2_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 3_2_reloadable1/Release/3_2_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 3_2_reloadable1/Release/3_2_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable1/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1 3_3_reloadable1/Release/3_3_reloadable1 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.map 3_3_reloadable1/Release/3_3_reloadable1.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.lst 3_3_reloadable1/Release/3_3_reloadable1.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.calltree 3_3_reloadable1/Release/3_3_reloadable1.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.sdr 3_3_reloadable1/Release/3_3_reloadable1.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.srv 3_3_reloadable1/Release/3_3_reloadable1.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.txt 3_3_reloadable1/Release/3_3_reloadable1.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmic2 3_3_reloadable1/Release/3_3_reloadable1.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable1/Release/0_0_reloadable1.cmico 3_3_reloadable1/Release/3_3_reloadable1.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable2.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable2/scripts/0_0_reloadable2.prx) 2>&1 |& tee -a 0_0_reloadable2/0_0_reloadable2.log 0_0_reloadable2/timestamped_log/0_0_reloadable2.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable2.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable2.ll -o../Release/chesswork1731/0_0_reloadable2.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork1731 ../Release/chesswork1731/0_0_reloadable2.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable2-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +chess-backend 0_0_reloadable2-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +chess-backend 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +chess-backend 0_0_reloadable2-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable2-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable2-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable2-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable2-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable2-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable2-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable2-F_Z13_b961_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +chess-backend 0_0_reloadable2-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable2-F_Z13_b961_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable2-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_Z13_b961_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_Z13_b961_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_Z13_b961_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_Z13_b961_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable2-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_Z13_b961_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable2 -L --common 0_0_reloadable2-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable2.objlist -o../0_0_reloadable2.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable2.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable2" +bridge -o../Release/0_0_reloadable2 ../Release/0_0_reloadable2.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable2.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork1731 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable2 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (165 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable2/Release/0_0_reloadable2 >> 0_0_reloadable2/Release/0_0_reloadable2.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_0_reloadable4/Release/0_0_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_0_reloadable4/Release/0_0_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_0_reloadable4/Release/0_0_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_0_reloadable4/Release/0_0_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_0_reloadable4/Release/0_0_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_0_reloadable4/Release/0_0_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_0_reloadable4/Release/0_0_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_0_reloadable4/Release/0_0_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_0_reloadable4/Release/0_0_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_0_reloadable6/Release/0_0_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_0_reloadable6/Release/0_0_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_0_reloadable6/Release/0_0_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_0_reloadable6/Release/0_0_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_0_reloadable6/Release/0_0_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_0_reloadable6/Release/0_0_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_0_reloadable6/Release/0_0_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_0_reloadable6/Release/0_0_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_0_reloadable6/Release/0_0_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_0_reloadable8/Release/0_0_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_0_reloadable8/Release/0_0_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_0_reloadable8/Release/0_0_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_0_reloadable8/Release/0_0_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_0_reloadable8/Release/0_0_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_0_reloadable8/Release/0_0_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_0_reloadable8/Release/0_0_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_0_reloadable8/Release/0_0_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_0_reloadable8/Release/0_0_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_0_reloadable10/Release/0_0_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_0_reloadable10/Release/0_0_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_0_reloadable10/Release/0_0_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_0_reloadable10/Release/0_0_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_0_reloadable10/Release/0_0_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_0_reloadable10/Release/0_0_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_0_reloadable10/Release/0_0_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_0_reloadable10/Release/0_0_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_0_reloadable10/Release/0_0_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_0_reloadable12/Release/0_0_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_0_reloadable12/Release/0_0_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_0_reloadable12/Release/0_0_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_0_reloadable12/Release/0_0_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_0_reloadable12/Release/0_0_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_0_reloadable12/Release/0_0_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_0_reloadable12/Release/0_0_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_0_reloadable12/Release/0_0_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_0_reloadable12/Release/0_0_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_0_reloadable14/Release/0_0_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_0_reloadable14/Release/0_0_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_0_reloadable14/Release/0_0_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_0_reloadable14/Release/0_0_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_0_reloadable14/Release/0_0_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_0_reloadable14/Release/0_0_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_0_reloadable14/Release/0_0_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_0_reloadable14/Release/0_0_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_0_reloadable14/Release/0_0_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_0_reloadable16/Release/0_0_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_0_reloadable16/Release/0_0_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_0_reloadable16/Release/0_0_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_0_reloadable16/Release/0_0_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_0_reloadable16/Release/0_0_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_0_reloadable16/Release/0_0_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_0_reloadable16/Release/0_0_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_0_reloadable16/Release/0_0_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_0_reloadable16/Release/0_0_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_0_reloadable18/Release/0_0_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_0_reloadable18/Release/0_0_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_0_reloadable18/Release/0_0_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_0_reloadable18/Release/0_0_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_0_reloadable18/Release/0_0_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_0_reloadable18/Release/0_0_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_0_reloadable18/Release/0_0_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_0_reloadable18/Release/0_0_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_0_reloadable18/Release/0_0_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_1_reloadable2/Release/0_1_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_1_reloadable2/Release/0_1_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_1_reloadable2/Release/0_1_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_1_reloadable2/Release/0_1_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_1_reloadable2/Release/0_1_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_1_reloadable2/Release/0_1_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_1_reloadable2/Release/0_1_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_1_reloadable2/Release/0_1_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_1_reloadable2/Release/0_1_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_1_reloadable4/Release/0_1_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_1_reloadable4/Release/0_1_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_1_reloadable4/Release/0_1_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_1_reloadable4/Release/0_1_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_1_reloadable4/Release/0_1_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_1_reloadable4/Release/0_1_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_1_reloadable4/Release/0_1_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_1_reloadable4/Release/0_1_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_1_reloadable4/Release/0_1_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_1_reloadable6/Release/0_1_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_1_reloadable6/Release/0_1_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_1_reloadable6/Release/0_1_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_1_reloadable6/Release/0_1_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_1_reloadable6/Release/0_1_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_1_reloadable6/Release/0_1_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_1_reloadable6/Release/0_1_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_1_reloadable6/Release/0_1_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_1_reloadable6/Release/0_1_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_1_reloadable8/Release/0_1_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_1_reloadable8/Release/0_1_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_1_reloadable8/Release/0_1_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_1_reloadable8/Release/0_1_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_1_reloadable8/Release/0_1_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_1_reloadable8/Release/0_1_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_1_reloadable8/Release/0_1_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_1_reloadable8/Release/0_1_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_1_reloadable8/Release/0_1_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_1_reloadable10/Release/0_1_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_1_reloadable10/Release/0_1_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_1_reloadable10/Release/0_1_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_1_reloadable10/Release/0_1_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_1_reloadable10/Release/0_1_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_1_reloadable10/Release/0_1_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_1_reloadable10/Release/0_1_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_1_reloadable10/Release/0_1_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_1_reloadable10/Release/0_1_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_1_reloadable12/Release/0_1_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_1_reloadable12/Release/0_1_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_1_reloadable12/Release/0_1_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_1_reloadable12/Release/0_1_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_1_reloadable12/Release/0_1_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_1_reloadable12/Release/0_1_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_1_reloadable12/Release/0_1_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_1_reloadable12/Release/0_1_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_1_reloadable12/Release/0_1_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_1_reloadable14/Release/0_1_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_1_reloadable14/Release/0_1_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_1_reloadable14/Release/0_1_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_1_reloadable14/Release/0_1_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_1_reloadable14/Release/0_1_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_1_reloadable14/Release/0_1_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_1_reloadable14/Release/0_1_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_1_reloadable14/Release/0_1_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_1_reloadable14/Release/0_1_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_1_reloadable16/Release/0_1_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_1_reloadable16/Release/0_1_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_1_reloadable16/Release/0_1_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_1_reloadable16/Release/0_1_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_1_reloadable16/Release/0_1_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_1_reloadable16/Release/0_1_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_1_reloadable16/Release/0_1_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_1_reloadable16/Release/0_1_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_1_reloadable16/Release/0_1_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_1_reloadable18/Release/0_1_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_1_reloadable18/Release/0_1_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_1_reloadable18/Release/0_1_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_1_reloadable18/Release/0_1_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_1_reloadable18/Release/0_1_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_1_reloadable18/Release/0_1_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_1_reloadable18/Release/0_1_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_1_reloadable18/Release/0_1_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_1_reloadable18/Release/0_1_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_2_reloadable2/Release/0_2_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_2_reloadable2/Release/0_2_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_2_reloadable2/Release/0_2_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_2_reloadable2/Release/0_2_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_2_reloadable2/Release/0_2_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_2_reloadable2/Release/0_2_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_2_reloadable2/Release/0_2_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_2_reloadable2/Release/0_2_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_2_reloadable2/Release/0_2_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_2_reloadable4/Release/0_2_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_2_reloadable4/Release/0_2_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_2_reloadable4/Release/0_2_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_2_reloadable4/Release/0_2_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_2_reloadable4/Release/0_2_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_2_reloadable4/Release/0_2_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_2_reloadable4/Release/0_2_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_2_reloadable4/Release/0_2_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_2_reloadable4/Release/0_2_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_2_reloadable6/Release/0_2_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_2_reloadable6/Release/0_2_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_2_reloadable6/Release/0_2_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_2_reloadable6/Release/0_2_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_2_reloadable6/Release/0_2_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_2_reloadable6/Release/0_2_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_2_reloadable6/Release/0_2_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_2_reloadable6/Release/0_2_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_2_reloadable6/Release/0_2_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_2_reloadable8/Release/0_2_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_2_reloadable8/Release/0_2_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_2_reloadable8/Release/0_2_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_2_reloadable8/Release/0_2_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_2_reloadable8/Release/0_2_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_2_reloadable8/Release/0_2_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_2_reloadable8/Release/0_2_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_2_reloadable8/Release/0_2_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_2_reloadable8/Release/0_2_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_2_reloadable10/Release/0_2_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_2_reloadable10/Release/0_2_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_2_reloadable10/Release/0_2_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_2_reloadable10/Release/0_2_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_2_reloadable10/Release/0_2_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_2_reloadable10/Release/0_2_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_2_reloadable10/Release/0_2_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_2_reloadable10/Release/0_2_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_2_reloadable10/Release/0_2_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_2_reloadable12/Release/0_2_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_2_reloadable12/Release/0_2_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_2_reloadable12/Release/0_2_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_2_reloadable12/Release/0_2_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_2_reloadable12/Release/0_2_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_2_reloadable12/Release/0_2_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_2_reloadable12/Release/0_2_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_2_reloadable12/Release/0_2_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_2_reloadable12/Release/0_2_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_2_reloadable14/Release/0_2_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_2_reloadable14/Release/0_2_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_2_reloadable14/Release/0_2_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_2_reloadable14/Release/0_2_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_2_reloadable14/Release/0_2_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_2_reloadable14/Release/0_2_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_2_reloadable14/Release/0_2_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_2_reloadable14/Release/0_2_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_2_reloadable14/Release/0_2_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_2_reloadable16/Release/0_2_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_2_reloadable16/Release/0_2_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_2_reloadable16/Release/0_2_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_2_reloadable16/Release/0_2_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_2_reloadable16/Release/0_2_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_2_reloadable16/Release/0_2_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_2_reloadable16/Release/0_2_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_2_reloadable16/Release/0_2_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_2_reloadable16/Release/0_2_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_2_reloadable18/Release/0_2_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_2_reloadable18/Release/0_2_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_2_reloadable18/Release/0_2_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_2_reloadable18/Release/0_2_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_2_reloadable18/Release/0_2_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_2_reloadable18/Release/0_2_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_2_reloadable18/Release/0_2_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_2_reloadable18/Release/0_2_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_2_reloadable18/Release/0_2_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_3_reloadable2/Release/0_3_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_3_reloadable2/Release/0_3_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_3_reloadable2/Release/0_3_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_3_reloadable2/Release/0_3_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_3_reloadable2/Release/0_3_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_3_reloadable2/Release/0_3_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_3_reloadable2/Release/0_3_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_3_reloadable2/Release/0_3_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_3_reloadable2/Release/0_3_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_3_reloadable4/Release/0_3_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_3_reloadable4/Release/0_3_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_3_reloadable4/Release/0_3_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_3_reloadable4/Release/0_3_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_3_reloadable4/Release/0_3_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_3_reloadable4/Release/0_3_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_3_reloadable4/Release/0_3_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_3_reloadable4/Release/0_3_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_3_reloadable4/Release/0_3_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_3_reloadable6/Release/0_3_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_3_reloadable6/Release/0_3_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_3_reloadable6/Release/0_3_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_3_reloadable6/Release/0_3_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_3_reloadable6/Release/0_3_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_3_reloadable6/Release/0_3_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_3_reloadable6/Release/0_3_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_3_reloadable6/Release/0_3_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_3_reloadable6/Release/0_3_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_3_reloadable8/Release/0_3_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_3_reloadable8/Release/0_3_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_3_reloadable8/Release/0_3_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_3_reloadable8/Release/0_3_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_3_reloadable8/Release/0_3_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_3_reloadable8/Release/0_3_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_3_reloadable8/Release/0_3_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_3_reloadable8/Release/0_3_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_3_reloadable8/Release/0_3_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_3_reloadable10/Release/0_3_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_3_reloadable10/Release/0_3_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_3_reloadable10/Release/0_3_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_3_reloadable10/Release/0_3_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_3_reloadable10/Release/0_3_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_3_reloadable10/Release/0_3_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_3_reloadable10/Release/0_3_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_3_reloadable10/Release/0_3_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_3_reloadable10/Release/0_3_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_3_reloadable12/Release/0_3_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_3_reloadable12/Release/0_3_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_3_reloadable12/Release/0_3_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_3_reloadable12/Release/0_3_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_3_reloadable12/Release/0_3_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_3_reloadable12/Release/0_3_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_3_reloadable12/Release/0_3_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_3_reloadable12/Release/0_3_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_3_reloadable12/Release/0_3_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_3_reloadable14/Release/0_3_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_3_reloadable14/Release/0_3_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_3_reloadable14/Release/0_3_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_3_reloadable14/Release/0_3_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_3_reloadable14/Release/0_3_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_3_reloadable14/Release/0_3_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_3_reloadable14/Release/0_3_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_3_reloadable14/Release/0_3_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_3_reloadable14/Release/0_3_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_3_reloadable16/Release/0_3_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_3_reloadable16/Release/0_3_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_3_reloadable16/Release/0_3_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_3_reloadable16/Release/0_3_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_3_reloadable16/Release/0_3_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_3_reloadable16/Release/0_3_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_3_reloadable16/Release/0_3_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_3_reloadable16/Release/0_3_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_3_reloadable16/Release/0_3_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 0_3_reloadable18/Release/0_3_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 0_3_reloadable18/Release/0_3_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 0_3_reloadable18/Release/0_3_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 0_3_reloadable18/Release/0_3_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 0_3_reloadable18/Release/0_3_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 0_3_reloadable18/Release/0_3_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 0_3_reloadable18/Release/0_3_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 0_3_reloadable18/Release/0_3_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 0_3_reloadable18/Release/0_3_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_0_reloadable2/Release/1_0_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_0_reloadable2/Release/1_0_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_0_reloadable2/Release/1_0_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_0_reloadable2/Release/1_0_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_0_reloadable2/Release/1_0_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_0_reloadable2/Release/1_0_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_0_reloadable2/Release/1_0_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_0_reloadable2/Release/1_0_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_0_reloadable2/Release/1_0_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_0_reloadable4/Release/1_0_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_0_reloadable4/Release/1_0_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_0_reloadable4/Release/1_0_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_0_reloadable4/Release/1_0_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_0_reloadable4/Release/1_0_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_0_reloadable4/Release/1_0_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_0_reloadable4/Release/1_0_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_0_reloadable4/Release/1_0_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_0_reloadable4/Release/1_0_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_0_reloadable6/Release/1_0_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_0_reloadable6/Release/1_0_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_0_reloadable6/Release/1_0_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_0_reloadable6/Release/1_0_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_0_reloadable6/Release/1_0_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_0_reloadable6/Release/1_0_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_0_reloadable6/Release/1_0_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_0_reloadable6/Release/1_0_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_0_reloadable6/Release/1_0_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_0_reloadable8/Release/1_0_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_0_reloadable8/Release/1_0_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_0_reloadable8/Release/1_0_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_0_reloadable8/Release/1_0_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_0_reloadable8/Release/1_0_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_0_reloadable8/Release/1_0_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_0_reloadable8/Release/1_0_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_0_reloadable8/Release/1_0_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_0_reloadable8/Release/1_0_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_0_reloadable10/Release/1_0_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_0_reloadable10/Release/1_0_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_0_reloadable10/Release/1_0_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_0_reloadable10/Release/1_0_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_0_reloadable10/Release/1_0_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_0_reloadable10/Release/1_0_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_0_reloadable10/Release/1_0_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_0_reloadable10/Release/1_0_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_0_reloadable10/Release/1_0_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_0_reloadable12/Release/1_0_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_0_reloadable12/Release/1_0_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_0_reloadable12/Release/1_0_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_0_reloadable12/Release/1_0_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_0_reloadable12/Release/1_0_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_0_reloadable12/Release/1_0_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_0_reloadable12/Release/1_0_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_0_reloadable12/Release/1_0_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_0_reloadable12/Release/1_0_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_0_reloadable14/Release/1_0_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_0_reloadable14/Release/1_0_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_0_reloadable14/Release/1_0_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_0_reloadable14/Release/1_0_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_0_reloadable14/Release/1_0_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_0_reloadable14/Release/1_0_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_0_reloadable14/Release/1_0_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_0_reloadable14/Release/1_0_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_0_reloadable14/Release/1_0_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_0_reloadable16/Release/1_0_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_0_reloadable16/Release/1_0_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_0_reloadable16/Release/1_0_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_0_reloadable16/Release/1_0_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_0_reloadable16/Release/1_0_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_0_reloadable16/Release/1_0_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_0_reloadable16/Release/1_0_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_0_reloadable16/Release/1_0_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_0_reloadable16/Release/1_0_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_0_reloadable18/Release/1_0_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_0_reloadable18/Release/1_0_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_0_reloadable18/Release/1_0_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_0_reloadable18/Release/1_0_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_0_reloadable18/Release/1_0_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_0_reloadable18/Release/1_0_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_0_reloadable18/Release/1_0_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_0_reloadable18/Release/1_0_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_0_reloadable18/Release/1_0_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_1_reloadable2/Release/1_1_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_1_reloadable2/Release/1_1_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_1_reloadable2/Release/1_1_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_1_reloadable2/Release/1_1_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_1_reloadable2/Release/1_1_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_1_reloadable2/Release/1_1_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_1_reloadable2/Release/1_1_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_1_reloadable2/Release/1_1_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_1_reloadable2/Release/1_1_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_1_reloadable4/Release/1_1_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_1_reloadable4/Release/1_1_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_1_reloadable4/Release/1_1_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_1_reloadable4/Release/1_1_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_1_reloadable4/Release/1_1_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_1_reloadable4/Release/1_1_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_1_reloadable4/Release/1_1_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_1_reloadable4/Release/1_1_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_1_reloadable4/Release/1_1_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_1_reloadable6/Release/1_1_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_1_reloadable6/Release/1_1_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_1_reloadable6/Release/1_1_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_1_reloadable6/Release/1_1_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_1_reloadable6/Release/1_1_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_1_reloadable6/Release/1_1_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_1_reloadable6/Release/1_1_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_1_reloadable6/Release/1_1_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_1_reloadable6/Release/1_1_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_1_reloadable8/Release/1_1_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_1_reloadable8/Release/1_1_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_1_reloadable8/Release/1_1_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_1_reloadable8/Release/1_1_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_1_reloadable8/Release/1_1_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_1_reloadable8/Release/1_1_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_1_reloadable8/Release/1_1_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_1_reloadable8/Release/1_1_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_1_reloadable8/Release/1_1_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_1_reloadable10/Release/1_1_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_1_reloadable10/Release/1_1_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_1_reloadable10/Release/1_1_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_1_reloadable10/Release/1_1_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_1_reloadable10/Release/1_1_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_1_reloadable10/Release/1_1_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_1_reloadable10/Release/1_1_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_1_reloadable10/Release/1_1_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_1_reloadable10/Release/1_1_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_1_reloadable12/Release/1_1_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_1_reloadable12/Release/1_1_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_1_reloadable12/Release/1_1_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_1_reloadable12/Release/1_1_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_1_reloadable12/Release/1_1_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_1_reloadable12/Release/1_1_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_1_reloadable12/Release/1_1_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_1_reloadable12/Release/1_1_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_1_reloadable12/Release/1_1_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_1_reloadable14/Release/1_1_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_1_reloadable14/Release/1_1_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_1_reloadable14/Release/1_1_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_1_reloadable14/Release/1_1_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_1_reloadable14/Release/1_1_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_1_reloadable14/Release/1_1_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_1_reloadable14/Release/1_1_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_1_reloadable14/Release/1_1_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_1_reloadable14/Release/1_1_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_1_reloadable16/Release/1_1_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_1_reloadable16/Release/1_1_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_1_reloadable16/Release/1_1_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_1_reloadable16/Release/1_1_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_1_reloadable16/Release/1_1_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_1_reloadable16/Release/1_1_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_1_reloadable16/Release/1_1_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_1_reloadable16/Release/1_1_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_1_reloadable16/Release/1_1_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_1_reloadable18/Release/1_1_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_1_reloadable18/Release/1_1_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_1_reloadable18/Release/1_1_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_1_reloadable18/Release/1_1_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_1_reloadable18/Release/1_1_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_1_reloadable18/Release/1_1_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_1_reloadable18/Release/1_1_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_1_reloadable18/Release/1_1_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_1_reloadable18/Release/1_1_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_2_reloadable2/Release/1_2_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_2_reloadable2/Release/1_2_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_2_reloadable2/Release/1_2_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_2_reloadable2/Release/1_2_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_2_reloadable2/Release/1_2_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_2_reloadable2/Release/1_2_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_2_reloadable2/Release/1_2_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_2_reloadable2/Release/1_2_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_2_reloadable2/Release/1_2_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_2_reloadable4/Release/1_2_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_2_reloadable4/Release/1_2_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_2_reloadable4/Release/1_2_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_2_reloadable4/Release/1_2_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_2_reloadable4/Release/1_2_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_2_reloadable4/Release/1_2_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_2_reloadable4/Release/1_2_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_2_reloadable4/Release/1_2_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_2_reloadable4/Release/1_2_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_2_reloadable6/Release/1_2_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_2_reloadable6/Release/1_2_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_2_reloadable6/Release/1_2_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_2_reloadable6/Release/1_2_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_2_reloadable6/Release/1_2_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_2_reloadable6/Release/1_2_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_2_reloadable6/Release/1_2_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_2_reloadable6/Release/1_2_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_2_reloadable6/Release/1_2_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_2_reloadable8/Release/1_2_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_2_reloadable8/Release/1_2_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_2_reloadable8/Release/1_2_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_2_reloadable8/Release/1_2_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_2_reloadable8/Release/1_2_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_2_reloadable8/Release/1_2_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_2_reloadable8/Release/1_2_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_2_reloadable8/Release/1_2_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_2_reloadable8/Release/1_2_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_2_reloadable10/Release/1_2_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_2_reloadable10/Release/1_2_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_2_reloadable10/Release/1_2_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_2_reloadable10/Release/1_2_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_2_reloadable10/Release/1_2_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_2_reloadable10/Release/1_2_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_2_reloadable10/Release/1_2_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_2_reloadable10/Release/1_2_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_2_reloadable10/Release/1_2_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_2_reloadable12/Release/1_2_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_2_reloadable12/Release/1_2_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_2_reloadable12/Release/1_2_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_2_reloadable12/Release/1_2_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_2_reloadable12/Release/1_2_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_2_reloadable12/Release/1_2_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_2_reloadable12/Release/1_2_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_2_reloadable12/Release/1_2_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_2_reloadable12/Release/1_2_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_2_reloadable14/Release/1_2_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_2_reloadable14/Release/1_2_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_2_reloadable14/Release/1_2_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_2_reloadable14/Release/1_2_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_2_reloadable14/Release/1_2_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_2_reloadable14/Release/1_2_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_2_reloadable14/Release/1_2_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_2_reloadable14/Release/1_2_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_2_reloadable14/Release/1_2_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_2_reloadable16/Release/1_2_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_2_reloadable16/Release/1_2_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_2_reloadable16/Release/1_2_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_2_reloadable16/Release/1_2_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_2_reloadable16/Release/1_2_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_2_reloadable16/Release/1_2_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_2_reloadable16/Release/1_2_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_2_reloadable16/Release/1_2_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_2_reloadable16/Release/1_2_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_2_reloadable18/Release/1_2_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_2_reloadable18/Release/1_2_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_2_reloadable18/Release/1_2_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_2_reloadable18/Release/1_2_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_2_reloadable18/Release/1_2_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_2_reloadable18/Release/1_2_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_2_reloadable18/Release/1_2_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_2_reloadable18/Release/1_2_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_2_reloadable18/Release/1_2_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_3_reloadable2/Release/1_3_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_3_reloadable2/Release/1_3_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_3_reloadable2/Release/1_3_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_3_reloadable2/Release/1_3_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_3_reloadable2/Release/1_3_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_3_reloadable2/Release/1_3_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_3_reloadable2/Release/1_3_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_3_reloadable2/Release/1_3_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_3_reloadable2/Release/1_3_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_3_reloadable4/Release/1_3_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_3_reloadable4/Release/1_3_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_3_reloadable4/Release/1_3_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_3_reloadable4/Release/1_3_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_3_reloadable4/Release/1_3_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_3_reloadable4/Release/1_3_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_3_reloadable4/Release/1_3_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_3_reloadable4/Release/1_3_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_3_reloadable4/Release/1_3_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_3_reloadable6/Release/1_3_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_3_reloadable6/Release/1_3_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_3_reloadable6/Release/1_3_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_3_reloadable6/Release/1_3_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_3_reloadable6/Release/1_3_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_3_reloadable6/Release/1_3_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_3_reloadable6/Release/1_3_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_3_reloadable6/Release/1_3_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_3_reloadable6/Release/1_3_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_3_reloadable8/Release/1_3_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_3_reloadable8/Release/1_3_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_3_reloadable8/Release/1_3_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_3_reloadable8/Release/1_3_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_3_reloadable8/Release/1_3_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_3_reloadable8/Release/1_3_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_3_reloadable8/Release/1_3_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_3_reloadable8/Release/1_3_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_3_reloadable8/Release/1_3_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_3_reloadable10/Release/1_3_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_3_reloadable10/Release/1_3_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_3_reloadable10/Release/1_3_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_3_reloadable10/Release/1_3_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_3_reloadable10/Release/1_3_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_3_reloadable10/Release/1_3_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_3_reloadable10/Release/1_3_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_3_reloadable10/Release/1_3_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_3_reloadable10/Release/1_3_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_3_reloadable12/Release/1_3_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_3_reloadable12/Release/1_3_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_3_reloadable12/Release/1_3_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_3_reloadable12/Release/1_3_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_3_reloadable12/Release/1_3_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_3_reloadable12/Release/1_3_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_3_reloadable12/Release/1_3_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_3_reloadable12/Release/1_3_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_3_reloadable12/Release/1_3_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_3_reloadable14/Release/1_3_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_3_reloadable14/Release/1_3_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_3_reloadable14/Release/1_3_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_3_reloadable14/Release/1_3_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_3_reloadable14/Release/1_3_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_3_reloadable14/Release/1_3_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_3_reloadable14/Release/1_3_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_3_reloadable14/Release/1_3_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_3_reloadable14/Release/1_3_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_3_reloadable16/Release/1_3_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_3_reloadable16/Release/1_3_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_3_reloadable16/Release/1_3_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_3_reloadable16/Release/1_3_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_3_reloadable16/Release/1_3_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_3_reloadable16/Release/1_3_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_3_reloadable16/Release/1_3_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_3_reloadable16/Release/1_3_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_3_reloadable16/Release/1_3_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 1_3_reloadable18/Release/1_3_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 1_3_reloadable18/Release/1_3_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 1_3_reloadable18/Release/1_3_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 1_3_reloadable18/Release/1_3_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 1_3_reloadable18/Release/1_3_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 1_3_reloadable18/Release/1_3_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 1_3_reloadable18/Release/1_3_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 1_3_reloadable18/Release/1_3_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 1_3_reloadable18/Release/1_3_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_0_reloadable2/Release/2_0_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_0_reloadable2/Release/2_0_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_0_reloadable2/Release/2_0_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_0_reloadable2/Release/2_0_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_0_reloadable2/Release/2_0_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_0_reloadable2/Release/2_0_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_0_reloadable2/Release/2_0_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_0_reloadable2/Release/2_0_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_0_reloadable2/Release/2_0_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_0_reloadable4/Release/2_0_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_0_reloadable4/Release/2_0_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_0_reloadable4/Release/2_0_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_0_reloadable4/Release/2_0_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_0_reloadable4/Release/2_0_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_0_reloadable4/Release/2_0_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_0_reloadable4/Release/2_0_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_0_reloadable4/Release/2_0_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_0_reloadable4/Release/2_0_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_0_reloadable6/Release/2_0_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_0_reloadable6/Release/2_0_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_0_reloadable6/Release/2_0_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_0_reloadable6/Release/2_0_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_0_reloadable6/Release/2_0_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_0_reloadable6/Release/2_0_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_0_reloadable6/Release/2_0_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_0_reloadable6/Release/2_0_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_0_reloadable6/Release/2_0_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_0_reloadable8/Release/2_0_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_0_reloadable8/Release/2_0_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_0_reloadable8/Release/2_0_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_0_reloadable8/Release/2_0_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_0_reloadable8/Release/2_0_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_0_reloadable8/Release/2_0_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_0_reloadable8/Release/2_0_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_0_reloadable8/Release/2_0_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_0_reloadable8/Release/2_0_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_0_reloadable10/Release/2_0_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_0_reloadable10/Release/2_0_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_0_reloadable10/Release/2_0_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_0_reloadable10/Release/2_0_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_0_reloadable10/Release/2_0_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_0_reloadable10/Release/2_0_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_0_reloadable10/Release/2_0_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_0_reloadable10/Release/2_0_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_0_reloadable10/Release/2_0_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_0_reloadable12/Release/2_0_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_0_reloadable12/Release/2_0_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_0_reloadable12/Release/2_0_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_0_reloadable12/Release/2_0_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_0_reloadable12/Release/2_0_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_0_reloadable12/Release/2_0_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_0_reloadable12/Release/2_0_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_0_reloadable12/Release/2_0_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_0_reloadable12/Release/2_0_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_0_reloadable14/Release/2_0_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_0_reloadable14/Release/2_0_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_0_reloadable14/Release/2_0_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_0_reloadable14/Release/2_0_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_0_reloadable14/Release/2_0_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_0_reloadable14/Release/2_0_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_0_reloadable14/Release/2_0_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_0_reloadable14/Release/2_0_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_0_reloadable14/Release/2_0_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_0_reloadable16/Release/2_0_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_0_reloadable16/Release/2_0_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_0_reloadable16/Release/2_0_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_0_reloadable16/Release/2_0_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_0_reloadable16/Release/2_0_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_0_reloadable16/Release/2_0_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_0_reloadable16/Release/2_0_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_0_reloadable16/Release/2_0_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_0_reloadable16/Release/2_0_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_0_reloadable18/Release/2_0_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_0_reloadable18/Release/2_0_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_0_reloadable18/Release/2_0_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_0_reloadable18/Release/2_0_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_0_reloadable18/Release/2_0_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_0_reloadable18/Release/2_0_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_0_reloadable18/Release/2_0_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_0_reloadable18/Release/2_0_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_0_reloadable18/Release/2_0_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_1_reloadable2/Release/2_1_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_1_reloadable2/Release/2_1_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_1_reloadable2/Release/2_1_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_1_reloadable2/Release/2_1_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_1_reloadable2/Release/2_1_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_1_reloadable2/Release/2_1_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_1_reloadable2/Release/2_1_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_1_reloadable2/Release/2_1_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_1_reloadable2/Release/2_1_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_1_reloadable4/Release/2_1_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_1_reloadable4/Release/2_1_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_1_reloadable4/Release/2_1_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_1_reloadable4/Release/2_1_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_1_reloadable4/Release/2_1_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_1_reloadable4/Release/2_1_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_1_reloadable4/Release/2_1_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_1_reloadable4/Release/2_1_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_1_reloadable4/Release/2_1_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_1_reloadable6/Release/2_1_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_1_reloadable6/Release/2_1_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_1_reloadable6/Release/2_1_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_1_reloadable6/Release/2_1_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_1_reloadable6/Release/2_1_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_1_reloadable6/Release/2_1_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_1_reloadable6/Release/2_1_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_1_reloadable6/Release/2_1_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_1_reloadable6/Release/2_1_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_1_reloadable8/Release/2_1_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_1_reloadable8/Release/2_1_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_1_reloadable8/Release/2_1_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_1_reloadable8/Release/2_1_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_1_reloadable8/Release/2_1_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_1_reloadable8/Release/2_1_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_1_reloadable8/Release/2_1_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_1_reloadable8/Release/2_1_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_1_reloadable8/Release/2_1_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_1_reloadable10/Release/2_1_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_1_reloadable10/Release/2_1_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_1_reloadable10/Release/2_1_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_1_reloadable10/Release/2_1_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_1_reloadable10/Release/2_1_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_1_reloadable10/Release/2_1_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_1_reloadable10/Release/2_1_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_1_reloadable10/Release/2_1_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_1_reloadable10/Release/2_1_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_1_reloadable12/Release/2_1_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_1_reloadable12/Release/2_1_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_1_reloadable12/Release/2_1_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_1_reloadable12/Release/2_1_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_1_reloadable12/Release/2_1_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_1_reloadable12/Release/2_1_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_1_reloadable12/Release/2_1_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_1_reloadable12/Release/2_1_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_1_reloadable12/Release/2_1_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_1_reloadable14/Release/2_1_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_1_reloadable14/Release/2_1_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_1_reloadable14/Release/2_1_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_1_reloadable14/Release/2_1_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_1_reloadable14/Release/2_1_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_1_reloadable14/Release/2_1_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_1_reloadable14/Release/2_1_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_1_reloadable14/Release/2_1_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_1_reloadable14/Release/2_1_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_1_reloadable16/Release/2_1_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_1_reloadable16/Release/2_1_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_1_reloadable16/Release/2_1_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_1_reloadable16/Release/2_1_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_1_reloadable16/Release/2_1_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_1_reloadable16/Release/2_1_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_1_reloadable16/Release/2_1_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_1_reloadable16/Release/2_1_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_1_reloadable16/Release/2_1_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_1_reloadable18/Release/2_1_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_1_reloadable18/Release/2_1_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_1_reloadable18/Release/2_1_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_1_reloadable18/Release/2_1_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_1_reloadable18/Release/2_1_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_1_reloadable18/Release/2_1_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_1_reloadable18/Release/2_1_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_1_reloadable18/Release/2_1_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_1_reloadable18/Release/2_1_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_2_reloadable2/Release/2_2_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_2_reloadable2/Release/2_2_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_2_reloadable2/Release/2_2_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_2_reloadable2/Release/2_2_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_2_reloadable2/Release/2_2_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_2_reloadable2/Release/2_2_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_2_reloadable2/Release/2_2_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_2_reloadable2/Release/2_2_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_2_reloadable2/Release/2_2_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_2_reloadable4/Release/2_2_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_2_reloadable4/Release/2_2_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_2_reloadable4/Release/2_2_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_2_reloadable4/Release/2_2_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_2_reloadable4/Release/2_2_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_2_reloadable4/Release/2_2_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_2_reloadable4/Release/2_2_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_2_reloadable4/Release/2_2_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_2_reloadable4/Release/2_2_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_2_reloadable6/Release/2_2_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_2_reloadable6/Release/2_2_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_2_reloadable6/Release/2_2_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_2_reloadable6/Release/2_2_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_2_reloadable6/Release/2_2_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_2_reloadable6/Release/2_2_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_2_reloadable6/Release/2_2_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_2_reloadable6/Release/2_2_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_2_reloadable6/Release/2_2_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_2_reloadable8/Release/2_2_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_2_reloadable8/Release/2_2_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_2_reloadable8/Release/2_2_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_2_reloadable8/Release/2_2_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_2_reloadable8/Release/2_2_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_2_reloadable8/Release/2_2_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_2_reloadable8/Release/2_2_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_2_reloadable8/Release/2_2_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_2_reloadable8/Release/2_2_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_2_reloadable10/Release/2_2_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_2_reloadable10/Release/2_2_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_2_reloadable10/Release/2_2_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_2_reloadable10/Release/2_2_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_2_reloadable10/Release/2_2_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_2_reloadable10/Release/2_2_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_2_reloadable10/Release/2_2_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_2_reloadable10/Release/2_2_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_2_reloadable10/Release/2_2_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_2_reloadable12/Release/2_2_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_2_reloadable12/Release/2_2_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_2_reloadable12/Release/2_2_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_2_reloadable12/Release/2_2_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_2_reloadable12/Release/2_2_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_2_reloadable12/Release/2_2_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_2_reloadable12/Release/2_2_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_2_reloadable12/Release/2_2_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_2_reloadable12/Release/2_2_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_2_reloadable14/Release/2_2_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_2_reloadable14/Release/2_2_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_2_reloadable14/Release/2_2_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_2_reloadable14/Release/2_2_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_2_reloadable14/Release/2_2_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_2_reloadable14/Release/2_2_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_2_reloadable14/Release/2_2_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_2_reloadable14/Release/2_2_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_2_reloadable14/Release/2_2_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_2_reloadable16/Release/2_2_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_2_reloadable16/Release/2_2_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_2_reloadable16/Release/2_2_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_2_reloadable16/Release/2_2_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_2_reloadable16/Release/2_2_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_2_reloadable16/Release/2_2_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_2_reloadable16/Release/2_2_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_2_reloadable16/Release/2_2_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_2_reloadable16/Release/2_2_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_2_reloadable18/Release/2_2_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_2_reloadable18/Release/2_2_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_2_reloadable18/Release/2_2_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_2_reloadable18/Release/2_2_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_2_reloadable18/Release/2_2_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_2_reloadable18/Release/2_2_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_2_reloadable18/Release/2_2_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_2_reloadable18/Release/2_2_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_2_reloadable18/Release/2_2_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_3_reloadable2/Release/2_3_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_3_reloadable2/Release/2_3_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_3_reloadable2/Release/2_3_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_3_reloadable2/Release/2_3_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_3_reloadable2/Release/2_3_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_3_reloadable2/Release/2_3_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_3_reloadable2/Release/2_3_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_3_reloadable2/Release/2_3_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_3_reloadable2/Release/2_3_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_3_reloadable4/Release/2_3_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_3_reloadable4/Release/2_3_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_3_reloadable4/Release/2_3_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_3_reloadable4/Release/2_3_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_3_reloadable4/Release/2_3_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_3_reloadable4/Release/2_3_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_3_reloadable4/Release/2_3_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_3_reloadable4/Release/2_3_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_3_reloadable4/Release/2_3_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_3_reloadable6/Release/2_3_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_3_reloadable6/Release/2_3_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_3_reloadable6/Release/2_3_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_3_reloadable6/Release/2_3_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_3_reloadable6/Release/2_3_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_3_reloadable6/Release/2_3_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_3_reloadable6/Release/2_3_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_3_reloadable6/Release/2_3_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_3_reloadable6/Release/2_3_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_3_reloadable8/Release/2_3_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_3_reloadable8/Release/2_3_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_3_reloadable8/Release/2_3_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_3_reloadable8/Release/2_3_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_3_reloadable8/Release/2_3_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_3_reloadable8/Release/2_3_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_3_reloadable8/Release/2_3_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_3_reloadable8/Release/2_3_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_3_reloadable8/Release/2_3_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_3_reloadable10/Release/2_3_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_3_reloadable10/Release/2_3_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_3_reloadable10/Release/2_3_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_3_reloadable10/Release/2_3_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_3_reloadable10/Release/2_3_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_3_reloadable10/Release/2_3_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_3_reloadable10/Release/2_3_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_3_reloadable10/Release/2_3_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_3_reloadable10/Release/2_3_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_3_reloadable12/Release/2_3_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_3_reloadable12/Release/2_3_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_3_reloadable12/Release/2_3_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_3_reloadable12/Release/2_3_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_3_reloadable12/Release/2_3_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_3_reloadable12/Release/2_3_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_3_reloadable12/Release/2_3_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_3_reloadable12/Release/2_3_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_3_reloadable12/Release/2_3_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_3_reloadable14/Release/2_3_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_3_reloadable14/Release/2_3_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_3_reloadable14/Release/2_3_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_3_reloadable14/Release/2_3_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_3_reloadable14/Release/2_3_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_3_reloadable14/Release/2_3_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_3_reloadable14/Release/2_3_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_3_reloadable14/Release/2_3_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_3_reloadable14/Release/2_3_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_3_reloadable16/Release/2_3_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_3_reloadable16/Release/2_3_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_3_reloadable16/Release/2_3_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_3_reloadable16/Release/2_3_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_3_reloadable16/Release/2_3_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_3_reloadable16/Release/2_3_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_3_reloadable16/Release/2_3_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_3_reloadable16/Release/2_3_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_3_reloadable16/Release/2_3_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 2_3_reloadable18/Release/2_3_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 2_3_reloadable18/Release/2_3_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 2_3_reloadable18/Release/2_3_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 2_3_reloadable18/Release/2_3_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 2_3_reloadable18/Release/2_3_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 2_3_reloadable18/Release/2_3_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 2_3_reloadable18/Release/2_3_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 2_3_reloadable18/Release/2_3_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 2_3_reloadable18/Release/2_3_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_0_reloadable2/Release/3_0_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_0_reloadable2/Release/3_0_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_0_reloadable2/Release/3_0_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_0_reloadable2/Release/3_0_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_0_reloadable2/Release/3_0_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_0_reloadable2/Release/3_0_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_0_reloadable2/Release/3_0_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_0_reloadable2/Release/3_0_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_0_reloadable2/Release/3_0_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_0_reloadable4/Release/3_0_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_0_reloadable4/Release/3_0_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_0_reloadable4/Release/3_0_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_0_reloadable4/Release/3_0_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_0_reloadable4/Release/3_0_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_0_reloadable4/Release/3_0_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_0_reloadable4/Release/3_0_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_0_reloadable4/Release/3_0_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_0_reloadable4/Release/3_0_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_0_reloadable6/Release/3_0_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_0_reloadable6/Release/3_0_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_0_reloadable6/Release/3_0_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_0_reloadable6/Release/3_0_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_0_reloadable6/Release/3_0_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_0_reloadable6/Release/3_0_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_0_reloadable6/Release/3_0_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_0_reloadable6/Release/3_0_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_0_reloadable6/Release/3_0_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_0_reloadable8/Release/3_0_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_0_reloadable8/Release/3_0_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_0_reloadable8/Release/3_0_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_0_reloadable8/Release/3_0_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_0_reloadable8/Release/3_0_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_0_reloadable8/Release/3_0_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_0_reloadable8/Release/3_0_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_0_reloadable8/Release/3_0_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_0_reloadable8/Release/3_0_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_0_reloadable10/Release/3_0_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_0_reloadable10/Release/3_0_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_0_reloadable10/Release/3_0_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_0_reloadable10/Release/3_0_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_0_reloadable10/Release/3_0_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_0_reloadable10/Release/3_0_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_0_reloadable10/Release/3_0_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_0_reloadable10/Release/3_0_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_0_reloadable10/Release/3_0_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_0_reloadable12/Release/3_0_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_0_reloadable12/Release/3_0_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_0_reloadable12/Release/3_0_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_0_reloadable12/Release/3_0_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_0_reloadable12/Release/3_0_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_0_reloadable12/Release/3_0_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_0_reloadable12/Release/3_0_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_0_reloadable12/Release/3_0_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_0_reloadable12/Release/3_0_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_0_reloadable14/Release/3_0_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_0_reloadable14/Release/3_0_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_0_reloadable14/Release/3_0_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_0_reloadable14/Release/3_0_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_0_reloadable14/Release/3_0_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_0_reloadable14/Release/3_0_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_0_reloadable14/Release/3_0_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_0_reloadable14/Release/3_0_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_0_reloadable14/Release/3_0_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_0_reloadable16/Release/3_0_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_0_reloadable16/Release/3_0_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_0_reloadable16/Release/3_0_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_0_reloadable16/Release/3_0_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_0_reloadable16/Release/3_0_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_0_reloadable16/Release/3_0_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_0_reloadable16/Release/3_0_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_0_reloadable16/Release/3_0_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_0_reloadable16/Release/3_0_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_0_reloadable18/Release/3_0_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_0_reloadable18/Release/3_0_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_0_reloadable18/Release/3_0_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_0_reloadable18/Release/3_0_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_0_reloadable18/Release/3_0_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_0_reloadable18/Release/3_0_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_0_reloadable18/Release/3_0_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_0_reloadable18/Release/3_0_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_0_reloadable18/Release/3_0_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_1_reloadable2/Release/3_1_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_1_reloadable2/Release/3_1_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_1_reloadable2/Release/3_1_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_1_reloadable2/Release/3_1_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_1_reloadable2/Release/3_1_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_1_reloadable2/Release/3_1_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_1_reloadable2/Release/3_1_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_1_reloadable2/Release/3_1_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_1_reloadable2/Release/3_1_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_1_reloadable4/Release/3_1_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_1_reloadable4/Release/3_1_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_1_reloadable4/Release/3_1_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_1_reloadable4/Release/3_1_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_1_reloadable4/Release/3_1_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_1_reloadable4/Release/3_1_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_1_reloadable4/Release/3_1_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_1_reloadable4/Release/3_1_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_1_reloadable4/Release/3_1_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_1_reloadable6/Release/3_1_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_1_reloadable6/Release/3_1_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_1_reloadable6/Release/3_1_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_1_reloadable6/Release/3_1_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_1_reloadable6/Release/3_1_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_1_reloadable6/Release/3_1_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_1_reloadable6/Release/3_1_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_1_reloadable6/Release/3_1_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_1_reloadable6/Release/3_1_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_1_reloadable8/Release/3_1_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_1_reloadable8/Release/3_1_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_1_reloadable8/Release/3_1_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_1_reloadable8/Release/3_1_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_1_reloadable8/Release/3_1_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_1_reloadable8/Release/3_1_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_1_reloadable8/Release/3_1_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_1_reloadable8/Release/3_1_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_1_reloadable8/Release/3_1_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_1_reloadable10/Release/3_1_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_1_reloadable10/Release/3_1_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_1_reloadable10/Release/3_1_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_1_reloadable10/Release/3_1_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_1_reloadable10/Release/3_1_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_1_reloadable10/Release/3_1_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_1_reloadable10/Release/3_1_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_1_reloadable10/Release/3_1_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_1_reloadable10/Release/3_1_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_1_reloadable12/Release/3_1_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_1_reloadable12/Release/3_1_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_1_reloadable12/Release/3_1_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_1_reloadable12/Release/3_1_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_1_reloadable12/Release/3_1_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_1_reloadable12/Release/3_1_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_1_reloadable12/Release/3_1_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_1_reloadable12/Release/3_1_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_1_reloadable12/Release/3_1_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_1_reloadable14/Release/3_1_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_1_reloadable14/Release/3_1_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_1_reloadable14/Release/3_1_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_1_reloadable14/Release/3_1_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_1_reloadable14/Release/3_1_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_1_reloadable14/Release/3_1_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_1_reloadable14/Release/3_1_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_1_reloadable14/Release/3_1_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_1_reloadable14/Release/3_1_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_1_reloadable16/Release/3_1_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_1_reloadable16/Release/3_1_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_1_reloadable16/Release/3_1_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_1_reloadable16/Release/3_1_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_1_reloadable16/Release/3_1_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_1_reloadable16/Release/3_1_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_1_reloadable16/Release/3_1_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_1_reloadable16/Release/3_1_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_1_reloadable16/Release/3_1_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_1_reloadable18/Release/3_1_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_1_reloadable18/Release/3_1_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_1_reloadable18/Release/3_1_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_1_reloadable18/Release/3_1_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_1_reloadable18/Release/3_1_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_1_reloadable18/Release/3_1_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_1_reloadable18/Release/3_1_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_1_reloadable18/Release/3_1_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_1_reloadable18/Release/3_1_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_2_reloadable2/Release/3_2_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_2_reloadable2/Release/3_2_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_2_reloadable2/Release/3_2_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_2_reloadable2/Release/3_2_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_2_reloadable2/Release/3_2_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_2_reloadable2/Release/3_2_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_2_reloadable2/Release/3_2_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_2_reloadable2/Release/3_2_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_2_reloadable2/Release/3_2_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_2_reloadable4/Release/3_2_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_2_reloadable4/Release/3_2_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_2_reloadable4/Release/3_2_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_2_reloadable4/Release/3_2_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_2_reloadable4/Release/3_2_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_2_reloadable4/Release/3_2_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_2_reloadable4/Release/3_2_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_2_reloadable4/Release/3_2_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_2_reloadable4/Release/3_2_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_2_reloadable6/Release/3_2_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_2_reloadable6/Release/3_2_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_2_reloadable6/Release/3_2_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_2_reloadable6/Release/3_2_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_2_reloadable6/Release/3_2_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_2_reloadable6/Release/3_2_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_2_reloadable6/Release/3_2_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_2_reloadable6/Release/3_2_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_2_reloadable6/Release/3_2_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_2_reloadable8/Release/3_2_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_2_reloadable8/Release/3_2_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_2_reloadable8/Release/3_2_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_2_reloadable8/Release/3_2_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_2_reloadable8/Release/3_2_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_2_reloadable8/Release/3_2_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_2_reloadable8/Release/3_2_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_2_reloadable8/Release/3_2_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_2_reloadable8/Release/3_2_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_2_reloadable10/Release/3_2_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_2_reloadable10/Release/3_2_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_2_reloadable10/Release/3_2_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_2_reloadable10/Release/3_2_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_2_reloadable10/Release/3_2_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_2_reloadable10/Release/3_2_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_2_reloadable10/Release/3_2_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_2_reloadable10/Release/3_2_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_2_reloadable10/Release/3_2_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_2_reloadable12/Release/3_2_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_2_reloadable12/Release/3_2_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_2_reloadable12/Release/3_2_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_2_reloadable12/Release/3_2_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_2_reloadable12/Release/3_2_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_2_reloadable12/Release/3_2_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_2_reloadable12/Release/3_2_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_2_reloadable12/Release/3_2_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_2_reloadable12/Release/3_2_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_2_reloadable14/Release/3_2_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_2_reloadable14/Release/3_2_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_2_reloadable14/Release/3_2_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_2_reloadable14/Release/3_2_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_2_reloadable14/Release/3_2_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_2_reloadable14/Release/3_2_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_2_reloadable14/Release/3_2_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_2_reloadable14/Release/3_2_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_2_reloadable14/Release/3_2_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_2_reloadable16/Release/3_2_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_2_reloadable16/Release/3_2_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_2_reloadable16/Release/3_2_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_2_reloadable16/Release/3_2_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_2_reloadable16/Release/3_2_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_2_reloadable16/Release/3_2_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_2_reloadable16/Release/3_2_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_2_reloadable16/Release/3_2_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_2_reloadable16/Release/3_2_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_2_reloadable18/Release/3_2_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_2_reloadable18/Release/3_2_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_2_reloadable18/Release/3_2_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_2_reloadable18/Release/3_2_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_2_reloadable18/Release/3_2_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_2_reloadable18/Release/3_2_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_2_reloadable18/Release/3_2_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_2_reloadable18/Release/3_2_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_2_reloadable18/Release/3_2_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable2/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_3_reloadable2/Release/3_3_reloadable2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_3_reloadable2/Release/3_3_reloadable2.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_3_reloadable2/Release/3_3_reloadable2.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_3_reloadable2/Release/3_3_reloadable2.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_3_reloadable2/Release/3_3_reloadable2.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_3_reloadable2/Release/3_3_reloadable2.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_3_reloadable2/Release/3_3_reloadable2.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_3_reloadable2/Release/3_3_reloadable2.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_3_reloadable2/Release/3_3_reloadable2.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable4/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_3_reloadable4/Release/3_3_reloadable4 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_3_reloadable4/Release/3_3_reloadable4.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_3_reloadable4/Release/3_3_reloadable4.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_3_reloadable4/Release/3_3_reloadable4.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_3_reloadable4/Release/3_3_reloadable4.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_3_reloadable4/Release/3_3_reloadable4.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_3_reloadable4/Release/3_3_reloadable4.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_3_reloadable4/Release/3_3_reloadable4.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_3_reloadable4/Release/3_3_reloadable4.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable6/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_3_reloadable6/Release/3_3_reloadable6 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_3_reloadable6/Release/3_3_reloadable6.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_3_reloadable6/Release/3_3_reloadable6.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_3_reloadable6/Release/3_3_reloadable6.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_3_reloadable6/Release/3_3_reloadable6.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_3_reloadable6/Release/3_3_reloadable6.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_3_reloadable6/Release/3_3_reloadable6.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_3_reloadable6/Release/3_3_reloadable6.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_3_reloadable6/Release/3_3_reloadable6.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable8/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_3_reloadable8/Release/3_3_reloadable8 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_3_reloadable8/Release/3_3_reloadable8.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_3_reloadable8/Release/3_3_reloadable8.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_3_reloadable8/Release/3_3_reloadable8.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_3_reloadable8/Release/3_3_reloadable8.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_3_reloadable8/Release/3_3_reloadable8.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_3_reloadable8/Release/3_3_reloadable8.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_3_reloadable8/Release/3_3_reloadable8.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_3_reloadable8/Release/3_3_reloadable8.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable10/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_3_reloadable10/Release/3_3_reloadable10 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_3_reloadable10/Release/3_3_reloadable10.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_3_reloadable10/Release/3_3_reloadable10.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_3_reloadable10/Release/3_3_reloadable10.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_3_reloadable10/Release/3_3_reloadable10.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_3_reloadable10/Release/3_3_reloadable10.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_3_reloadable10/Release/3_3_reloadable10.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_3_reloadable10/Release/3_3_reloadable10.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_3_reloadable10/Release/3_3_reloadable10.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable12/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_3_reloadable12/Release/3_3_reloadable12 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_3_reloadable12/Release/3_3_reloadable12.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_3_reloadable12/Release/3_3_reloadable12.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_3_reloadable12/Release/3_3_reloadable12.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_3_reloadable12/Release/3_3_reloadable12.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_3_reloadable12/Release/3_3_reloadable12.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_3_reloadable12/Release/3_3_reloadable12.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_3_reloadable12/Release/3_3_reloadable12.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_3_reloadable12/Release/3_3_reloadable12.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable14/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_3_reloadable14/Release/3_3_reloadable14 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_3_reloadable14/Release/3_3_reloadable14.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_3_reloadable14/Release/3_3_reloadable14.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_3_reloadable14/Release/3_3_reloadable14.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_3_reloadable14/Release/3_3_reloadable14.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_3_reloadable14/Release/3_3_reloadable14.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_3_reloadable14/Release/3_3_reloadable14.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_3_reloadable14/Release/3_3_reloadable14.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_3_reloadable14/Release/3_3_reloadable14.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable16/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_3_reloadable16/Release/3_3_reloadable16 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_3_reloadable16/Release/3_3_reloadable16.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_3_reloadable16/Release/3_3_reloadable16.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_3_reloadable16/Release/3_3_reloadable16.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_3_reloadable16/Release/3_3_reloadable16.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_3_reloadable16/Release/3_3_reloadable16.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_3_reloadable16/Release/3_3_reloadable16.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_3_reloadable16/Release/3_3_reloadable16.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_3_reloadable16/Release/3_3_reloadable16.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable18/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2 3_3_reloadable18/Release/3_3_reloadable18 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.map 3_3_reloadable18/Release/3_3_reloadable18.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.lst 3_3_reloadable18/Release/3_3_reloadable18.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.calltree 3_3_reloadable18/Release/3_3_reloadable18.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.sdr 3_3_reloadable18/Release/3_3_reloadable18.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.srv 3_3_reloadable18/Release/3_3_reloadable18.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.txt 3_3_reloadable18/Release/3_3_reloadable18.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmic2 3_3_reloadable18/Release/3_3_reloadable18.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable2/Release/0_0_reloadable2.cmico 3_3_reloadable18/Release/3_3_reloadable18.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable3.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable3/scripts/0_0_reloadable3.prx) 2>&1 |& tee -a 0_0_reloadable3/0_0_reloadable3.log 0_0_reloadable3/timestamped_log/0_0_reloadable3.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable3.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable3.ll -o../Release/chesswork3342/0_0_reloadable3.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork3342 ../Release/chesswork3342/0_0_reloadable3.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable3-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +chess-backend 0_0_reloadable3-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +chess-backend 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +chess-backend 0_0_reloadable3-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--mist1 -k64 --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable3-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable3-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable3-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable3-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable3-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable3-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable3-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable3-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable3 -L --common 0_0_reloadable3-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable3.objlist -o../0_0_reloadable3.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable3.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable3" +bridge -o../Release/0_0_reloadable3 ../Release/0_0_reloadable3.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable3.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork3342 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable3 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (180 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable3/Release/0_0_reloadable3 >> 0_0_reloadable3/Release/0_0_reloadable3.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_0_reloadable9/Release/0_0_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_0_reloadable9/Release/0_0_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_0_reloadable9/Release/0_0_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_0_reloadable9/Release/0_0_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_0_reloadable9/Release/0_0_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_0_reloadable9/Release/0_0_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_0_reloadable9/Release/0_0_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_0_reloadable9/Release/0_0_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_0_reloadable9/Release/0_0_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_0_reloadable13/Release/0_0_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_0_reloadable13/Release/0_0_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_0_reloadable13/Release/0_0_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_0_reloadable13/Release/0_0_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_0_reloadable13/Release/0_0_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_0_reloadable13/Release/0_0_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_0_reloadable13/Release/0_0_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_0_reloadable13/Release/0_0_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_0_reloadable13/Release/0_0_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_1_reloadable3/Release/0_1_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_1_reloadable3/Release/0_1_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_1_reloadable3/Release/0_1_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_1_reloadable3/Release/0_1_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_1_reloadable3/Release/0_1_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_1_reloadable3/Release/0_1_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_1_reloadable3/Release/0_1_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_1_reloadable3/Release/0_1_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_1_reloadable3/Release/0_1_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_1_reloadable9/Release/0_1_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_1_reloadable9/Release/0_1_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_1_reloadable9/Release/0_1_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_1_reloadable9/Release/0_1_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_1_reloadable9/Release/0_1_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_1_reloadable9/Release/0_1_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_1_reloadable9/Release/0_1_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_1_reloadable9/Release/0_1_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_1_reloadable9/Release/0_1_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_1_reloadable13/Release/0_1_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_1_reloadable13/Release/0_1_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_1_reloadable13/Release/0_1_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_1_reloadable13/Release/0_1_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_1_reloadable13/Release/0_1_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_1_reloadable13/Release/0_1_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_1_reloadable13/Release/0_1_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_1_reloadable13/Release/0_1_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_1_reloadable13/Release/0_1_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_2_reloadable3/Release/0_2_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_2_reloadable3/Release/0_2_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_2_reloadable3/Release/0_2_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_2_reloadable3/Release/0_2_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_2_reloadable3/Release/0_2_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_2_reloadable3/Release/0_2_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_2_reloadable3/Release/0_2_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_2_reloadable3/Release/0_2_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_2_reloadable3/Release/0_2_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_2_reloadable9/Release/0_2_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_2_reloadable9/Release/0_2_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_2_reloadable9/Release/0_2_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_2_reloadable9/Release/0_2_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_2_reloadable9/Release/0_2_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_2_reloadable9/Release/0_2_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_2_reloadable9/Release/0_2_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_2_reloadable9/Release/0_2_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_2_reloadable9/Release/0_2_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_2_reloadable13/Release/0_2_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_2_reloadable13/Release/0_2_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_2_reloadable13/Release/0_2_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_2_reloadable13/Release/0_2_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_2_reloadable13/Release/0_2_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_2_reloadable13/Release/0_2_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_2_reloadable13/Release/0_2_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_2_reloadable13/Release/0_2_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_2_reloadable13/Release/0_2_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_3_reloadable3/Release/0_3_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_3_reloadable3/Release/0_3_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_3_reloadable3/Release/0_3_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_3_reloadable3/Release/0_3_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_3_reloadable3/Release/0_3_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_3_reloadable3/Release/0_3_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_3_reloadable3/Release/0_3_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_3_reloadable3/Release/0_3_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_3_reloadable3/Release/0_3_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_3_reloadable9/Release/0_3_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_3_reloadable9/Release/0_3_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_3_reloadable9/Release/0_3_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_3_reloadable9/Release/0_3_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_3_reloadable9/Release/0_3_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_3_reloadable9/Release/0_3_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_3_reloadable9/Release/0_3_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_3_reloadable9/Release/0_3_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_3_reloadable9/Release/0_3_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 0_3_reloadable13/Release/0_3_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 0_3_reloadable13/Release/0_3_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 0_3_reloadable13/Release/0_3_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 0_3_reloadable13/Release/0_3_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 0_3_reloadable13/Release/0_3_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 0_3_reloadable13/Release/0_3_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 0_3_reloadable13/Release/0_3_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 0_3_reloadable13/Release/0_3_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 0_3_reloadable13/Release/0_3_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_0_reloadable3/Release/1_0_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_0_reloadable3/Release/1_0_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_0_reloadable3/Release/1_0_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_0_reloadable3/Release/1_0_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_0_reloadable3/Release/1_0_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_0_reloadable3/Release/1_0_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_0_reloadable3/Release/1_0_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_0_reloadable3/Release/1_0_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_0_reloadable3/Release/1_0_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_0_reloadable9/Release/1_0_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_0_reloadable9/Release/1_0_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_0_reloadable9/Release/1_0_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_0_reloadable9/Release/1_0_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_0_reloadable9/Release/1_0_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_0_reloadable9/Release/1_0_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_0_reloadable9/Release/1_0_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_0_reloadable9/Release/1_0_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_0_reloadable9/Release/1_0_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_0_reloadable13/Release/1_0_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_0_reloadable13/Release/1_0_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_0_reloadable13/Release/1_0_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_0_reloadable13/Release/1_0_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_0_reloadable13/Release/1_0_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_0_reloadable13/Release/1_0_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_0_reloadable13/Release/1_0_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_0_reloadable13/Release/1_0_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_0_reloadable13/Release/1_0_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_1_reloadable3/Release/1_1_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_1_reloadable3/Release/1_1_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_1_reloadable3/Release/1_1_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_1_reloadable3/Release/1_1_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_1_reloadable3/Release/1_1_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_1_reloadable3/Release/1_1_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_1_reloadable3/Release/1_1_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_1_reloadable3/Release/1_1_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_1_reloadable3/Release/1_1_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_1_reloadable9/Release/1_1_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_1_reloadable9/Release/1_1_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_1_reloadable9/Release/1_1_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_1_reloadable9/Release/1_1_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_1_reloadable9/Release/1_1_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_1_reloadable9/Release/1_1_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_1_reloadable9/Release/1_1_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_1_reloadable9/Release/1_1_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_1_reloadable9/Release/1_1_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_1_reloadable13/Release/1_1_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_1_reloadable13/Release/1_1_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_1_reloadable13/Release/1_1_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_1_reloadable13/Release/1_1_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_1_reloadable13/Release/1_1_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_1_reloadable13/Release/1_1_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_1_reloadable13/Release/1_1_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_1_reloadable13/Release/1_1_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_1_reloadable13/Release/1_1_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_2_reloadable3/Release/1_2_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_2_reloadable3/Release/1_2_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_2_reloadable3/Release/1_2_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_2_reloadable3/Release/1_2_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_2_reloadable3/Release/1_2_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_2_reloadable3/Release/1_2_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_2_reloadable3/Release/1_2_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_2_reloadable3/Release/1_2_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_2_reloadable3/Release/1_2_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_2_reloadable9/Release/1_2_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_2_reloadable9/Release/1_2_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_2_reloadable9/Release/1_2_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_2_reloadable9/Release/1_2_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_2_reloadable9/Release/1_2_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_2_reloadable9/Release/1_2_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_2_reloadable9/Release/1_2_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_2_reloadable9/Release/1_2_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_2_reloadable9/Release/1_2_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_2_reloadable13/Release/1_2_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_2_reloadable13/Release/1_2_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_2_reloadable13/Release/1_2_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_2_reloadable13/Release/1_2_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_2_reloadable13/Release/1_2_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_2_reloadable13/Release/1_2_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_2_reloadable13/Release/1_2_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_2_reloadable13/Release/1_2_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_2_reloadable13/Release/1_2_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_3_reloadable3/Release/1_3_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_3_reloadable3/Release/1_3_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_3_reloadable3/Release/1_3_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_3_reloadable3/Release/1_3_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_3_reloadable3/Release/1_3_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_3_reloadable3/Release/1_3_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_3_reloadable3/Release/1_3_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_3_reloadable3/Release/1_3_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_3_reloadable3/Release/1_3_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_3_reloadable9/Release/1_3_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_3_reloadable9/Release/1_3_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_3_reloadable9/Release/1_3_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_3_reloadable9/Release/1_3_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_3_reloadable9/Release/1_3_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_3_reloadable9/Release/1_3_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_3_reloadable9/Release/1_3_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_3_reloadable9/Release/1_3_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_3_reloadable9/Release/1_3_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 1_3_reloadable13/Release/1_3_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 1_3_reloadable13/Release/1_3_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 1_3_reloadable13/Release/1_3_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 1_3_reloadable13/Release/1_3_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 1_3_reloadable13/Release/1_3_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 1_3_reloadable13/Release/1_3_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 1_3_reloadable13/Release/1_3_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 1_3_reloadable13/Release/1_3_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 1_3_reloadable13/Release/1_3_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_0_reloadable3/Release/2_0_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_0_reloadable3/Release/2_0_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_0_reloadable3/Release/2_0_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_0_reloadable3/Release/2_0_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_0_reloadable3/Release/2_0_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_0_reloadable3/Release/2_0_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_0_reloadable3/Release/2_0_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_0_reloadable3/Release/2_0_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_0_reloadable3/Release/2_0_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_0_reloadable9/Release/2_0_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_0_reloadable9/Release/2_0_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_0_reloadable9/Release/2_0_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_0_reloadable9/Release/2_0_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_0_reloadable9/Release/2_0_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_0_reloadable9/Release/2_0_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_0_reloadable9/Release/2_0_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_0_reloadable9/Release/2_0_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_0_reloadable9/Release/2_0_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_0_reloadable13/Release/2_0_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_0_reloadable13/Release/2_0_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_0_reloadable13/Release/2_0_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_0_reloadable13/Release/2_0_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_0_reloadable13/Release/2_0_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_0_reloadable13/Release/2_0_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_0_reloadable13/Release/2_0_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_0_reloadable13/Release/2_0_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_0_reloadable13/Release/2_0_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_1_reloadable3/Release/2_1_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_1_reloadable3/Release/2_1_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_1_reloadable3/Release/2_1_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_1_reloadable3/Release/2_1_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_1_reloadable3/Release/2_1_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_1_reloadable3/Release/2_1_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_1_reloadable3/Release/2_1_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_1_reloadable3/Release/2_1_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_1_reloadable3/Release/2_1_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_1_reloadable9/Release/2_1_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_1_reloadable9/Release/2_1_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_1_reloadable9/Release/2_1_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_1_reloadable9/Release/2_1_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_1_reloadable9/Release/2_1_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_1_reloadable9/Release/2_1_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_1_reloadable9/Release/2_1_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_1_reloadable9/Release/2_1_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_1_reloadable9/Release/2_1_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_1_reloadable13/Release/2_1_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_1_reloadable13/Release/2_1_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_1_reloadable13/Release/2_1_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_1_reloadable13/Release/2_1_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_1_reloadable13/Release/2_1_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_1_reloadable13/Release/2_1_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_1_reloadable13/Release/2_1_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_1_reloadable13/Release/2_1_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_1_reloadable13/Release/2_1_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_2_reloadable3/Release/2_2_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_2_reloadable3/Release/2_2_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_2_reloadable3/Release/2_2_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_2_reloadable3/Release/2_2_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_2_reloadable3/Release/2_2_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_2_reloadable3/Release/2_2_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_2_reloadable3/Release/2_2_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_2_reloadable3/Release/2_2_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_2_reloadable3/Release/2_2_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_2_reloadable9/Release/2_2_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_2_reloadable9/Release/2_2_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_2_reloadable9/Release/2_2_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_2_reloadable9/Release/2_2_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_2_reloadable9/Release/2_2_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_2_reloadable9/Release/2_2_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_2_reloadable9/Release/2_2_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_2_reloadable9/Release/2_2_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_2_reloadable9/Release/2_2_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_2_reloadable13/Release/2_2_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_2_reloadable13/Release/2_2_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_2_reloadable13/Release/2_2_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_2_reloadable13/Release/2_2_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_2_reloadable13/Release/2_2_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_2_reloadable13/Release/2_2_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_2_reloadable13/Release/2_2_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_2_reloadable13/Release/2_2_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_2_reloadable13/Release/2_2_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_3_reloadable3/Release/2_3_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_3_reloadable3/Release/2_3_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_3_reloadable3/Release/2_3_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_3_reloadable3/Release/2_3_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_3_reloadable3/Release/2_3_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_3_reloadable3/Release/2_3_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_3_reloadable3/Release/2_3_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_3_reloadable3/Release/2_3_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_3_reloadable3/Release/2_3_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_3_reloadable9/Release/2_3_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_3_reloadable9/Release/2_3_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_3_reloadable9/Release/2_3_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_3_reloadable9/Release/2_3_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_3_reloadable9/Release/2_3_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_3_reloadable9/Release/2_3_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_3_reloadable9/Release/2_3_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_3_reloadable9/Release/2_3_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_3_reloadable9/Release/2_3_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 2_3_reloadable13/Release/2_3_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 2_3_reloadable13/Release/2_3_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 2_3_reloadable13/Release/2_3_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 2_3_reloadable13/Release/2_3_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 2_3_reloadable13/Release/2_3_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 2_3_reloadable13/Release/2_3_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 2_3_reloadable13/Release/2_3_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 2_3_reloadable13/Release/2_3_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 2_3_reloadable13/Release/2_3_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_0_reloadable3/Release/3_0_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_0_reloadable3/Release/3_0_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_0_reloadable3/Release/3_0_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_0_reloadable3/Release/3_0_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_0_reloadable3/Release/3_0_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_0_reloadable3/Release/3_0_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_0_reloadable3/Release/3_0_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_0_reloadable3/Release/3_0_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_0_reloadable3/Release/3_0_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_0_reloadable9/Release/3_0_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_0_reloadable9/Release/3_0_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_0_reloadable9/Release/3_0_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_0_reloadable9/Release/3_0_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_0_reloadable9/Release/3_0_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_0_reloadable9/Release/3_0_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_0_reloadable9/Release/3_0_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_0_reloadable9/Release/3_0_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_0_reloadable9/Release/3_0_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_0_reloadable13/Release/3_0_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_0_reloadable13/Release/3_0_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_0_reloadable13/Release/3_0_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_0_reloadable13/Release/3_0_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_0_reloadable13/Release/3_0_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_0_reloadable13/Release/3_0_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_0_reloadable13/Release/3_0_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_0_reloadable13/Release/3_0_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_0_reloadable13/Release/3_0_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_1_reloadable3/Release/3_1_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_1_reloadable3/Release/3_1_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_1_reloadable3/Release/3_1_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_1_reloadable3/Release/3_1_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_1_reloadable3/Release/3_1_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_1_reloadable3/Release/3_1_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_1_reloadable3/Release/3_1_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_1_reloadable3/Release/3_1_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_1_reloadable3/Release/3_1_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_1_reloadable9/Release/3_1_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_1_reloadable9/Release/3_1_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_1_reloadable9/Release/3_1_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_1_reloadable9/Release/3_1_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_1_reloadable9/Release/3_1_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_1_reloadable9/Release/3_1_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_1_reloadable9/Release/3_1_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_1_reloadable9/Release/3_1_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_1_reloadable9/Release/3_1_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_1_reloadable13/Release/3_1_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_1_reloadable13/Release/3_1_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_1_reloadable13/Release/3_1_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_1_reloadable13/Release/3_1_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_1_reloadable13/Release/3_1_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_1_reloadable13/Release/3_1_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_1_reloadable13/Release/3_1_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_1_reloadable13/Release/3_1_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_1_reloadable13/Release/3_1_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_2_reloadable3/Release/3_2_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_2_reloadable3/Release/3_2_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_2_reloadable3/Release/3_2_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_2_reloadable3/Release/3_2_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_2_reloadable3/Release/3_2_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_2_reloadable3/Release/3_2_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_2_reloadable3/Release/3_2_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_2_reloadable3/Release/3_2_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_2_reloadable3/Release/3_2_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_2_reloadable9/Release/3_2_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_2_reloadable9/Release/3_2_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_2_reloadable9/Release/3_2_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_2_reloadable9/Release/3_2_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_2_reloadable9/Release/3_2_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_2_reloadable9/Release/3_2_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_2_reloadable9/Release/3_2_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_2_reloadable9/Release/3_2_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_2_reloadable9/Release/3_2_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_2_reloadable13/Release/3_2_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_2_reloadable13/Release/3_2_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_2_reloadable13/Release/3_2_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_2_reloadable13/Release/3_2_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_2_reloadable13/Release/3_2_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_2_reloadable13/Release/3_2_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_2_reloadable13/Release/3_2_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_2_reloadable13/Release/3_2_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_2_reloadable13/Release/3_2_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable3/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_3_reloadable3/Release/3_3_reloadable3 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_3_reloadable3/Release/3_3_reloadable3.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_3_reloadable3/Release/3_3_reloadable3.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_3_reloadable3/Release/3_3_reloadable3.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_3_reloadable3/Release/3_3_reloadable3.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_3_reloadable3/Release/3_3_reloadable3.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_3_reloadable3/Release/3_3_reloadable3.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_3_reloadable3/Release/3_3_reloadable3.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_3_reloadable3/Release/3_3_reloadable3.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable9/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_3_reloadable9/Release/3_3_reloadable9 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_3_reloadable9/Release/3_3_reloadable9.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_3_reloadable9/Release/3_3_reloadable9.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_3_reloadable9/Release/3_3_reloadable9.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_3_reloadable9/Release/3_3_reloadable9.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_3_reloadable9/Release/3_3_reloadable9.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_3_reloadable9/Release/3_3_reloadable9.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_3_reloadable9/Release/3_3_reloadable9.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_3_reloadable9/Release/3_3_reloadable9.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable13/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3 3_3_reloadable13/Release/3_3_reloadable13 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.map 3_3_reloadable13/Release/3_3_reloadable13.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.lst 3_3_reloadable13/Release/3_3_reloadable13.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.calltree 3_3_reloadable13/Release/3_3_reloadable13.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.sdr 3_3_reloadable13/Release/3_3_reloadable13.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.srv 3_3_reloadable13/Release/3_3_reloadable13.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.txt 3_3_reloadable13/Release/3_3_reloadable13.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmic2 3_3_reloadable13/Release/3_3_reloadable13.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable3/Release/0_0_reloadable3.cmico 3_3_reloadable13/Release/3_3_reloadable13.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable5.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable5/scripts/0_0_reloadable5.prx) 2>&1 |& tee -a 0_0_reloadable5/0_0_reloadable5.log 0_0_reloadable5/timestamped_log/0_0_reloadable5.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable5.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable5.ll -o../Release/chesswork4008/0_0_reloadable5.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork4008 ../Release/chesswork4008/0_0_reloadable5.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable5-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +chess-backend 0_0_reloadable5-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +chess-backend 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +chess-backend 0_0_reloadable5-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable5-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable5-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable5-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT1_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--mist1 -k64 --common 0_0_reloadable5-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable5-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEERN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE_N_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyn_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable5-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable5-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable5-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable5-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable5-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable5-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable5-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_Z13_b919_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable5-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable5 -L --common 0_0_reloadable5-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable5.objlist -o../0_0_reloadable5.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable5.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable5" +bridge -o../Release/0_0_reloadable5 ../Release/0_0_reloadable5.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable5.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork4008 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable5 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (183 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable5/Release/0_0_reloadable5 >> 0_0_reloadable5/Release/0_0_reloadable5.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_0_reloadable7/Release/0_0_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_0_reloadable7/Release/0_0_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_0_reloadable7/Release/0_0_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_0_reloadable7/Release/0_0_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_0_reloadable7/Release/0_0_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_0_reloadable7/Release/0_0_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_0_reloadable7/Release/0_0_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_0_reloadable7/Release/0_0_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_0_reloadable7/Release/0_0_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_0_reloadable11/Release/0_0_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_0_reloadable11/Release/0_0_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_0_reloadable11/Release/0_0_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_0_reloadable11/Release/0_0_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_0_reloadable11/Release/0_0_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_0_reloadable11/Release/0_0_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_0_reloadable11/Release/0_0_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_0_reloadable11/Release/0_0_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_0_reloadable11/Release/0_0_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_0_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_0_reloadable15/Release/0_0_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_0_reloadable15/Release/0_0_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_0_reloadable15/Release/0_0_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_0_reloadable15/Release/0_0_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_0_reloadable15/Release/0_0_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_0_reloadable15/Release/0_0_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_0_reloadable15/Release/0_0_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_0_reloadable15/Release/0_0_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_0_reloadable15/Release/0_0_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_1_reloadable5/Release/0_1_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_1_reloadable5/Release/0_1_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_1_reloadable5/Release/0_1_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_1_reloadable5/Release/0_1_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_1_reloadable5/Release/0_1_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_1_reloadable5/Release/0_1_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_1_reloadable5/Release/0_1_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_1_reloadable5/Release/0_1_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_1_reloadable5/Release/0_1_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_1_reloadable7/Release/0_1_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_1_reloadable7/Release/0_1_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_1_reloadable7/Release/0_1_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_1_reloadable7/Release/0_1_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_1_reloadable7/Release/0_1_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_1_reloadable7/Release/0_1_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_1_reloadable7/Release/0_1_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_1_reloadable7/Release/0_1_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_1_reloadable7/Release/0_1_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_1_reloadable11/Release/0_1_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_1_reloadable11/Release/0_1_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_1_reloadable11/Release/0_1_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_1_reloadable11/Release/0_1_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_1_reloadable11/Release/0_1_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_1_reloadable11/Release/0_1_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_1_reloadable11/Release/0_1_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_1_reloadable11/Release/0_1_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_1_reloadable11/Release/0_1_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_1_reloadable15/Release/0_1_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_1_reloadable15/Release/0_1_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_1_reloadable15/Release/0_1_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_1_reloadable15/Release/0_1_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_1_reloadable15/Release/0_1_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_1_reloadable15/Release/0_1_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_1_reloadable15/Release/0_1_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_1_reloadable15/Release/0_1_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_1_reloadable15/Release/0_1_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_2_reloadable5/Release/0_2_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_2_reloadable5/Release/0_2_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_2_reloadable5/Release/0_2_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_2_reloadable5/Release/0_2_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_2_reloadable5/Release/0_2_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_2_reloadable5/Release/0_2_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_2_reloadable5/Release/0_2_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_2_reloadable5/Release/0_2_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_2_reloadable5/Release/0_2_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_2_reloadable7/Release/0_2_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_2_reloadable7/Release/0_2_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_2_reloadable7/Release/0_2_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_2_reloadable7/Release/0_2_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_2_reloadable7/Release/0_2_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_2_reloadable7/Release/0_2_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_2_reloadable7/Release/0_2_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_2_reloadable7/Release/0_2_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_2_reloadable7/Release/0_2_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_2_reloadable11/Release/0_2_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_2_reloadable11/Release/0_2_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_2_reloadable11/Release/0_2_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_2_reloadable11/Release/0_2_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_2_reloadable11/Release/0_2_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_2_reloadable11/Release/0_2_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_2_reloadable11/Release/0_2_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_2_reloadable11/Release/0_2_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_2_reloadable11/Release/0_2_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_2_reloadable15/Release/0_2_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_2_reloadable15/Release/0_2_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_2_reloadable15/Release/0_2_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_2_reloadable15/Release/0_2_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_2_reloadable15/Release/0_2_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_2_reloadable15/Release/0_2_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_2_reloadable15/Release/0_2_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_2_reloadable15/Release/0_2_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_2_reloadable15/Release/0_2_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_3_reloadable5/Release/0_3_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_3_reloadable5/Release/0_3_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_3_reloadable5/Release/0_3_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_3_reloadable5/Release/0_3_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_3_reloadable5/Release/0_3_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_3_reloadable5/Release/0_3_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_3_reloadable5/Release/0_3_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_3_reloadable5/Release/0_3_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_3_reloadable5/Release/0_3_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_3_reloadable7/Release/0_3_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_3_reloadable7/Release/0_3_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_3_reloadable7/Release/0_3_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_3_reloadable7/Release/0_3_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_3_reloadable7/Release/0_3_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_3_reloadable7/Release/0_3_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_3_reloadable7/Release/0_3_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_3_reloadable7/Release/0_3_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_3_reloadable7/Release/0_3_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_3_reloadable11/Release/0_3_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_3_reloadable11/Release/0_3_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_3_reloadable11/Release/0_3_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_3_reloadable11/Release/0_3_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_3_reloadable11/Release/0_3_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_3_reloadable11/Release/0_3_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_3_reloadable11/Release/0_3_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_3_reloadable11/Release/0_3_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_3_reloadable11/Release/0_3_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 0_3_reloadable15/Release/0_3_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 0_3_reloadable15/Release/0_3_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 0_3_reloadable15/Release/0_3_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 0_3_reloadable15/Release/0_3_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 0_3_reloadable15/Release/0_3_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 0_3_reloadable15/Release/0_3_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 0_3_reloadable15/Release/0_3_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 0_3_reloadable15/Release/0_3_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 0_3_reloadable15/Release/0_3_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_0_reloadable5/Release/1_0_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_0_reloadable5/Release/1_0_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_0_reloadable5/Release/1_0_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_0_reloadable5/Release/1_0_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_0_reloadable5/Release/1_0_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_0_reloadable5/Release/1_0_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_0_reloadable5/Release/1_0_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_0_reloadable5/Release/1_0_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_0_reloadable5/Release/1_0_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_0_reloadable7/Release/1_0_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_0_reloadable7/Release/1_0_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_0_reloadable7/Release/1_0_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_0_reloadable7/Release/1_0_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_0_reloadable7/Release/1_0_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_0_reloadable7/Release/1_0_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_0_reloadable7/Release/1_0_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_0_reloadable7/Release/1_0_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_0_reloadable7/Release/1_0_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_0_reloadable11/Release/1_0_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_0_reloadable11/Release/1_0_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_0_reloadable11/Release/1_0_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_0_reloadable11/Release/1_0_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_0_reloadable11/Release/1_0_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_0_reloadable11/Release/1_0_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_0_reloadable11/Release/1_0_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_0_reloadable11/Release/1_0_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_0_reloadable11/Release/1_0_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_0_reloadable15/Release/1_0_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_0_reloadable15/Release/1_0_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_0_reloadable15/Release/1_0_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_0_reloadable15/Release/1_0_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_0_reloadable15/Release/1_0_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_0_reloadable15/Release/1_0_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_0_reloadable15/Release/1_0_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_0_reloadable15/Release/1_0_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_0_reloadable15/Release/1_0_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_1_reloadable5/Release/1_1_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_1_reloadable5/Release/1_1_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_1_reloadable5/Release/1_1_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_1_reloadable5/Release/1_1_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_1_reloadable5/Release/1_1_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_1_reloadable5/Release/1_1_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_1_reloadable5/Release/1_1_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_1_reloadable5/Release/1_1_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_1_reloadable5/Release/1_1_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_1_reloadable7/Release/1_1_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_1_reloadable7/Release/1_1_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_1_reloadable7/Release/1_1_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_1_reloadable7/Release/1_1_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_1_reloadable7/Release/1_1_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_1_reloadable7/Release/1_1_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_1_reloadable7/Release/1_1_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_1_reloadable7/Release/1_1_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_1_reloadable7/Release/1_1_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_1_reloadable11/Release/1_1_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_1_reloadable11/Release/1_1_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_1_reloadable11/Release/1_1_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_1_reloadable11/Release/1_1_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_1_reloadable11/Release/1_1_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_1_reloadable11/Release/1_1_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_1_reloadable11/Release/1_1_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_1_reloadable11/Release/1_1_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_1_reloadable11/Release/1_1_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_1_reloadable15/Release/1_1_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_1_reloadable15/Release/1_1_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_1_reloadable15/Release/1_1_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_1_reloadable15/Release/1_1_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_1_reloadable15/Release/1_1_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_1_reloadable15/Release/1_1_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_1_reloadable15/Release/1_1_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_1_reloadable15/Release/1_1_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_1_reloadable15/Release/1_1_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_2_reloadable5/Release/1_2_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_2_reloadable5/Release/1_2_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_2_reloadable5/Release/1_2_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_2_reloadable5/Release/1_2_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_2_reloadable5/Release/1_2_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_2_reloadable5/Release/1_2_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_2_reloadable5/Release/1_2_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_2_reloadable5/Release/1_2_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_2_reloadable5/Release/1_2_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_2_reloadable7/Release/1_2_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_2_reloadable7/Release/1_2_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_2_reloadable7/Release/1_2_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_2_reloadable7/Release/1_2_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_2_reloadable7/Release/1_2_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_2_reloadable7/Release/1_2_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_2_reloadable7/Release/1_2_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_2_reloadable7/Release/1_2_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_2_reloadable7/Release/1_2_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_2_reloadable11/Release/1_2_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_2_reloadable11/Release/1_2_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_2_reloadable11/Release/1_2_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_2_reloadable11/Release/1_2_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_2_reloadable11/Release/1_2_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_2_reloadable11/Release/1_2_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_2_reloadable11/Release/1_2_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_2_reloadable11/Release/1_2_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_2_reloadable11/Release/1_2_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_2_reloadable15/Release/1_2_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_2_reloadable15/Release/1_2_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_2_reloadable15/Release/1_2_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_2_reloadable15/Release/1_2_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_2_reloadable15/Release/1_2_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_2_reloadable15/Release/1_2_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_2_reloadable15/Release/1_2_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_2_reloadable15/Release/1_2_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_2_reloadable15/Release/1_2_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_3_reloadable5/Release/1_3_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_3_reloadable5/Release/1_3_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_3_reloadable5/Release/1_3_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_3_reloadable5/Release/1_3_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_3_reloadable5/Release/1_3_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_3_reloadable5/Release/1_3_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_3_reloadable5/Release/1_3_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_3_reloadable5/Release/1_3_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_3_reloadable5/Release/1_3_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_3_reloadable7/Release/1_3_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_3_reloadable7/Release/1_3_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_3_reloadable7/Release/1_3_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_3_reloadable7/Release/1_3_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_3_reloadable7/Release/1_3_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_3_reloadable7/Release/1_3_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_3_reloadable7/Release/1_3_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_3_reloadable7/Release/1_3_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_3_reloadable7/Release/1_3_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_3_reloadable11/Release/1_3_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_3_reloadable11/Release/1_3_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_3_reloadable11/Release/1_3_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_3_reloadable11/Release/1_3_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_3_reloadable11/Release/1_3_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_3_reloadable11/Release/1_3_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_3_reloadable11/Release/1_3_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_3_reloadable11/Release/1_3_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_3_reloadable11/Release/1_3_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 1_3_reloadable15/Release/1_3_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 1_3_reloadable15/Release/1_3_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 1_3_reloadable15/Release/1_3_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 1_3_reloadable15/Release/1_3_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 1_3_reloadable15/Release/1_3_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 1_3_reloadable15/Release/1_3_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 1_3_reloadable15/Release/1_3_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 1_3_reloadable15/Release/1_3_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 1_3_reloadable15/Release/1_3_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_0_reloadable5/Release/2_0_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_0_reloadable5/Release/2_0_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_0_reloadable5/Release/2_0_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_0_reloadable5/Release/2_0_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_0_reloadable5/Release/2_0_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_0_reloadable5/Release/2_0_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_0_reloadable5/Release/2_0_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_0_reloadable5/Release/2_0_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_0_reloadable5/Release/2_0_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_0_reloadable7/Release/2_0_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_0_reloadable7/Release/2_0_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_0_reloadable7/Release/2_0_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_0_reloadable7/Release/2_0_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_0_reloadable7/Release/2_0_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_0_reloadable7/Release/2_0_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_0_reloadable7/Release/2_0_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_0_reloadable7/Release/2_0_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_0_reloadable7/Release/2_0_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_0_reloadable11/Release/2_0_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_0_reloadable11/Release/2_0_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_0_reloadable11/Release/2_0_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_0_reloadable11/Release/2_0_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_0_reloadable11/Release/2_0_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_0_reloadable11/Release/2_0_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_0_reloadable11/Release/2_0_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_0_reloadable11/Release/2_0_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_0_reloadable11/Release/2_0_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_0_reloadable15/Release/2_0_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_0_reloadable15/Release/2_0_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_0_reloadable15/Release/2_0_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_0_reloadable15/Release/2_0_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_0_reloadable15/Release/2_0_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_0_reloadable15/Release/2_0_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_0_reloadable15/Release/2_0_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_0_reloadable15/Release/2_0_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_0_reloadable15/Release/2_0_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_1_reloadable5/Release/2_1_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_1_reloadable5/Release/2_1_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_1_reloadable5/Release/2_1_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_1_reloadable5/Release/2_1_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_1_reloadable5/Release/2_1_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_1_reloadable5/Release/2_1_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_1_reloadable5/Release/2_1_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_1_reloadable5/Release/2_1_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_1_reloadable5/Release/2_1_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_1_reloadable7/Release/2_1_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_1_reloadable7/Release/2_1_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_1_reloadable7/Release/2_1_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_1_reloadable7/Release/2_1_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_1_reloadable7/Release/2_1_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_1_reloadable7/Release/2_1_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_1_reloadable7/Release/2_1_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_1_reloadable7/Release/2_1_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_1_reloadable7/Release/2_1_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_1_reloadable11/Release/2_1_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_1_reloadable11/Release/2_1_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_1_reloadable11/Release/2_1_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_1_reloadable11/Release/2_1_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_1_reloadable11/Release/2_1_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_1_reloadable11/Release/2_1_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_1_reloadable11/Release/2_1_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_1_reloadable11/Release/2_1_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_1_reloadable11/Release/2_1_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_1_reloadable15/Release/2_1_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_1_reloadable15/Release/2_1_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_1_reloadable15/Release/2_1_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_1_reloadable15/Release/2_1_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_1_reloadable15/Release/2_1_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_1_reloadable15/Release/2_1_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_1_reloadable15/Release/2_1_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_1_reloadable15/Release/2_1_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_1_reloadable15/Release/2_1_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_2_reloadable5/Release/2_2_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_2_reloadable5/Release/2_2_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_2_reloadable5/Release/2_2_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_2_reloadable5/Release/2_2_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_2_reloadable5/Release/2_2_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_2_reloadable5/Release/2_2_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_2_reloadable5/Release/2_2_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_2_reloadable5/Release/2_2_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_2_reloadable5/Release/2_2_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_2_reloadable7/Release/2_2_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_2_reloadable7/Release/2_2_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_2_reloadable7/Release/2_2_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_2_reloadable7/Release/2_2_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_2_reloadable7/Release/2_2_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_2_reloadable7/Release/2_2_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_2_reloadable7/Release/2_2_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_2_reloadable7/Release/2_2_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_2_reloadable7/Release/2_2_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_2_reloadable11/Release/2_2_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_2_reloadable11/Release/2_2_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_2_reloadable11/Release/2_2_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_2_reloadable11/Release/2_2_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_2_reloadable11/Release/2_2_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_2_reloadable11/Release/2_2_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_2_reloadable11/Release/2_2_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_2_reloadable11/Release/2_2_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_2_reloadable11/Release/2_2_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_2_reloadable15/Release/2_2_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_2_reloadable15/Release/2_2_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_2_reloadable15/Release/2_2_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_2_reloadable15/Release/2_2_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_2_reloadable15/Release/2_2_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_2_reloadable15/Release/2_2_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_2_reloadable15/Release/2_2_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_2_reloadable15/Release/2_2_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_2_reloadable15/Release/2_2_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_3_reloadable5/Release/2_3_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_3_reloadable5/Release/2_3_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_3_reloadable5/Release/2_3_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_3_reloadable5/Release/2_3_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_3_reloadable5/Release/2_3_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_3_reloadable5/Release/2_3_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_3_reloadable5/Release/2_3_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_3_reloadable5/Release/2_3_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_3_reloadable5/Release/2_3_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_3_reloadable7/Release/2_3_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_3_reloadable7/Release/2_3_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_3_reloadable7/Release/2_3_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_3_reloadable7/Release/2_3_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_3_reloadable7/Release/2_3_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_3_reloadable7/Release/2_3_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_3_reloadable7/Release/2_3_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_3_reloadable7/Release/2_3_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_3_reloadable7/Release/2_3_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_3_reloadable11/Release/2_3_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_3_reloadable11/Release/2_3_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_3_reloadable11/Release/2_3_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_3_reloadable11/Release/2_3_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_3_reloadable11/Release/2_3_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_3_reloadable11/Release/2_3_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_3_reloadable11/Release/2_3_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_3_reloadable11/Release/2_3_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_3_reloadable11/Release/2_3_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 2_3_reloadable15/Release/2_3_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 2_3_reloadable15/Release/2_3_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 2_3_reloadable15/Release/2_3_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 2_3_reloadable15/Release/2_3_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 2_3_reloadable15/Release/2_3_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 2_3_reloadable15/Release/2_3_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 2_3_reloadable15/Release/2_3_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 2_3_reloadable15/Release/2_3_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 2_3_reloadable15/Release/2_3_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_0_reloadable5/Release/3_0_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_0_reloadable5/Release/3_0_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_0_reloadable5/Release/3_0_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_0_reloadable5/Release/3_0_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_0_reloadable5/Release/3_0_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_0_reloadable5/Release/3_0_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_0_reloadable5/Release/3_0_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_0_reloadable5/Release/3_0_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_0_reloadable5/Release/3_0_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_0_reloadable7/Release/3_0_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_0_reloadable7/Release/3_0_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_0_reloadable7/Release/3_0_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_0_reloadable7/Release/3_0_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_0_reloadable7/Release/3_0_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_0_reloadable7/Release/3_0_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_0_reloadable7/Release/3_0_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_0_reloadable7/Release/3_0_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_0_reloadable7/Release/3_0_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_0_reloadable11/Release/3_0_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_0_reloadable11/Release/3_0_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_0_reloadable11/Release/3_0_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_0_reloadable11/Release/3_0_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_0_reloadable11/Release/3_0_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_0_reloadable11/Release/3_0_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_0_reloadable11/Release/3_0_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_0_reloadable11/Release/3_0_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_0_reloadable11/Release/3_0_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_0_reloadable15/Release/3_0_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_0_reloadable15/Release/3_0_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_0_reloadable15/Release/3_0_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_0_reloadable15/Release/3_0_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_0_reloadable15/Release/3_0_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_0_reloadable15/Release/3_0_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_0_reloadable15/Release/3_0_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_0_reloadable15/Release/3_0_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_0_reloadable15/Release/3_0_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_1_reloadable5/Release/3_1_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_1_reloadable5/Release/3_1_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_1_reloadable5/Release/3_1_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_1_reloadable5/Release/3_1_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_1_reloadable5/Release/3_1_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_1_reloadable5/Release/3_1_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_1_reloadable5/Release/3_1_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_1_reloadable5/Release/3_1_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_1_reloadable5/Release/3_1_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_1_reloadable7/Release/3_1_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_1_reloadable7/Release/3_1_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_1_reloadable7/Release/3_1_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_1_reloadable7/Release/3_1_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_1_reloadable7/Release/3_1_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_1_reloadable7/Release/3_1_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_1_reloadable7/Release/3_1_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_1_reloadable7/Release/3_1_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_1_reloadable7/Release/3_1_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_1_reloadable11/Release/3_1_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_1_reloadable11/Release/3_1_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_1_reloadable11/Release/3_1_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_1_reloadable11/Release/3_1_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_1_reloadable11/Release/3_1_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_1_reloadable11/Release/3_1_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_1_reloadable11/Release/3_1_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_1_reloadable11/Release/3_1_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_1_reloadable11/Release/3_1_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_1_reloadable15/Release/3_1_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_1_reloadable15/Release/3_1_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_1_reloadable15/Release/3_1_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_1_reloadable15/Release/3_1_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_1_reloadable15/Release/3_1_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_1_reloadable15/Release/3_1_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_1_reloadable15/Release/3_1_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_1_reloadable15/Release/3_1_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_1_reloadable15/Release/3_1_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_2_reloadable5/Release/3_2_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_2_reloadable5/Release/3_2_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_2_reloadable5/Release/3_2_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_2_reloadable5/Release/3_2_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_2_reloadable5/Release/3_2_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_2_reloadable5/Release/3_2_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_2_reloadable5/Release/3_2_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_2_reloadable5/Release/3_2_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_2_reloadable5/Release/3_2_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_2_reloadable7/Release/3_2_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_2_reloadable7/Release/3_2_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_2_reloadable7/Release/3_2_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_2_reloadable7/Release/3_2_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_2_reloadable7/Release/3_2_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_2_reloadable7/Release/3_2_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_2_reloadable7/Release/3_2_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_2_reloadable7/Release/3_2_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_2_reloadable7/Release/3_2_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_2_reloadable11/Release/3_2_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_2_reloadable11/Release/3_2_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_2_reloadable11/Release/3_2_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_2_reloadable11/Release/3_2_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_2_reloadable11/Release/3_2_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_2_reloadable11/Release/3_2_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_2_reloadable11/Release/3_2_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_2_reloadable11/Release/3_2_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_2_reloadable11/Release/3_2_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_2_reloadable15/Release/3_2_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_2_reloadable15/Release/3_2_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_2_reloadable15/Release/3_2_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_2_reloadable15/Release/3_2_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_2_reloadable15/Release/3_2_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_2_reloadable15/Release/3_2_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_2_reloadable15/Release/3_2_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_2_reloadable15/Release/3_2_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_2_reloadable15/Release/3_2_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable5/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_3_reloadable5/Release/3_3_reloadable5 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_3_reloadable5/Release/3_3_reloadable5.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_3_reloadable5/Release/3_3_reloadable5.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_3_reloadable5/Release/3_3_reloadable5.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_3_reloadable5/Release/3_3_reloadable5.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_3_reloadable5/Release/3_3_reloadable5.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_3_reloadable5/Release/3_3_reloadable5.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_3_reloadable5/Release/3_3_reloadable5.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_3_reloadable5/Release/3_3_reloadable5.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable7/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_3_reloadable7/Release/3_3_reloadable7 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_3_reloadable7/Release/3_3_reloadable7.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_3_reloadable7/Release/3_3_reloadable7.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_3_reloadable7/Release/3_3_reloadable7.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_3_reloadable7/Release/3_3_reloadable7.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_3_reloadable7/Release/3_3_reloadable7.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_3_reloadable7/Release/3_3_reloadable7.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_3_reloadable7/Release/3_3_reloadable7.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_3_reloadable7/Release/3_3_reloadable7.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable11/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_3_reloadable11/Release/3_3_reloadable11 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_3_reloadable11/Release/3_3_reloadable11.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_3_reloadable11/Release/3_3_reloadable11.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_3_reloadable11/Release/3_3_reloadable11.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_3_reloadable11/Release/3_3_reloadable11.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_3_reloadable11/Release/3_3_reloadable11.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_3_reloadable11/Release/3_3_reloadable11.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_3_reloadable11/Release/3_3_reloadable11.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_3_reloadable11/Release/3_3_reloadable11.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable15/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5 3_3_reloadable15/Release/3_3_reloadable15 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.map 3_3_reloadable15/Release/3_3_reloadable15.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.lst 3_3_reloadable15/Release/3_3_reloadable15.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.calltree 3_3_reloadable15/Release/3_3_reloadable15.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.sdr 3_3_reloadable15/Release/3_3_reloadable15.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.srv 3_3_reloadable15/Release/3_3_reloadable15.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.txt 3_3_reloadable15/Release/3_3_reloadable15.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmic2 3_3_reloadable15/Release/3_3_reloadable15.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable5/Release/0_0_reloadable5.cmico 3_3_reloadable15/Release/3_3_reloadable15.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable17.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable17/scripts/0_0_reloadable17.prx) 2>&1 |& tee -a 0_0_reloadable17/0_0_reloadable17.log 0_0_reloadable17/timestamped_log/0_0_reloadable17.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable17.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable17.ll -o../Release/chesswork4837/0_0_reloadable17.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork4837 ../Release/chesswork4837/0_0_reloadable17.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable17-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +chess-backend 0_0_reloadable17-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +chess-backend 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +chess-backend 0_0_reloadable17-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable17-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--mist1 -k64 --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable17-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable17-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable17-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z13_b901_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable17-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z13_b881_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable17-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable17-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable17-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable17-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable17-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable17 -L --common 0_0_reloadable17-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable17.objlist -o../0_0_reloadable17.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable17.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable17" +bridge -o../Release/0_0_reloadable17 ../Release/0_0_reloadable17.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable17.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork4837 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable17 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (180 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable17/Release/0_0_reloadable17 >> 0_0_reloadable17/Release/0_0_reloadable17.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 0_1_reloadable17/Release/0_1_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 0_1_reloadable17/Release/0_1_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 0_1_reloadable17/Release/0_1_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 0_1_reloadable17/Release/0_1_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 0_1_reloadable17/Release/0_1_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 0_1_reloadable17/Release/0_1_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 0_1_reloadable17/Release/0_1_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 0_1_reloadable17/Release/0_1_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 0_1_reloadable17/Release/0_1_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 0_2_reloadable17/Release/0_2_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 0_2_reloadable17/Release/0_2_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 0_2_reloadable17/Release/0_2_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 0_2_reloadable17/Release/0_2_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 0_2_reloadable17/Release/0_2_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 0_2_reloadable17/Release/0_2_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 0_2_reloadable17/Release/0_2_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 0_2_reloadable17/Release/0_2_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 0_2_reloadable17/Release/0_2_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 0_3_reloadable17/Release/0_3_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 0_3_reloadable17/Release/0_3_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 0_3_reloadable17/Release/0_3_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 0_3_reloadable17/Release/0_3_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 0_3_reloadable17/Release/0_3_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 0_3_reloadable17/Release/0_3_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 0_3_reloadable17/Release/0_3_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 0_3_reloadable17/Release/0_3_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 0_3_reloadable17/Release/0_3_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 1_0_reloadable17/Release/1_0_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 1_0_reloadable17/Release/1_0_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 1_0_reloadable17/Release/1_0_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 1_0_reloadable17/Release/1_0_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 1_0_reloadable17/Release/1_0_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 1_0_reloadable17/Release/1_0_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 1_0_reloadable17/Release/1_0_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 1_0_reloadable17/Release/1_0_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 1_0_reloadable17/Release/1_0_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 1_1_reloadable17/Release/1_1_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 1_1_reloadable17/Release/1_1_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 1_1_reloadable17/Release/1_1_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 1_1_reloadable17/Release/1_1_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 1_1_reloadable17/Release/1_1_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 1_1_reloadable17/Release/1_1_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 1_1_reloadable17/Release/1_1_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 1_1_reloadable17/Release/1_1_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 1_1_reloadable17/Release/1_1_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 1_2_reloadable17/Release/1_2_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 1_2_reloadable17/Release/1_2_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 1_2_reloadable17/Release/1_2_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 1_2_reloadable17/Release/1_2_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 1_2_reloadable17/Release/1_2_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 1_2_reloadable17/Release/1_2_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 1_2_reloadable17/Release/1_2_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 1_2_reloadable17/Release/1_2_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 1_2_reloadable17/Release/1_2_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 1_3_reloadable17/Release/1_3_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 1_3_reloadable17/Release/1_3_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 1_3_reloadable17/Release/1_3_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 1_3_reloadable17/Release/1_3_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 1_3_reloadable17/Release/1_3_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 1_3_reloadable17/Release/1_3_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 1_3_reloadable17/Release/1_3_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 1_3_reloadable17/Release/1_3_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 1_3_reloadable17/Release/1_3_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 2_0_reloadable17/Release/2_0_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 2_0_reloadable17/Release/2_0_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 2_0_reloadable17/Release/2_0_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 2_0_reloadable17/Release/2_0_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 2_0_reloadable17/Release/2_0_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 2_0_reloadable17/Release/2_0_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 2_0_reloadable17/Release/2_0_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 2_0_reloadable17/Release/2_0_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 2_0_reloadable17/Release/2_0_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 2_1_reloadable17/Release/2_1_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 2_1_reloadable17/Release/2_1_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 2_1_reloadable17/Release/2_1_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 2_1_reloadable17/Release/2_1_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 2_1_reloadable17/Release/2_1_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 2_1_reloadable17/Release/2_1_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 2_1_reloadable17/Release/2_1_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 2_1_reloadable17/Release/2_1_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 2_1_reloadable17/Release/2_1_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 2_2_reloadable17/Release/2_2_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 2_2_reloadable17/Release/2_2_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 2_2_reloadable17/Release/2_2_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 2_2_reloadable17/Release/2_2_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 2_2_reloadable17/Release/2_2_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 2_2_reloadable17/Release/2_2_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 2_2_reloadable17/Release/2_2_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 2_2_reloadable17/Release/2_2_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 2_2_reloadable17/Release/2_2_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 2_3_reloadable17/Release/2_3_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 2_3_reloadable17/Release/2_3_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 2_3_reloadable17/Release/2_3_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 2_3_reloadable17/Release/2_3_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 2_3_reloadable17/Release/2_3_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 2_3_reloadable17/Release/2_3_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 2_3_reloadable17/Release/2_3_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 2_3_reloadable17/Release/2_3_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 2_3_reloadable17/Release/2_3_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 3_0_reloadable17/Release/3_0_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 3_0_reloadable17/Release/3_0_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 3_0_reloadable17/Release/3_0_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 3_0_reloadable17/Release/3_0_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 3_0_reloadable17/Release/3_0_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 3_0_reloadable17/Release/3_0_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 3_0_reloadable17/Release/3_0_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 3_0_reloadable17/Release/3_0_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 3_0_reloadable17/Release/3_0_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 3_1_reloadable17/Release/3_1_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 3_1_reloadable17/Release/3_1_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 3_1_reloadable17/Release/3_1_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 3_1_reloadable17/Release/3_1_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 3_1_reloadable17/Release/3_1_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 3_1_reloadable17/Release/3_1_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 3_1_reloadable17/Release/3_1_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 3_1_reloadable17/Release/3_1_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 3_1_reloadable17/Release/3_1_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 3_2_reloadable17/Release/3_2_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 3_2_reloadable17/Release/3_2_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 3_2_reloadable17/Release/3_2_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 3_2_reloadable17/Release/3_2_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 3_2_reloadable17/Release/3_2_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 3_2_reloadable17/Release/3_2_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 3_2_reloadable17/Release/3_2_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 3_2_reloadable17/Release/3_2_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 3_2_reloadable17/Release/3_2_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable17/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17 3_3_reloadable17/Release/3_3_reloadable17 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.map 3_3_reloadable17/Release/3_3_reloadable17.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.lst 3_3_reloadable17/Release/3_3_reloadable17.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.calltree 3_3_reloadable17/Release/3_3_reloadable17.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.sdr 3_3_reloadable17/Release/3_3_reloadable17.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.srv 3_3_reloadable17/Release/3_3_reloadable17.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.txt 3_3_reloadable17/Release/3_3_reloadable17.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmic2 3_3_reloadable17/Release/3_3_reloadable17.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable17/Release/0_0_reloadable17.cmico 3_3_reloadable17/Release/3_3_reloadable17.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable19.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable19/scripts/0_0_reloadable19.prx) 2>&1 |& tee -a 0_0_reloadable19/0_0_reloadable19.log 0_0_reloadable19/timestamped_log/0_0_reloadable19.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable19.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable19.ll -o../Release/chesswork5183/0_0_reloadable19.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork5183 ../Release/chesswork5183/0_0_reloadable19.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable19-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +chess-backend 0_0_reloadable19-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +chess-backend 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +chess-backend 0_0_reloadable19-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--mist1 -k64 --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable19-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable19-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z13_b924_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable19-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable19-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable19-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable19-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable19-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable19 -L --common 0_0_reloadable19-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable19.objlist -o../0_0_reloadable19.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable19.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable19" +bridge -o../Release/0_0_reloadable19 ../Release/0_0_reloadable19.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable19.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork5183 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable19 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (183 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable19/Release/0_0_reloadable19 >> 0_0_reloadable19/Release/0_0_reloadable19.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 0_1_reloadable19/Release/0_1_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 0_1_reloadable19/Release/0_1_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 0_1_reloadable19/Release/0_1_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 0_1_reloadable19/Release/0_1_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 0_1_reloadable19/Release/0_1_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 0_1_reloadable19/Release/0_1_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 0_1_reloadable19/Release/0_1_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 0_1_reloadable19/Release/0_1_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 0_1_reloadable19/Release/0_1_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 0_2_reloadable19/Release/0_2_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 0_2_reloadable19/Release/0_2_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 0_2_reloadable19/Release/0_2_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 0_2_reloadable19/Release/0_2_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 0_2_reloadable19/Release/0_2_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 0_2_reloadable19/Release/0_2_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 0_2_reloadable19/Release/0_2_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 0_2_reloadable19/Release/0_2_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 0_2_reloadable19/Release/0_2_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 0_3_reloadable19/Release/0_3_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 0_3_reloadable19/Release/0_3_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 0_3_reloadable19/Release/0_3_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 0_3_reloadable19/Release/0_3_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 0_3_reloadable19/Release/0_3_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 0_3_reloadable19/Release/0_3_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 0_3_reloadable19/Release/0_3_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 0_3_reloadable19/Release/0_3_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 0_3_reloadable19/Release/0_3_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 1_0_reloadable19/Release/1_0_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 1_0_reloadable19/Release/1_0_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 1_0_reloadable19/Release/1_0_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 1_0_reloadable19/Release/1_0_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 1_0_reloadable19/Release/1_0_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 1_0_reloadable19/Release/1_0_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 1_0_reloadable19/Release/1_0_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 1_0_reloadable19/Release/1_0_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 1_0_reloadable19/Release/1_0_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 1_1_reloadable19/Release/1_1_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 1_1_reloadable19/Release/1_1_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 1_1_reloadable19/Release/1_1_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 1_1_reloadable19/Release/1_1_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 1_1_reloadable19/Release/1_1_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 1_1_reloadable19/Release/1_1_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 1_1_reloadable19/Release/1_1_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 1_1_reloadable19/Release/1_1_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 1_1_reloadable19/Release/1_1_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 1_2_reloadable19/Release/1_2_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 1_2_reloadable19/Release/1_2_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 1_2_reloadable19/Release/1_2_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 1_2_reloadable19/Release/1_2_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 1_2_reloadable19/Release/1_2_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 1_2_reloadable19/Release/1_2_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 1_2_reloadable19/Release/1_2_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 1_2_reloadable19/Release/1_2_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 1_2_reloadable19/Release/1_2_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 1_3_reloadable19/Release/1_3_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 1_3_reloadable19/Release/1_3_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 1_3_reloadable19/Release/1_3_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 1_3_reloadable19/Release/1_3_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 1_3_reloadable19/Release/1_3_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 1_3_reloadable19/Release/1_3_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 1_3_reloadable19/Release/1_3_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 1_3_reloadable19/Release/1_3_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 1_3_reloadable19/Release/1_3_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 2_0_reloadable19/Release/2_0_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 2_0_reloadable19/Release/2_0_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 2_0_reloadable19/Release/2_0_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 2_0_reloadable19/Release/2_0_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 2_0_reloadable19/Release/2_0_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 2_0_reloadable19/Release/2_0_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 2_0_reloadable19/Release/2_0_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 2_0_reloadable19/Release/2_0_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 2_0_reloadable19/Release/2_0_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 2_1_reloadable19/Release/2_1_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 2_1_reloadable19/Release/2_1_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 2_1_reloadable19/Release/2_1_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 2_1_reloadable19/Release/2_1_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 2_1_reloadable19/Release/2_1_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 2_1_reloadable19/Release/2_1_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 2_1_reloadable19/Release/2_1_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 2_1_reloadable19/Release/2_1_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 2_1_reloadable19/Release/2_1_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 2_2_reloadable19/Release/2_2_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 2_2_reloadable19/Release/2_2_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 2_2_reloadable19/Release/2_2_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 2_2_reloadable19/Release/2_2_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 2_2_reloadable19/Release/2_2_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 2_2_reloadable19/Release/2_2_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 2_2_reloadable19/Release/2_2_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 2_2_reloadable19/Release/2_2_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 2_2_reloadable19/Release/2_2_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 2_3_reloadable19/Release/2_3_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 2_3_reloadable19/Release/2_3_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 2_3_reloadable19/Release/2_3_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 2_3_reloadable19/Release/2_3_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 2_3_reloadable19/Release/2_3_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 2_3_reloadable19/Release/2_3_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 2_3_reloadable19/Release/2_3_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 2_3_reloadable19/Release/2_3_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 2_3_reloadable19/Release/2_3_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 3_0_reloadable19/Release/3_0_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 3_0_reloadable19/Release/3_0_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 3_0_reloadable19/Release/3_0_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 3_0_reloadable19/Release/3_0_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 3_0_reloadable19/Release/3_0_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 3_0_reloadable19/Release/3_0_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 3_0_reloadable19/Release/3_0_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 3_0_reloadable19/Release/3_0_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 3_0_reloadable19/Release/3_0_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 3_1_reloadable19/Release/3_1_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 3_1_reloadable19/Release/3_1_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 3_1_reloadable19/Release/3_1_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 3_1_reloadable19/Release/3_1_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 3_1_reloadable19/Release/3_1_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 3_1_reloadable19/Release/3_1_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 3_1_reloadable19/Release/3_1_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 3_1_reloadable19/Release/3_1_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 3_1_reloadable19/Release/3_1_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 3_2_reloadable19/Release/3_2_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 3_2_reloadable19/Release/3_2_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 3_2_reloadable19/Release/3_2_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 3_2_reloadable19/Release/3_2_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 3_2_reloadable19/Release/3_2_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 3_2_reloadable19/Release/3_2_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 3_2_reloadable19/Release/3_2_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 3_2_reloadable19/Release/3_2_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 3_2_reloadable19/Release/3_2_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable19/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19 3_3_reloadable19/Release/3_3_reloadable19 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.map 3_3_reloadable19/Release/3_3_reloadable19.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.lst 3_3_reloadable19/Release/3_3_reloadable19.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.calltree 3_3_reloadable19/Release/3_3_reloadable19.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.sdr 3_3_reloadable19/Release/3_3_reloadable19.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.srv 3_3_reloadable19/Release/3_3_reloadable19.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.txt 3_3_reloadable19/Release/3_3_reloadable19.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmic2 3_3_reloadable19/Release/3_3_reloadable19.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable19/Release/0_0_reloadable19.cmico 3_3_reloadable19/Release/3_3_reloadable19.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable20.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable20/scripts/0_0_reloadable20.prx) 2>&1 |& tee -a 0_0_reloadable20/0_0_reloadable20.log 0_0_reloadable20/timestamped_log/0_0_reloadable20.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable20.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable20.ll -o../Release/chesswork5532/0_0_reloadable20.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork5532 ../Release/chesswork5532/0_0_reloadable20.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable20-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +chess-backend 0_0_reloadable20-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +chess-backend 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +chess-backend 0_0_reloadable20-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable20-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable20-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable20-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z14_b1685_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z14_b1685_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z14_b1685_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z14_b1685_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z14_b1685_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z14_b1685_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z14_b1685_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable20-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable20-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable20-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable20-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--mist1 -k64 --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable20 -L --common 0_0_reloadable20-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable20.objlist -o../0_0_reloadable20.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable20.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable20" +bridge -o../Release/0_0_reloadable20 ../Release/0_0_reloadable20.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable20.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork5532 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable20 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (179 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable20/Release/0_0_reloadable20 >> 0_0_reloadable20/Release/0_0_reloadable20.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 0_1_reloadable20/Release/0_1_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 0_1_reloadable20/Release/0_1_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 0_1_reloadable20/Release/0_1_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 0_1_reloadable20/Release/0_1_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 0_1_reloadable20/Release/0_1_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 0_1_reloadable20/Release/0_1_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 0_1_reloadable20/Release/0_1_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 0_1_reloadable20/Release/0_1_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 0_1_reloadable20/Release/0_1_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 0_2_reloadable20/Release/0_2_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 0_2_reloadable20/Release/0_2_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 0_2_reloadable20/Release/0_2_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 0_2_reloadable20/Release/0_2_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 0_2_reloadable20/Release/0_2_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 0_2_reloadable20/Release/0_2_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 0_2_reloadable20/Release/0_2_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 0_2_reloadable20/Release/0_2_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 0_2_reloadable20/Release/0_2_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 0_3_reloadable20/Release/0_3_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 0_3_reloadable20/Release/0_3_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 0_3_reloadable20/Release/0_3_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 0_3_reloadable20/Release/0_3_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 0_3_reloadable20/Release/0_3_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 0_3_reloadable20/Release/0_3_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 0_3_reloadable20/Release/0_3_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 0_3_reloadable20/Release/0_3_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 0_3_reloadable20/Release/0_3_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 1_0_reloadable20/Release/1_0_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 1_0_reloadable20/Release/1_0_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 1_0_reloadable20/Release/1_0_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 1_0_reloadable20/Release/1_0_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 1_0_reloadable20/Release/1_0_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 1_0_reloadable20/Release/1_0_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 1_0_reloadable20/Release/1_0_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 1_0_reloadable20/Release/1_0_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 1_0_reloadable20/Release/1_0_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 1_1_reloadable20/Release/1_1_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 1_1_reloadable20/Release/1_1_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 1_1_reloadable20/Release/1_1_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 1_1_reloadable20/Release/1_1_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 1_1_reloadable20/Release/1_1_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 1_1_reloadable20/Release/1_1_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 1_1_reloadable20/Release/1_1_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 1_1_reloadable20/Release/1_1_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 1_1_reloadable20/Release/1_1_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 1_2_reloadable20/Release/1_2_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 1_2_reloadable20/Release/1_2_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 1_2_reloadable20/Release/1_2_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 1_2_reloadable20/Release/1_2_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 1_2_reloadable20/Release/1_2_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 1_2_reloadable20/Release/1_2_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 1_2_reloadable20/Release/1_2_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 1_2_reloadable20/Release/1_2_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 1_2_reloadable20/Release/1_2_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 1_3_reloadable20/Release/1_3_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 1_3_reloadable20/Release/1_3_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 1_3_reloadable20/Release/1_3_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 1_3_reloadable20/Release/1_3_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 1_3_reloadable20/Release/1_3_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 1_3_reloadable20/Release/1_3_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 1_3_reloadable20/Release/1_3_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 1_3_reloadable20/Release/1_3_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 1_3_reloadable20/Release/1_3_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 2_0_reloadable20/Release/2_0_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 2_0_reloadable20/Release/2_0_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 2_0_reloadable20/Release/2_0_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 2_0_reloadable20/Release/2_0_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 2_0_reloadable20/Release/2_0_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 2_0_reloadable20/Release/2_0_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 2_0_reloadable20/Release/2_0_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 2_0_reloadable20/Release/2_0_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 2_0_reloadable20/Release/2_0_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 2_1_reloadable20/Release/2_1_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 2_1_reloadable20/Release/2_1_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 2_1_reloadable20/Release/2_1_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 2_1_reloadable20/Release/2_1_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 2_1_reloadable20/Release/2_1_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 2_1_reloadable20/Release/2_1_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 2_1_reloadable20/Release/2_1_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 2_1_reloadable20/Release/2_1_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 2_1_reloadable20/Release/2_1_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 2_2_reloadable20/Release/2_2_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 2_2_reloadable20/Release/2_2_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 2_2_reloadable20/Release/2_2_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 2_2_reloadable20/Release/2_2_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 2_2_reloadable20/Release/2_2_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 2_2_reloadable20/Release/2_2_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 2_2_reloadable20/Release/2_2_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 2_2_reloadable20/Release/2_2_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 2_2_reloadable20/Release/2_2_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 2_3_reloadable20/Release/2_3_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 2_3_reloadable20/Release/2_3_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 2_3_reloadable20/Release/2_3_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 2_3_reloadable20/Release/2_3_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 2_3_reloadable20/Release/2_3_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 2_3_reloadable20/Release/2_3_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 2_3_reloadable20/Release/2_3_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 2_3_reloadable20/Release/2_3_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 2_3_reloadable20/Release/2_3_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 3_0_reloadable20/Release/3_0_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 3_0_reloadable20/Release/3_0_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 3_0_reloadable20/Release/3_0_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 3_0_reloadable20/Release/3_0_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 3_0_reloadable20/Release/3_0_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 3_0_reloadable20/Release/3_0_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 3_0_reloadable20/Release/3_0_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 3_0_reloadable20/Release/3_0_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 3_0_reloadable20/Release/3_0_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 3_1_reloadable20/Release/3_1_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 3_1_reloadable20/Release/3_1_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 3_1_reloadable20/Release/3_1_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 3_1_reloadable20/Release/3_1_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 3_1_reloadable20/Release/3_1_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 3_1_reloadable20/Release/3_1_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 3_1_reloadable20/Release/3_1_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 3_1_reloadable20/Release/3_1_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 3_1_reloadable20/Release/3_1_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 3_2_reloadable20/Release/3_2_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 3_2_reloadable20/Release/3_2_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 3_2_reloadable20/Release/3_2_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 3_2_reloadable20/Release/3_2_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 3_2_reloadable20/Release/3_2_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 3_2_reloadable20/Release/3_2_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 3_2_reloadable20/Release/3_2_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 3_2_reloadable20/Release/3_2_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 3_2_reloadable20/Release/3_2_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable20/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20 3_3_reloadable20/Release/3_3_reloadable20 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.map 3_3_reloadable20/Release/3_3_reloadable20.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.lst 3_3_reloadable20/Release/3_3_reloadable20.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.calltree 3_3_reloadable20/Release/3_3_reloadable20.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.sdr 3_3_reloadable20/Release/3_3_reloadable20.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.srv 3_3_reloadable20/Release/3_3_reloadable20.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.txt 3_3_reloadable20/Release/3_3_reloadable20.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmic2 3_3_reloadable20/Release/3_3_reloadable20.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable20/Release/0_0_reloadable20.cmico 3_3_reloadable20/Release/3_3_reloadable20.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable21.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable21/scripts/0_0_reloadable21.prx) 2>&1 |& tee -a 0_0_reloadable21/0_0_reloadable21.log 0_0_reloadable21/timestamped_log/0_0_reloadable21.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable21.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable21.ll -o../Release/chesswork5877/0_0_reloadable21.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork5877 ../Release/chesswork5877/0_0_reloadable21.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable21-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +chess-backend 0_0_reloadable21-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +chess-backend 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +chess-backend 0_0_reloadable21-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +chess-backend 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable21-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable21-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable21-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable21-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable21-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable21-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable21-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable21 -L --common 0_0_reloadable21-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable21.objlist -o../0_0_reloadable21.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable21.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable21" +bridge -o../Release/0_0_reloadable21 ../Release/0_0_reloadable21.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable21.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork5877 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable21 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (180 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable21/Release/0_0_reloadable21 >> 0_0_reloadable21/Release/0_0_reloadable21.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 0_1_reloadable21/Release/0_1_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 0_1_reloadable21/Release/0_1_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 0_1_reloadable21/Release/0_1_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 0_1_reloadable21/Release/0_1_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 0_1_reloadable21/Release/0_1_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 0_1_reloadable21/Release/0_1_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 0_1_reloadable21/Release/0_1_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 0_1_reloadable21/Release/0_1_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 0_1_reloadable21/Release/0_1_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 0_2_reloadable21/Release/0_2_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 0_2_reloadable21/Release/0_2_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 0_2_reloadable21/Release/0_2_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 0_2_reloadable21/Release/0_2_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 0_2_reloadable21/Release/0_2_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 0_2_reloadable21/Release/0_2_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 0_2_reloadable21/Release/0_2_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 0_2_reloadable21/Release/0_2_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 0_2_reloadable21/Release/0_2_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 0_3_reloadable21/Release/0_3_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 0_3_reloadable21/Release/0_3_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 0_3_reloadable21/Release/0_3_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 0_3_reloadable21/Release/0_3_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 0_3_reloadable21/Release/0_3_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 0_3_reloadable21/Release/0_3_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 0_3_reloadable21/Release/0_3_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 0_3_reloadable21/Release/0_3_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 0_3_reloadable21/Release/0_3_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 1_0_reloadable21/Release/1_0_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 1_0_reloadable21/Release/1_0_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 1_0_reloadable21/Release/1_0_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 1_0_reloadable21/Release/1_0_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 1_0_reloadable21/Release/1_0_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 1_0_reloadable21/Release/1_0_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 1_0_reloadable21/Release/1_0_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 1_0_reloadable21/Release/1_0_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 1_0_reloadable21/Release/1_0_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 1_1_reloadable21/Release/1_1_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 1_1_reloadable21/Release/1_1_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 1_1_reloadable21/Release/1_1_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 1_1_reloadable21/Release/1_1_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 1_1_reloadable21/Release/1_1_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 1_1_reloadable21/Release/1_1_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 1_1_reloadable21/Release/1_1_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 1_1_reloadable21/Release/1_1_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 1_1_reloadable21/Release/1_1_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 1_2_reloadable21/Release/1_2_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 1_2_reloadable21/Release/1_2_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 1_2_reloadable21/Release/1_2_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 1_2_reloadable21/Release/1_2_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 1_2_reloadable21/Release/1_2_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 1_2_reloadable21/Release/1_2_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 1_2_reloadable21/Release/1_2_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 1_2_reloadable21/Release/1_2_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 1_2_reloadable21/Release/1_2_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 1_3_reloadable21/Release/1_3_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 1_3_reloadable21/Release/1_3_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 1_3_reloadable21/Release/1_3_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 1_3_reloadable21/Release/1_3_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 1_3_reloadable21/Release/1_3_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 1_3_reloadable21/Release/1_3_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 1_3_reloadable21/Release/1_3_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 1_3_reloadable21/Release/1_3_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 1_3_reloadable21/Release/1_3_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 2_0_reloadable21/Release/2_0_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 2_0_reloadable21/Release/2_0_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 2_0_reloadable21/Release/2_0_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 2_0_reloadable21/Release/2_0_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 2_0_reloadable21/Release/2_0_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 2_0_reloadable21/Release/2_0_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 2_0_reloadable21/Release/2_0_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 2_0_reloadable21/Release/2_0_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 2_0_reloadable21/Release/2_0_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 2_1_reloadable21/Release/2_1_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 2_1_reloadable21/Release/2_1_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 2_1_reloadable21/Release/2_1_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 2_1_reloadable21/Release/2_1_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 2_1_reloadable21/Release/2_1_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 2_1_reloadable21/Release/2_1_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 2_1_reloadable21/Release/2_1_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 2_1_reloadable21/Release/2_1_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 2_1_reloadable21/Release/2_1_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 2_2_reloadable21/Release/2_2_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 2_2_reloadable21/Release/2_2_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 2_2_reloadable21/Release/2_2_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 2_2_reloadable21/Release/2_2_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 2_2_reloadable21/Release/2_2_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 2_2_reloadable21/Release/2_2_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 2_2_reloadable21/Release/2_2_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 2_2_reloadable21/Release/2_2_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 2_2_reloadable21/Release/2_2_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 2_3_reloadable21/Release/2_3_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 2_3_reloadable21/Release/2_3_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 2_3_reloadable21/Release/2_3_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 2_3_reloadable21/Release/2_3_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 2_3_reloadable21/Release/2_3_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 2_3_reloadable21/Release/2_3_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 2_3_reloadable21/Release/2_3_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 2_3_reloadable21/Release/2_3_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 2_3_reloadable21/Release/2_3_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 3_0_reloadable21/Release/3_0_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 3_0_reloadable21/Release/3_0_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 3_0_reloadable21/Release/3_0_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 3_0_reloadable21/Release/3_0_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 3_0_reloadable21/Release/3_0_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 3_0_reloadable21/Release/3_0_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 3_0_reloadable21/Release/3_0_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 3_0_reloadable21/Release/3_0_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 3_0_reloadable21/Release/3_0_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 3_1_reloadable21/Release/3_1_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 3_1_reloadable21/Release/3_1_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 3_1_reloadable21/Release/3_1_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 3_1_reloadable21/Release/3_1_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 3_1_reloadable21/Release/3_1_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 3_1_reloadable21/Release/3_1_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 3_1_reloadable21/Release/3_1_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 3_1_reloadable21/Release/3_1_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 3_1_reloadable21/Release/3_1_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 3_2_reloadable21/Release/3_2_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 3_2_reloadable21/Release/3_2_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 3_2_reloadable21/Release/3_2_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 3_2_reloadable21/Release/3_2_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 3_2_reloadable21/Release/3_2_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 3_2_reloadable21/Release/3_2_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 3_2_reloadable21/Release/3_2_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 3_2_reloadable21/Release/3_2_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 3_2_reloadable21/Release/3_2_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable21/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21 3_3_reloadable21/Release/3_3_reloadable21 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.map 3_3_reloadable21/Release/3_3_reloadable21.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.lst 3_3_reloadable21/Release/3_3_reloadable21.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.calltree 3_3_reloadable21/Release/3_3_reloadable21.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.sdr 3_3_reloadable21/Release/3_3_reloadable21.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.srv 3_3_reloadable21/Release/3_3_reloadable21.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.txt 3_3_reloadable21/Release/3_3_reloadable21.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmic2 3_3_reloadable21/Release/3_3_reloadable21.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable21/Release/0_0_reloadable21.cmico 3_3_reloadable21/Release/3_3_reloadable21.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable22.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable22/scripts/0_0_reloadable22.prx) 2>&1 |& tee -a 0_0_reloadable22/0_0_reloadable22.log 0_0_reloadable22/timestamped_log/0_0_reloadable22.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable22.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable22.ll -o../Release/chesswork6223/0_0_reloadable22.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork6223 ../Release/chesswork6223/0_0_reloadable22.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable22-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +chess-backend 0_0_reloadable22-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +chess-backend 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +chess-backend 0_0_reloadable22-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable22-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable22-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z13_b787_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi0EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable22-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp", line 112, column 4: (loop #961) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. +--mist1 -k64 --common 0_0_reloadable22-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable22-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable22-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable22-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--mist1 -k64 --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable22 -L --common 0_0_reloadable22-F_Z15resize_bilinearI8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_aS0_EEvPS1_S2_R29resize_common_internal_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable22.objlist -o../0_0_reloadable22.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable22.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable22" +bridge -o../Release/0_0_reloadable22 ../Release/0_0_reloadable22.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable22.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork6223 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable22 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (194 errors, 5 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable22/Release/0_0_reloadable22 >> 0_0_reloadable22/Release/0_0_reloadable22.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 0_1_reloadable22/Release/0_1_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 0_1_reloadable22/Release/0_1_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 0_1_reloadable22/Release/0_1_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 0_1_reloadable22/Release/0_1_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 0_1_reloadable22/Release/0_1_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 0_1_reloadable22/Release/0_1_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 0_1_reloadable22/Release/0_1_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 0_1_reloadable22/Release/0_1_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 0_1_reloadable22/Release/0_1_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 0_2_reloadable22/Release/0_2_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 0_2_reloadable22/Release/0_2_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 0_2_reloadable22/Release/0_2_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 0_2_reloadable22/Release/0_2_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 0_2_reloadable22/Release/0_2_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 0_2_reloadable22/Release/0_2_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 0_2_reloadable22/Release/0_2_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 0_2_reloadable22/Release/0_2_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 0_2_reloadable22/Release/0_2_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 0_3_reloadable22/Release/0_3_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 0_3_reloadable22/Release/0_3_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 0_3_reloadable22/Release/0_3_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 0_3_reloadable22/Release/0_3_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 0_3_reloadable22/Release/0_3_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 0_3_reloadable22/Release/0_3_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 0_3_reloadable22/Release/0_3_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 0_3_reloadable22/Release/0_3_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 0_3_reloadable22/Release/0_3_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 1_0_reloadable22/Release/1_0_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 1_0_reloadable22/Release/1_0_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 1_0_reloadable22/Release/1_0_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 1_0_reloadable22/Release/1_0_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 1_0_reloadable22/Release/1_0_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 1_0_reloadable22/Release/1_0_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 1_0_reloadable22/Release/1_0_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 1_0_reloadable22/Release/1_0_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 1_0_reloadable22/Release/1_0_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 1_1_reloadable22/Release/1_1_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 1_1_reloadable22/Release/1_1_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 1_1_reloadable22/Release/1_1_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 1_1_reloadable22/Release/1_1_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 1_1_reloadable22/Release/1_1_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 1_1_reloadable22/Release/1_1_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 1_1_reloadable22/Release/1_1_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 1_1_reloadable22/Release/1_1_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 1_1_reloadable22/Release/1_1_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 1_2_reloadable22/Release/1_2_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 1_2_reloadable22/Release/1_2_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 1_2_reloadable22/Release/1_2_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 1_2_reloadable22/Release/1_2_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 1_2_reloadable22/Release/1_2_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 1_2_reloadable22/Release/1_2_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 1_2_reloadable22/Release/1_2_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 1_2_reloadable22/Release/1_2_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 1_2_reloadable22/Release/1_2_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 1_3_reloadable22/Release/1_3_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 1_3_reloadable22/Release/1_3_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 1_3_reloadable22/Release/1_3_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 1_3_reloadable22/Release/1_3_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 1_3_reloadable22/Release/1_3_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 1_3_reloadable22/Release/1_3_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 1_3_reloadable22/Release/1_3_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 1_3_reloadable22/Release/1_3_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 1_3_reloadable22/Release/1_3_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 2_0_reloadable22/Release/2_0_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 2_0_reloadable22/Release/2_0_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 2_0_reloadable22/Release/2_0_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 2_0_reloadable22/Release/2_0_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 2_0_reloadable22/Release/2_0_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 2_0_reloadable22/Release/2_0_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 2_0_reloadable22/Release/2_0_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 2_0_reloadable22/Release/2_0_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 2_0_reloadable22/Release/2_0_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 2_1_reloadable22/Release/2_1_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 2_1_reloadable22/Release/2_1_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 2_1_reloadable22/Release/2_1_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 2_1_reloadable22/Release/2_1_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 2_1_reloadable22/Release/2_1_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 2_1_reloadable22/Release/2_1_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 2_1_reloadable22/Release/2_1_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 2_1_reloadable22/Release/2_1_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 2_1_reloadable22/Release/2_1_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 2_2_reloadable22/Release/2_2_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 2_2_reloadable22/Release/2_2_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 2_2_reloadable22/Release/2_2_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 2_2_reloadable22/Release/2_2_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 2_2_reloadable22/Release/2_2_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 2_2_reloadable22/Release/2_2_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 2_2_reloadable22/Release/2_2_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 2_2_reloadable22/Release/2_2_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 2_2_reloadable22/Release/2_2_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 2_3_reloadable22/Release/2_3_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 2_3_reloadable22/Release/2_3_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 2_3_reloadable22/Release/2_3_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 2_3_reloadable22/Release/2_3_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 2_3_reloadable22/Release/2_3_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 2_3_reloadable22/Release/2_3_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 2_3_reloadable22/Release/2_3_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 2_3_reloadable22/Release/2_3_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 2_3_reloadable22/Release/2_3_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 3_0_reloadable22/Release/3_0_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 3_0_reloadable22/Release/3_0_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 3_0_reloadable22/Release/3_0_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 3_0_reloadable22/Release/3_0_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 3_0_reloadable22/Release/3_0_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 3_0_reloadable22/Release/3_0_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 3_0_reloadable22/Release/3_0_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 3_0_reloadable22/Release/3_0_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 3_0_reloadable22/Release/3_0_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 3_1_reloadable22/Release/3_1_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 3_1_reloadable22/Release/3_1_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 3_1_reloadable22/Release/3_1_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 3_1_reloadable22/Release/3_1_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 3_1_reloadable22/Release/3_1_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 3_1_reloadable22/Release/3_1_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 3_1_reloadable22/Release/3_1_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 3_1_reloadable22/Release/3_1_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 3_1_reloadable22/Release/3_1_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 3_2_reloadable22/Release/3_2_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 3_2_reloadable22/Release/3_2_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 3_2_reloadable22/Release/3_2_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 3_2_reloadable22/Release/3_2_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 3_2_reloadable22/Release/3_2_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 3_2_reloadable22/Release/3_2_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 3_2_reloadable22/Release/3_2_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 3_2_reloadable22/Release/3_2_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 3_2_reloadable22/Release/3_2_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable22/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22 3_3_reloadable22/Release/3_3_reloadable22 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.map 3_3_reloadable22/Release/3_3_reloadable22.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.lst 3_3_reloadable22/Release/3_3_reloadable22.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.calltree 3_3_reloadable22/Release/3_3_reloadable22.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.sdr 3_3_reloadable22/Release/3_3_reloadable22.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.srv 3_3_reloadable22/Release/3_3_reloadable22.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.txt 3_3_reloadable22/Release/3_3_reloadable22.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmic2 3_3_reloadable22/Release/3_3_reloadable22.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable22/Release/0_0_reloadable22.cmico 3_3_reloadable22/Release/3_3_reloadable22.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable23.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable23/scripts/0_0_reloadable23.prx) 2>&1 |& tee -a 0_0_reloadable23/0_0_reloadable23.log 0_0_reloadable23/timestamped_log/0_0_reloadable23.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable23.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable23.ll -o../Release/chesswork6583/0_0_reloadable23.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork6583 ../Release/chesswork6583/0_0_reloadable23.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable23-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +chess-backend 0_0_reloadable23-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +chess-backend 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +chess-backend 0_0_reloadable23-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable23-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable23-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable23-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +chess-backend 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable23-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable23-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable23-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable23-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z15concat_adf_initv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable23-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable23-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable23-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z13_b820_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf/concat_adf_wrapper.cpp", line 112, column 4: (loop #961) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. +chess-backend 0_0_reloadable23-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable23-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable23-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable23-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable23-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable23-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN12mllib_graphs18concat_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_SC_EEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNSD_ISE_SG__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable23-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable23-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable23 -L --common 0_0_reloadable23-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable23.objlist -o../0_0_reloadable23.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable23.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable23" +bridge -o../Release/0_0_reloadable23 ../Release/0_0_reloadable23.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable23.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork6583 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable23 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (198 errors, 5 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable23/Release/0_0_reloadable23 >> 0_0_reloadable23/Release/0_0_reloadable23.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 0_1_reloadable23/Release/0_1_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 0_1_reloadable23/Release/0_1_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 0_1_reloadable23/Release/0_1_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 0_1_reloadable23/Release/0_1_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 0_1_reloadable23/Release/0_1_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 0_1_reloadable23/Release/0_1_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 0_1_reloadable23/Release/0_1_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 0_1_reloadable23/Release/0_1_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 0_1_reloadable23/Release/0_1_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 0_2_reloadable23/Release/0_2_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 0_2_reloadable23/Release/0_2_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 0_2_reloadable23/Release/0_2_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 0_2_reloadable23/Release/0_2_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 0_2_reloadable23/Release/0_2_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 0_2_reloadable23/Release/0_2_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 0_2_reloadable23/Release/0_2_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 0_2_reloadable23/Release/0_2_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 0_2_reloadable23/Release/0_2_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 0_3_reloadable23/Release/0_3_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 0_3_reloadable23/Release/0_3_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 0_3_reloadable23/Release/0_3_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 0_3_reloadable23/Release/0_3_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 0_3_reloadable23/Release/0_3_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 0_3_reloadable23/Release/0_3_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 0_3_reloadable23/Release/0_3_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 0_3_reloadable23/Release/0_3_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 0_3_reloadable23/Release/0_3_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 1_0_reloadable23/Release/1_0_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 1_0_reloadable23/Release/1_0_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 1_0_reloadable23/Release/1_0_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 1_0_reloadable23/Release/1_0_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 1_0_reloadable23/Release/1_0_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 1_0_reloadable23/Release/1_0_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 1_0_reloadable23/Release/1_0_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 1_0_reloadable23/Release/1_0_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 1_0_reloadable23/Release/1_0_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 1_1_reloadable23/Release/1_1_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 1_1_reloadable23/Release/1_1_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 1_1_reloadable23/Release/1_1_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 1_1_reloadable23/Release/1_1_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 1_1_reloadable23/Release/1_1_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 1_1_reloadable23/Release/1_1_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 1_1_reloadable23/Release/1_1_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 1_1_reloadable23/Release/1_1_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 1_1_reloadable23/Release/1_1_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 1_2_reloadable23/Release/1_2_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 1_2_reloadable23/Release/1_2_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 1_2_reloadable23/Release/1_2_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 1_2_reloadable23/Release/1_2_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 1_2_reloadable23/Release/1_2_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 1_2_reloadable23/Release/1_2_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 1_2_reloadable23/Release/1_2_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 1_2_reloadable23/Release/1_2_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 1_2_reloadable23/Release/1_2_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 1_3_reloadable23/Release/1_3_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 1_3_reloadable23/Release/1_3_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 1_3_reloadable23/Release/1_3_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 1_3_reloadable23/Release/1_3_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 1_3_reloadable23/Release/1_3_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 1_3_reloadable23/Release/1_3_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 1_3_reloadable23/Release/1_3_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 1_3_reloadable23/Release/1_3_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 1_3_reloadable23/Release/1_3_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 2_0_reloadable23/Release/2_0_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 2_0_reloadable23/Release/2_0_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 2_0_reloadable23/Release/2_0_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 2_0_reloadable23/Release/2_0_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 2_0_reloadable23/Release/2_0_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 2_0_reloadable23/Release/2_0_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 2_0_reloadable23/Release/2_0_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 2_0_reloadable23/Release/2_0_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 2_0_reloadable23/Release/2_0_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 2_1_reloadable23/Release/2_1_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 2_1_reloadable23/Release/2_1_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 2_1_reloadable23/Release/2_1_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 2_1_reloadable23/Release/2_1_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 2_1_reloadable23/Release/2_1_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 2_1_reloadable23/Release/2_1_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 2_1_reloadable23/Release/2_1_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 2_1_reloadable23/Release/2_1_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 2_1_reloadable23/Release/2_1_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 2_2_reloadable23/Release/2_2_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 2_2_reloadable23/Release/2_2_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 2_2_reloadable23/Release/2_2_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 2_2_reloadable23/Release/2_2_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 2_2_reloadable23/Release/2_2_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 2_2_reloadable23/Release/2_2_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 2_2_reloadable23/Release/2_2_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 2_2_reloadable23/Release/2_2_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 2_2_reloadable23/Release/2_2_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 2_3_reloadable23/Release/2_3_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 2_3_reloadable23/Release/2_3_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 2_3_reloadable23/Release/2_3_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 2_3_reloadable23/Release/2_3_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 2_3_reloadable23/Release/2_3_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 2_3_reloadable23/Release/2_3_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 2_3_reloadable23/Release/2_3_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 2_3_reloadable23/Release/2_3_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 2_3_reloadable23/Release/2_3_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 3_0_reloadable23/Release/3_0_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 3_0_reloadable23/Release/3_0_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 3_0_reloadable23/Release/3_0_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 3_0_reloadable23/Release/3_0_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 3_0_reloadable23/Release/3_0_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 3_0_reloadable23/Release/3_0_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 3_0_reloadable23/Release/3_0_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 3_0_reloadable23/Release/3_0_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 3_0_reloadable23/Release/3_0_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 3_1_reloadable23/Release/3_1_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 3_1_reloadable23/Release/3_1_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 3_1_reloadable23/Release/3_1_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 3_1_reloadable23/Release/3_1_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 3_1_reloadable23/Release/3_1_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 3_1_reloadable23/Release/3_1_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 3_1_reloadable23/Release/3_1_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 3_1_reloadable23/Release/3_1_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 3_1_reloadable23/Release/3_1_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 3_2_reloadable23/Release/3_2_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 3_2_reloadable23/Release/3_2_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 3_2_reloadable23/Release/3_2_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 3_2_reloadable23/Release/3_2_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 3_2_reloadable23/Release/3_2_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 3_2_reloadable23/Release/3_2_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 3_2_reloadable23/Release/3_2_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 3_2_reloadable23/Release/3_2_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 3_2_reloadable23/Release/3_2_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable23/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23 3_3_reloadable23/Release/3_3_reloadable23 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.map 3_3_reloadable23/Release/3_3_reloadable23.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.lst 3_3_reloadable23/Release/3_3_reloadable23.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.calltree 3_3_reloadable23/Release/3_3_reloadable23.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.sdr 3_3_reloadable23/Release/3_3_reloadable23.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.srv 3_3_reloadable23/Release/3_3_reloadable23.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.txt 3_3_reloadable23/Release/3_3_reloadable23.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmic2 3_3_reloadable23/Release/3_3_reloadable23.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable23/Release/0_0_reloadable23.cmico 3_3_reloadable23/Release/3_3_reloadable23.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable24.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable24/scripts/0_0_reloadable24.prx) 2>&1 |& tee -a 0_0_reloadable24/0_0_reloadable24.log 0_0_reloadable24/timestamped_log/0_0_reloadable24.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable24.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable24.ll -o../Release/chesswork6947/0_0_reloadable24.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork6947 ../Release/chesswork6947/0_0_reloadable24.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable24-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +chess-backend 0_0_reloadable24-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +chess-backend 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +chess-backend 0_0_reloadable24-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable24-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable24-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable24-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable24-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN12mllib_graphs18resize_adf_wrapperI8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_Li1ELi0ELi2EEEvRNS2_9io_bufferIT_NS2_9direction2inET0_EERNS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z13_b853_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z13_b853_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z13_b853_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z13_b853_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z13_b853_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z13_b853_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z13_b853_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable24-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z13_b896_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z14_b1638_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z14_b1655_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z13_b891_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable24-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z14_b1672_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable24-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable24-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z23resize_bilinear_scale_2I8bfloat16Qsr5mllib5utilsE11is_one_of_vIT_S0_EEvPS1_PfS2_jjjj_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable24-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable24-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable24-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable24-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable24 -L --common 0_0_reloadable24-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable24.objlist -o../0_0_reloadable24.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable24.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable24" +bridge -o../Release/0_0_reloadable24 ../Release/0_0_reloadable24.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable24.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork6947 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable24 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (189 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable24/Release/0_0_reloadable24 >> 0_0_reloadable24/Release/0_0_reloadable24.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 0_1_reloadable24/Release/0_1_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 0_1_reloadable24/Release/0_1_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 0_1_reloadable24/Release/0_1_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 0_1_reloadable24/Release/0_1_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 0_1_reloadable24/Release/0_1_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 0_1_reloadable24/Release/0_1_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 0_1_reloadable24/Release/0_1_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 0_1_reloadable24/Release/0_1_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 0_1_reloadable24/Release/0_1_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 0_2_reloadable24/Release/0_2_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 0_2_reloadable24/Release/0_2_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 0_2_reloadable24/Release/0_2_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 0_2_reloadable24/Release/0_2_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 0_2_reloadable24/Release/0_2_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 0_2_reloadable24/Release/0_2_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 0_2_reloadable24/Release/0_2_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 0_2_reloadable24/Release/0_2_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 0_2_reloadable24/Release/0_2_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 0_3_reloadable24/Release/0_3_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 0_3_reloadable24/Release/0_3_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 0_3_reloadable24/Release/0_3_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 0_3_reloadable24/Release/0_3_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 0_3_reloadable24/Release/0_3_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 0_3_reloadable24/Release/0_3_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 0_3_reloadable24/Release/0_3_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 0_3_reloadable24/Release/0_3_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 0_3_reloadable24/Release/0_3_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 1_0_reloadable24/Release/1_0_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 1_0_reloadable24/Release/1_0_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 1_0_reloadable24/Release/1_0_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 1_0_reloadable24/Release/1_0_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 1_0_reloadable24/Release/1_0_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 1_0_reloadable24/Release/1_0_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 1_0_reloadable24/Release/1_0_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 1_0_reloadable24/Release/1_0_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 1_0_reloadable24/Release/1_0_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 1_1_reloadable24/Release/1_1_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 1_1_reloadable24/Release/1_1_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 1_1_reloadable24/Release/1_1_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 1_1_reloadable24/Release/1_1_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 1_1_reloadable24/Release/1_1_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 1_1_reloadable24/Release/1_1_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 1_1_reloadable24/Release/1_1_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 1_1_reloadable24/Release/1_1_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 1_1_reloadable24/Release/1_1_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 1_2_reloadable24/Release/1_2_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 1_2_reloadable24/Release/1_2_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 1_2_reloadable24/Release/1_2_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 1_2_reloadable24/Release/1_2_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 1_2_reloadable24/Release/1_2_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 1_2_reloadable24/Release/1_2_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 1_2_reloadable24/Release/1_2_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 1_2_reloadable24/Release/1_2_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 1_2_reloadable24/Release/1_2_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 1_3_reloadable24/Release/1_3_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 1_3_reloadable24/Release/1_3_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 1_3_reloadable24/Release/1_3_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 1_3_reloadable24/Release/1_3_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 1_3_reloadable24/Release/1_3_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 1_3_reloadable24/Release/1_3_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 1_3_reloadable24/Release/1_3_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 1_3_reloadable24/Release/1_3_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 1_3_reloadable24/Release/1_3_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 2_0_reloadable24/Release/2_0_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 2_0_reloadable24/Release/2_0_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 2_0_reloadable24/Release/2_0_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 2_0_reloadable24/Release/2_0_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 2_0_reloadable24/Release/2_0_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 2_0_reloadable24/Release/2_0_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 2_0_reloadable24/Release/2_0_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 2_0_reloadable24/Release/2_0_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 2_0_reloadable24/Release/2_0_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 2_1_reloadable24/Release/2_1_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 2_1_reloadable24/Release/2_1_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 2_1_reloadable24/Release/2_1_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 2_1_reloadable24/Release/2_1_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 2_1_reloadable24/Release/2_1_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 2_1_reloadable24/Release/2_1_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 2_1_reloadable24/Release/2_1_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 2_1_reloadable24/Release/2_1_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 2_1_reloadable24/Release/2_1_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 2_2_reloadable24/Release/2_2_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 2_2_reloadable24/Release/2_2_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 2_2_reloadable24/Release/2_2_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 2_2_reloadable24/Release/2_2_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 2_2_reloadable24/Release/2_2_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 2_2_reloadable24/Release/2_2_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 2_2_reloadable24/Release/2_2_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 2_2_reloadable24/Release/2_2_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 2_2_reloadable24/Release/2_2_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 2_3_reloadable24/Release/2_3_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 2_3_reloadable24/Release/2_3_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 2_3_reloadable24/Release/2_3_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 2_3_reloadable24/Release/2_3_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 2_3_reloadable24/Release/2_3_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 2_3_reloadable24/Release/2_3_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 2_3_reloadable24/Release/2_3_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 2_3_reloadable24/Release/2_3_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 2_3_reloadable24/Release/2_3_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 3_0_reloadable24/Release/3_0_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 3_0_reloadable24/Release/3_0_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 3_0_reloadable24/Release/3_0_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 3_0_reloadable24/Release/3_0_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 3_0_reloadable24/Release/3_0_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 3_0_reloadable24/Release/3_0_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 3_0_reloadable24/Release/3_0_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 3_0_reloadable24/Release/3_0_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 3_0_reloadable24/Release/3_0_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 3_1_reloadable24/Release/3_1_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 3_1_reloadable24/Release/3_1_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 3_1_reloadable24/Release/3_1_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 3_1_reloadable24/Release/3_1_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 3_1_reloadable24/Release/3_1_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 3_1_reloadable24/Release/3_1_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 3_1_reloadable24/Release/3_1_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 3_1_reloadable24/Release/3_1_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 3_1_reloadable24/Release/3_1_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 3_2_reloadable24/Release/3_2_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 3_2_reloadable24/Release/3_2_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 3_2_reloadable24/Release/3_2_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 3_2_reloadable24/Release/3_2_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 3_2_reloadable24/Release/3_2_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 3_2_reloadable24/Release/3_2_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 3_2_reloadable24/Release/3_2_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 3_2_reloadable24/Release/3_2_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 3_2_reloadable24/Release/3_2_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable24/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24 3_3_reloadable24/Release/3_3_reloadable24 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.map 3_3_reloadable24/Release/3_3_reloadable24.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.lst 3_3_reloadable24/Release/3_3_reloadable24.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.calltree 3_3_reloadable24/Release/3_3_reloadable24.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.sdr 3_3_reloadable24/Release/3_3_reloadable24.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.srv 3_3_reloadable24/Release/3_3_reloadable24.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.txt 3_3_reloadable24/Release/3_3_reloadable24.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmic2 3_3_reloadable24/Release/3_3_reloadable24.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable24/Release/0_0_reloadable24.cmico 3_3_reloadable24/Release/3_3_reloadable24.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/aie -O -j1 -f 0_0_reloadable25.elfgen.Makefile all 2>&1 + +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +set -o pipefail;(/usr/local/lib/python3.10/dist-packages/tps/lnx64/target_aie2p/bin/LNa64bin/chessmk -C Release_LLVM -P /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +P 4 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR +w +o ../Release 0_0_reloadable25/scripts/0_0_reloadable25.prx) 2>&1 |& tee -a 0_0_reloadable25/0_0_reloadable25.log 0_0_reloadable25/timestamped_log/0_0_reloadable25.log +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Configuration: Release_LLVM +Compiling "0_0_reloadable25.ll" +chess-clang --chess-proc-dir=/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -S -O2 -std=c++2a -fno-builtin-memcpy -mllvm -instcombine-code-sinking=false -mllvm -disable-lsr -mllvm -replexitval=never -mllvm -enable-load-pre=false -mllvm -chess-disable-add-to-or -mllvm -chess-combine-gep-indices=none -mllvm -chess-disable-fold-phi-of-loads -mllvm -chess-aainfo2chains-algo=4 -mllvm -chess-aggressive-aainfo=false -mllvm -chess-enable-indvarsimplify=0 -mllvm -chess-disable-cse-across-loopboundary -mllvm -chess-tbaa-detect-common-underlying-object=true -mllvm -chess-protect-llvm-global-reg-access=true -fno-jump-tables -fno-discard-value-names -g ../../ir/0_0_reloadable25.ll -o../Release/chesswork7302/0_0_reloadable25.sfg --chess-proc-name=me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +noodle -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -iaie_core.h +Sinl +Olbb=200 +Opmsa +NOpld +Olzyinl +w../Release/chesswork7302 ../Release/chesswork7302/0_0_reloadable25.sfg +Q1=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q2=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Q3=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl +Qfast=+Sinl,+Olbb=1000,+Opmsa,+NOpld,+Olzyinl,+Opfp +Qs=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl +Qz=+Sinl,+Olbb=200,+Opmsa,+NOpld,+Olzyinl me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable25-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +chess-backend 0_0_reloadable25-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +chess-backend 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +chess-backend 0_0_reloadable25-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/conv/conv2d_bf16.h", line 702: (loop #3) + further loop software pipelining (to 3 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 6 + ... consider annotating the loop with `chess_loop_range(6,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z14conv2d_maxpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_SC_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z21convert_bf16_to_bfp16I8bfloat16Lb0EEvPT_PS0_RK13BfToBfpParams_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN41elementwise_binary_attribute_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN31elementwise_binary_broadcastingI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z40superkernel_add1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_clipIS0_E20clip_internal_paramsIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z18superkernel_clip1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166, column 4: in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_binary_shared.h", line 166: (loop #19) + further loop software pipelining (to 2 cycles) is feasible with `chess_prepare_for_pipelining' + but requires a minimum loop count of 7 + ... consider annotating the loop with `chess_loop_range(7,)' if applicable, + ... or remove the current `chess_loop_range(4,)` annotation + +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z40superkernel_mul1d_attribute_broadcastingRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outEN_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZL19shared_run_backboneI8bfloat16L5act_t0EEKvPT_S4_S4_R27elementwise_binary_params_tI15shared_params_tIS3_EE_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat1626mul_impl_broadcasting_attrIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z21superkernel_sigmoid1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncES_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE5setupER26elementwise_unary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1619elementwise_sigmoidIS0_E26sigmoid_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z18superkernel_tanh1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA_S_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z24setup_conv2d_bf16_paramsILb1ELb0EEvPKjR18conv2d_bf16_paramshh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z22setup_avgpool2d_paramsI8bfloat16EvPT_R25avgpool2d_internal_paramsIS1_Eh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z19superkernel_avgpoolRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asyncESA__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z11conv2d_bf16ILh1EL5act_t0E8bfloat16S1_S1_N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_NS3_IS5_NS6_5asyncES9_SB_EELb0ELb0ELb1ELb0EEvRNS2_9io_bufferIT_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE21shared_setup_backboneER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS5_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168add_implIS0_L5act_t0EE15shared_params_tIS0_ELS2_0EE5setupER27elementwise_binary_params_tIS5_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/elementwise_unary.h", line 154, column 8: (loop #3) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z17superkernel_add1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN17elementwise_unaryI8bfloat1616elementwise_tanhIS0_E23tanh_templated_params_tIS0_EE3runEPS0_S6_R26elementwise_unary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN18elementwise_binaryIJ8bfloat168mul_implIS0_E15shared_params_tIS0_EEE3runEPS0_S6_S6_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z17superkernel_mul1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE21shared_setup_backboneER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +chess-backend 0_0_reloadable25-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE5setupER27elementwise_binary_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN25elementwise_binary_sharedI8bfloat168sub_implIS0_E15shared_params_tIS0_EL5act_t0EE3runEPS0_S7_S7_R27elementwise_binary_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z17superkernel_sub1dRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC_EEEER_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZL27setup_conv2d_dw_params_bf16PKjR21conv2d_dw_bf16_paramsh_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z9conv2d_dwILh1E8bfloat16S0_S0_N3adf16io_buffer_configINS1_7extentsIJEEENS1_7locking4syncENS1_10addressing6linearENS1_6marginILj0EEEEESB_NS2_IS4_NS5_5asyncES8_SA_EEQsr3stdE9is_same_vIT0_S0_EEvRNS1_9io_bufferISE__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z22superkernel_conv2d_dwcRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5asy_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE5setupER18reduce_c8_params_tIS4_EPKv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z6pad_3dIL11pad_3d_mode0E8bfloat16Li1EEvPT0_S3_R15pad_3d_params_t_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z26superkernel_conv_eltbinaryRN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEESF_RNS0_IS1_S3_NS4_IS6_NS7_5asyncESA_SC__ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z9avgpool2dILh1E8bfloat16Qsr5mllib5utilsE11is_one_of_vIT0_ahS0_EEvPS1_S2_R25avgpool2d_internal_paramsIS1_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z26superkernel_reduce_mean_c8RN3adf9io_bufferI8bfloat16NS_9direction2inENS_16io_buffer_configINS_7extentsIJEEENS_7locking4syncENS_10addressing6linearENS_6marginILj0EEEEEEERA16_KjRNS0_IS1_NS2_3outENS4_IS6_NS7_5as_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z11slice_hcwc8I8bfloat16EvPT_S2_R18slice_hcwc8_params_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist1 -k64 --common 0_0_reloadable25-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist1 -k64 --common 0_0_reloadable25-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z13_b806_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend 0_0_reloadable25-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--cosel -m +ef +s -M3 --common 0_0_reloadable25-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z13_b906_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN12mllib_graphs17slice_adf_wrapperILi1E8bfloat16N3adf16io_buffer_configINS2_7extentsIJEEENS2_7locking4syncENS2_10addressing6linearENS2_6marginILj0EEEEESC_EEvRNS2_9io_bufferIT0_NS2_9direction2inET1_EERNSD_ISE_NS_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +chess-backend 0_0_reloadable25-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x --print-subtools --cosel -m +ef +s -M3 --amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --mist1 -k64 --showcolor -b -Obbl --mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--cosel -m +ef +s -M3 --common 0_0_reloadable25-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +chess-backend --gvt me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation --tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--showcolor -b -Obbl --common 0_0_reloadable25-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--amnesia -p10 -q2 -ecrSCDEn -ecrMCDEn -ecrVaddSign -ecrUnpackSign -ecrPackSign -ecrUPSSign -ecrUPSMode -ecrSRSSign -ecrSRSMode -ecrF2IMask -ecrUnpackSize -ecrPackSize -ecrSat -ecrRnd +Oefc +Opbr +Odhls +Oprefer-local-reg-moves -Onocb --common 0_0_reloadable25-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_Z13_b886_wrapperPPv_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--showcolor -b -Obbl --common 0_0_reloadable25-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-kernelWrapper_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation -x +--mist1 -k64 --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--showcolor -b -Obbl --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +--mist2 -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 +Omod -k64 +Omsbr=100 +Opnll +A +pnopALU +pnopLDA +pnopLDB +pnopSTS +pnopVEC +Ofexm +Omsanafs +Onzmem +Onombt +Ochex +Omsmfi +Omslactc=lckLdaRsrc_E1,lckLdbRsrc_E1:2 +Odra +Oslr=crSRSSign +Oslr=crUPSMode +Oslr=crSRSMode +Oslr=crRnd +Oslr=crSat +Onop-syntax=NOPA +Onop-syntax=NOPB +Onop-syntax=NOPM +Onop-syntax=NOPV +Onop-syntax=NOPS +Onop-syntax=NOPX +Onop-syntax=NOPXM --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Warning in "/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common/../../include/misc/reduce_mean_c8_impl.h", line 223, column 9: (loop #95) + `chess_prepare_for_pipelining' was not used: + loop software pipelining has not succeeded. + This may result in a redundant 'loop_count += 0' instruction. +--tale -g -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -V0_0_reloadable25 -L --common 0_0_reloadable25-F_ZN18reduce_skeleton_c8I8bfloat1619reduce_mean_c8_implIS0_E23reduce_mean_c8_params_tIS0_EE3runEPS0_S6_R18reduce_c8_params_tIS4_E_ me /usr/local/lib/python3.10/dist-packages/data/aie2p/lib +H/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/elongation +bridge -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -i -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 0_0_reloadable25.objlist -o../0_0_reloadable25.o -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno ../Release/0_0_reloadable25.o me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Linking "../Release/0_0_reloadable25" +bridge -o../Release/0_0_reloadable25 ../Release/0_0_reloadable25.o -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/isg -g -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -c0_0_reloadable25.bcf -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/softfloat/lib/Release -L/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/lib/Release_LLVM -lme -lc -lm -lc++lite -lsoftfloat -S -export-locals -iconfig extra_memories.bcf -yTM -m -fC -fS -fH +m -T +work ../Release/chesswork7302 -pme +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +darts -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib -d -h -I/usr/local/lib/python3.10/dist-packages/include -I/app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I/usr/local/lib/python3.10/site-packages/include/aie_api -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I/usr/local/lib/python3.10/dist-packages/vitis_mllib -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I/usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -I/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/. -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libcxx-lite/include -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime_cxx/libs/libcxx-9.0.0/include-lite -I/usr/local/lib/python3.10/dist-packages/data/aie2p/lib/runtime/include -D__AIENGINE__ -D__AIE_ARCH__=21 -DDEPLOYMENT_ELF=1 -D__LOCK_FENCE_MODE__=0 -D__IO_BUFFER_FORCE_LIGHT_WEIGHT__ -DAIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16=1 -DAIE_OPTION_SCALAR_FLOAT_ON_VECTOR -D__tct_tgt__=241219 -L +Ihex +nanno +u ../Release/0_0_reloadable25 me +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +Compilation finished successfully (177 errors, 4 warnings) +(readelf --debug-dump=decodedline 0_0_reloadable25/Release/0_0_reloadable25 >> 0_0_reloadable25/Release/0_0_reloadable25.txt) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_1_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 0_1_reloadable25/Release/0_1_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 0_1_reloadable25/Release/0_1_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 0_1_reloadable25/Release/0_1_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 0_1_reloadable25/Release/0_1_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 0_1_reloadable25/Release/0_1_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 0_1_reloadable25/Release/0_1_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 0_1_reloadable25/Release/0_1_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 0_1_reloadable25/Release/0_1_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 0_1_reloadable25/Release/0_1_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_2_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 0_2_reloadable25/Release/0_2_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 0_2_reloadable25/Release/0_2_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 0_2_reloadable25/Release/0_2_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 0_2_reloadable25/Release/0_2_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 0_2_reloadable25/Release/0_2_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 0_2_reloadable25/Release/0_2_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 0_2_reloadable25/Release/0_2_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 0_2_reloadable25/Release/0_2_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 0_2_reloadable25/Release/0_2_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 0_3_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 0_3_reloadable25/Release/0_3_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 0_3_reloadable25/Release/0_3_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 0_3_reloadable25/Release/0_3_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 0_3_reloadable25/Release/0_3_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 0_3_reloadable25/Release/0_3_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 0_3_reloadable25/Release/0_3_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 0_3_reloadable25/Release/0_3_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 0_3_reloadable25/Release/0_3_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 0_3_reloadable25/Release/0_3_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_0_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 1_0_reloadable25/Release/1_0_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 1_0_reloadable25/Release/1_0_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 1_0_reloadable25/Release/1_0_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 1_0_reloadable25/Release/1_0_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 1_0_reloadable25/Release/1_0_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 1_0_reloadable25/Release/1_0_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 1_0_reloadable25/Release/1_0_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 1_0_reloadable25/Release/1_0_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 1_0_reloadable25/Release/1_0_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_1_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 1_1_reloadable25/Release/1_1_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 1_1_reloadable25/Release/1_1_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 1_1_reloadable25/Release/1_1_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 1_1_reloadable25/Release/1_1_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 1_1_reloadable25/Release/1_1_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 1_1_reloadable25/Release/1_1_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 1_1_reloadable25/Release/1_1_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 1_1_reloadable25/Release/1_1_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 1_1_reloadable25/Release/1_1_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_2_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 1_2_reloadable25/Release/1_2_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 1_2_reloadable25/Release/1_2_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 1_2_reloadable25/Release/1_2_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 1_2_reloadable25/Release/1_2_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 1_2_reloadable25/Release/1_2_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 1_2_reloadable25/Release/1_2_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 1_2_reloadable25/Release/1_2_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 1_2_reloadable25/Release/1_2_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 1_2_reloadable25/Release/1_2_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 1_3_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 1_3_reloadable25/Release/1_3_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 1_3_reloadable25/Release/1_3_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 1_3_reloadable25/Release/1_3_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 1_3_reloadable25/Release/1_3_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 1_3_reloadable25/Release/1_3_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 1_3_reloadable25/Release/1_3_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 1_3_reloadable25/Release/1_3_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 1_3_reloadable25/Release/1_3_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 1_3_reloadable25/Release/1_3_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_0_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 2_0_reloadable25/Release/2_0_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 2_0_reloadable25/Release/2_0_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 2_0_reloadable25/Release/2_0_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 2_0_reloadable25/Release/2_0_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 2_0_reloadable25/Release/2_0_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 2_0_reloadable25/Release/2_0_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 2_0_reloadable25/Release/2_0_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 2_0_reloadable25/Release/2_0_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 2_0_reloadable25/Release/2_0_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_1_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 2_1_reloadable25/Release/2_1_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 2_1_reloadable25/Release/2_1_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 2_1_reloadable25/Release/2_1_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 2_1_reloadable25/Release/2_1_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 2_1_reloadable25/Release/2_1_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 2_1_reloadable25/Release/2_1_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 2_1_reloadable25/Release/2_1_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 2_1_reloadable25/Release/2_1_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 2_1_reloadable25/Release/2_1_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_2_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 2_2_reloadable25/Release/2_2_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 2_2_reloadable25/Release/2_2_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 2_2_reloadable25/Release/2_2_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 2_2_reloadable25/Release/2_2_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 2_2_reloadable25/Release/2_2_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 2_2_reloadable25/Release/2_2_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 2_2_reloadable25/Release/2_2_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 2_2_reloadable25/Release/2_2_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 2_2_reloadable25/Release/2_2_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 2_3_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 2_3_reloadable25/Release/2_3_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 2_3_reloadable25/Release/2_3_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 2_3_reloadable25/Release/2_3_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 2_3_reloadable25/Release/2_3_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 2_3_reloadable25/Release/2_3_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 2_3_reloadable25/Release/2_3_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 2_3_reloadable25/Release/2_3_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 2_3_reloadable25/Release/2_3_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 2_3_reloadable25/Release/2_3_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_0_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 3_0_reloadable25/Release/3_0_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 3_0_reloadable25/Release/3_0_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 3_0_reloadable25/Release/3_0_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 3_0_reloadable25/Release/3_0_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 3_0_reloadable25/Release/3_0_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 3_0_reloadable25/Release/3_0_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 3_0_reloadable25/Release/3_0_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 3_0_reloadable25/Release/3_0_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 3_0_reloadable25/Release/3_0_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_1_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 3_1_reloadable25/Release/3_1_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 3_1_reloadable25/Release/3_1_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 3_1_reloadable25/Release/3_1_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 3_1_reloadable25/Release/3_1_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 3_1_reloadable25/Release/3_1_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 3_1_reloadable25/Release/3_1_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 3_1_reloadable25/Release/3_1_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 3_1_reloadable25/Release/3_1_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 3_1_reloadable25/Release/3_1_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_2_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 3_2_reloadable25/Release/3_2_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 3_2_reloadable25/Release/3_2_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 3_2_reloadable25/Release/3_2_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 3_2_reloadable25/Release/3_2_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 3_2_reloadable25/Release/3_2_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 3_2_reloadable25/Release/3_2_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 3_2_reloadable25/Release/3_2_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 3_2_reloadable25/Release/3_2_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 3_2_reloadable25/Release/3_2_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +set -o pipefail;mkdir -p 3_3_reloadable25/Release +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25 3_3_reloadable25/Release/3_3_reloadable25 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.map 3_3_reloadable25/Release/3_3_reloadable25.map +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.lst 3_3_reloadable25/Release/3_3_reloadable25.lst +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.calltree 3_3_reloadable25/Release/3_3_reloadable25.calltree +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.sdr 3_3_reloadable25/Release/3_3_reloadable25.sdr +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.srv 3_3_reloadable25/Release/3_3_reloadable25.srv +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.txt 3_3_reloadable25/Release/3_3_reloadable25.txt +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmic2 3_3_reloadable25/Release/3_3_reloadable25.cmic2 +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +cp 0_0_reloadable25/Release/0_0_reloadable25.cmico 3_3_reloadable25/Release/3_3_reloadable25.cmico +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/aie' +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c4f0 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c464 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c470 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c468 +WARNING: AIE driver read operation not supported for elf packets - XAIE_IO_MASKWRITE 0x250c458 +INFO: [aiecompiler 77-22140] Generating Control Packet Binary +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0/elf_ctrl_pkt.bin +for file: Work/aie/0_0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1/elf_ctrl_pkt.bin +for file: Work/aie/0_1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2/elf_ctrl_pkt.bin +for file: Work/aie/0_2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3/elf_ctrl_pkt.bin +for file: Work/aie/0_3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0/elf_ctrl_pkt.bin +for file: Work/aie/1_0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1/elf_ctrl_pkt.bin +for file: Work/aie/1_1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2/elf_ctrl_pkt.bin +for file: Work/aie/1_2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3/elf_ctrl_pkt.bin +for file: Work/aie/1_3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0/elf_ctrl_pkt.bin +for file: Work/aie/2_0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1/elf_ctrl_pkt.bin +for file: Work/aie/2_1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2/elf_ctrl_pkt.bin +for file: Work/aie/2_2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3/elf_ctrl_pkt.bin +for file: Work/aie/2_3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0/elf_ctrl_pkt.bin +for file: Work/aie/3_0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1/elf_ctrl_pkt.bin +for file: Work/aie/3_1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2/elf_ctrl_pkt.bin +for file: Work/aie/3_2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3/elf_ctrl_pkt.bin +for file: Work/aie/3_3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6148 +gcount is: 6148 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable0/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable0/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18660 +gcount is: 18660 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable1/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable1/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15740 +gcount is: 15740 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable2/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable2/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable3/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable3/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable4/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable4/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable5/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable5/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable6/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable6/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable7/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable7/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable8/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable8/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable9/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable9/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable10/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable10/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable11/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable11/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable12/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable12/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable13/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable13/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19344 +gcount is: 19344 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable14/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable14/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable15/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable15/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21564 +gcount is: 21564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable16/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable16/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable17/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable17/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:18076 +gcount is: 18076 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable18/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable18/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:15504 +gcount is: 15504 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable19/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable19/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:19580 +gcount is: 19580 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable20/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable20/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21160 +gcount is: 21160 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable21/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable21/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:17816 +gcount is: 17816 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable22/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable22/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21244 +gcount is: 21244 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable23/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable23/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21620 +gcount is: 21620 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable24/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable24/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:21484 +gcount is: 21484 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_0_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/0_0_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_1_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/0_1_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_2_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/0_2_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/0_3_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/0_3_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_0_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/1_0_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_1_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/1_1_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_2_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/1_2_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/1_3_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/1_3_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_0_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/2_0_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_1_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/2_1_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_2_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/2_2_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/2_3_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/2_3_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_0_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/3_0_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_1_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/3_1_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_2_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/3_2_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +inside merge ctrl pkt +Inside merge control packet binary for file: Work/aie/3_3_reloadable25/elf_ctrl_pkt.bin +for file: Work/aie/3_3_reloadable25/elf_ctrl_pkt.bin +ctrlPktBinaryFileStream size is:6564 +gcount is: 6564 +INFO: [aiecompiler 77-22511] Creating External Buffers Json Work/ps/c_rts/external_buffer_id.json +INFO: [aiecompiler 77-22141] Generating Transaction Binary +INFO: [aiecompiler 77-6311] Opened file : Work/ps/cdo/generated-source/cdo_main.cpp +INFO: [aiecompiler 77-6311] Opened file : Work/ps/cdo/generated-source/gen_cdo.cpp +INFO: [aiecompiler 77-6311] Opened file : Work/ps/cdo/generated-source/gen_cdo.h +INFO: [aiecompiler 77-6311] Opened file : Work/ps/cdo/Makefile +INFO: [aiecompiler 77-404] Executing Cmd: make -C Work/ps/cdo -f Makefile all +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +make: Entering directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/ps/cdo' +(rm -rf generated-objects/* *.bin ../../config/aie_resources.bin) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +(/usr/local/lib/python3.10/dist-packages/tps/lnx64/gcc/bin/g++ -fPIC -std=c++17 -Wno-deprecated-declarations -Og -flto -D__AIE_ARCH__=21 -D__CDO__ -D__PS_INIT_AIE__ -D__AIE2IPU_BASE_ADDR_CDO__ -I /usr/local/lib/python3.10/dist-packages/include -I /usr/local/lib/python3.10/dist-packages/include/adf -I . -I . -I /usr/local/lib/python3.10/dist-packages/include/drivers/aiengine_aig -I ../../.. -I /app/vaiml_1.3_examples/camo/./segmentation_1_4_0_fp32_combined/vaiml_par_0/0/backend -I /usr/local/lib/python3.10/site-packages/include/aie_api -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/include/common -I /usr/local/lib/python3.10/dist-packages/vitis_mllib -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L1/include/misc -I /usr/local/lib/python3.10/dist-packages/vitis_mllib/L2/src/ml_adf -o "generated-objects/cdo_main.out" generated-source/gen_cdo.cpp generated-source/cdo_main.cpp -Wl,--allow-shlib-undefined -lcdo_driver -lxaiengine_aig -ladf_api_aig -l stdc++fs -L /usr/local/lib/python3.10/dist-packages/lib/lnx64.o ) +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +generated-objects/cdo_main.out --work-dir-path /app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work +ERROR: ld.so: object '/lib64/libudev.so.1' from LD_PRELOAD cannot be preloaded (cannot open shared object file): ignored. +XAIEFAL: INFO: Resource group Avail is created. +XAIEFAL: INFO: Resource group Static is created. +XAIEFAL: INFO: Resource group Generic is created. +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():398: Program memory section is skipped as XAIE_LOAD_ELF_TXT flag is not set +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +[AIE WARNING] _XAie_WriteProgramSection():407: Mismatch in program header to data memory loadable section. Skipping this program section. +Generating: aie_cdo_elfs_dm.bin +Generating: aie_cdo_elfs_pm.bin +make: Leaving directory '/app/vaiml_1.3_examples/camo/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/Work/ps/cdo' +INFO: [aiecompiler 77-6439] Run completed. Find additional information in: + Guidance: Work/reports/guidance.html + +INFO: [aiecompiler 77-6440] Use the vitis_analyzer tool to visualize and navigate the relevant reports. Run: + vitis_analyzer Work/top.aiecompile_summary + +Compilation Complete +(WARNING:62, CRITICAL-WARNING:0, ERROR:0) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/aiesim_cmd_opts.txt b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/aiesim_cmd_opts.txt new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/xcd.log b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/xcd.log new file mode 100644 index 0000000000000000000000000000000000000000..32c03218750d3f1b2b20a6f3198157263c248136 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/aiecompiler/xcd.log @@ -0,0 +1,30 @@ +Fri Mar 21 03:30:56 2025: Running xcd server. This server was built by gradle. +Fri Mar 21 03:30:56 2025: Server was asked to start on port: '34917' +Fri Mar 21 03:30:56 2025: Server is using token UUID: 'e14f6479-c44a-42cf-8f07-308e6ee9c741' +Fri Mar 21 03:30:56 2025: SESSION_MGR: creating session with session uuid: '6d04a3e9-ade5-4ab9-8d63-1f78fcd2ec76' and name 'default' +Fri Mar 21 03:30:56 2025: Attempting to start server on port '34917' +Fri Mar 21 03:30:56 2025: XCD main fifo: created main fifo '/tmp/xcdmaine14f6479-c44a-42cf-8f07-308e6ee9c741', fd read = 12 +Fri Mar 21 03:30:56 2025: Running Dispatch Server +Fri Mar 21 03:30:56 2025: Version 2.2.0 +Fri Mar 21 03:30:57 2025: Dispatch Server: Accepted socket connection from client +Fri Mar 21 03:30:57 2025: Starting Socket connection +Fri Mar 21 03:30:57 2025: EXCHANGE_TOKEN received, server token: e14f6479-c44a-42cf-8f07-308e6ee9c741, passed token: e14f6479-c44a-42cf-8f07-308e6ee9c741 +Fri Mar 21 03:30:57 2025: SESSION_MGR: creating session with session uuid: 'bef5681c-a6ee-487a-8b13-f1c9f1805b32' and name 'AIE Compilation' +Fri Mar 21 03:30:57 2025: KERNEL_SVC: creating session with uuid: 'bef5681c-a6ee-487a-8b13-f1c9f1805b32' and name 'AIE Compilation' +Fri Mar 21 03:30:57 2025: GUIDANCE_SMGR: creating guidance session with uuid: 'bef5681c-a6ee-487a-8b13-f1c9f1805b32' and name 'AIE Compilation' +Fri Mar 21 03:31:54 2025: Dispatch Server: Accepted socket connection from client +Fri Mar 21 03:31:54 2025: Starting Socket connection +Fri Mar 21 04:04:52 2025: SESSION_MGR: closing session for UUID 'bef5681c-a6ee-487a-8b13-f1c9f1805b32' with name '' +Fri Mar 21 04:04:52 2025: KERNEL_SVC: closing session for UUID 'bef5681c-a6ee-487a-8b13-f1c9f1805b32' with name '' +Fri Mar 21 04:04:52 2025: GUIDANCE_SMGR: closing session for UUID 'bef5681c-a6ee-487a-8b13-f1c9f1805b32' with name '' +Fri Mar 21 04:04:52 2025: STOP_SERVER received, server token: e14f6479-c44a-42cf-8f07-308e6ee9c741, passed token: e14f6479-c44a-42cf-8f07-308e6ee9c741 +Fri Mar 21 04:04:52 2025: Socket received request to stop server. +Fri Mar 21 04:04:52 2025: Dispatch Server: do_await_stop +Fri Mar 21 04:04:52 2025: Dispatch Server: stopping mainFifo_ +Fri Mar 21 04:04:52 2025: XCD main fifo: Stopping +Fri Mar 21 04:04:52 2025: XCD main fifo: do_read error code: Operation canceled +Fri Mar 21 04:04:52 2025: XCD main fifo: bytes_transferred: 0 +Fri Mar 21 04:04:52 2025: SERVER: destructor for server bound to port 34917 +Fri Mar 21 04:04:52 2025: XCD main fifo: Good unlink of /tmp/xcdmaine14f6479-c44a-42cf-8f07-308e6ee9c741 +Fri Mar 21 04:04:52 2025: Server exiting with status 0. +Fri Mar 21 04:04:52 2025: Closing log file. diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/buffer_info.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/buffer_info.json new file mode 100644 index 0000000000000000000000000000000000000000..c4cbe536d80c45179140093a512b088ef936ee1a --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/buffer_info.json @@ -0,0 +1,40268 @@ +{ + ".meta": { + "layout": [ + 4, + 4 + ], + "version": "1.2" + }, + "0001": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_for_input0_0_port0", + "compute_graph.ifm_ddr" + ], + "out": [ + "compute_graph.l2_1", + "compute_graph.l2l3_1_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[0].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[0].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[0].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[0].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[0].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[0].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[0].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[0].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[0].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[0].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[0].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[0].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[0].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[0].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[0].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[0].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 0, + 235520, + 235520 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_for_input0_0_port0" + ], + "l3": { + "ifm_ddr": { + "size": 235520 + } + }, + "l3_buffer_names": [ + "compute_graph.ifm_ddr" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_for_input0_0_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_1.out[0]", + "to": "compute_graph.l2l3_1_spill.in[0]" + }, + { + "from": "compute_graph.l2_1.out[1]", + "to": "compute_graph.l2l3_1_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr.out[0]", + "to": "compute_graph.L2_IFM_Buffer_for_input0_0_port0.in[0]" + }, + { + "from": "compute_graph.ifm_ddr.out[1]", + "to": "compute_graph.L2_IFM_Buffer_for_input0_0_port0.in[1]" + } + ], + "layer_name": "Div_2", + "layer_object_name": "compute_graph.flexml_layers[0]", + "layer_order": 1, + "mlir_dag_id_map": { + "dag_layer": 1, + "mlir_layer": 1 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[0].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[0].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[0].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[0].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[0].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[0].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[0].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[0].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[0].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[0].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[0].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[0].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[0].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[0].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[0].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[0].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 471040, + 245760, + 235520 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_1" + ], + "l3": { + "l2l3_1_spill": { + "size": 235520 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_1_spill" + ] + }, + "super_iter": 1 + }, + "0002": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_1_spill" + ], + "out": [ + "compute_graph.l2l3_2_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_1_spill": { + "size": 235520 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_1_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Slice_7", + "layer_object_name": "compute_graph.templated_graph_2", + "layer_order": 2, + "mlir_dag_id_map": { + "dag_layer": 2, + "mlir_layer": 2 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_2_spill": { + "size": 176640 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_2_spill" + ] + }, + "templated_graph": 1 + }, + "0003": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_2_spill", + "compute_graph.l2l3_scratch_0_3_spill" + ], + "out": [ + "compute_graph.l2l3_3_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_2_spill": { + "size": 176640 + }, + "l2l3_scratch_0_3_spill": { + "size": 942080 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_2_spill", + "compute_graph.l2l3_scratch_0_3_spill" + ] + }, + "kernel_name": [ + "BufferPadAdf" + ], + "layer_name": "Generated-#0", + "layer_object_name": "compute_graph.templated_graph_3", + "layer_order": 3, + "mlir_dag_id_map": { + "dag_layer": 3, + "mlir_layer": 3 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_3_spill": { + "size": 471040 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_3_spill" + ] + }, + "templated_graph": 1 + }, + "0004": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_3_spill" + ], + "out": [ + "compute_graph.l2l3_4_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_3_spill": { + "size": 471040 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_3_spill" + ] + }, + "kernel_name": [ + "Transpose4dAdf" + ], + "layer_name": "Generated-#2", + "layer_object_name": "compute_graph.templated_graph_4", + "layer_order": 4, + "mlir_dag_id_map": { + "dag_layer": 4, + "mlir_layer": 4 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_4_spill": { + "size": 471040 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_4_spill" + ] + }, + "templated_graph": 1 + }, + "0005": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_4_spill", + "compute_graph.l2l3_scratch_0_5_spill" + ], + "out": [ + "compute_graph.l2l3_5_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_4_spill": { + "size": 471040 + }, + "l2l3_scratch_0_5_spill": { + "size": 921600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_4_spill", + "compute_graph.l2l3_scratch_0_5_spill" + ] + }, + "kernel_name": [ + "BufferUnpadAdf" + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_294.in[2]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_294.in[3]" + } + ], + "layer_name": "Generated-#4", + "layer_object_name": "compute_graph.templated_graph_5", + "layer_order": 5, + "mlir_dag_id_map": { + "dag_layer": 5, + "mlir_layer": 5 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_5_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_5_spill" + ] + }, + "templated_graph": 1 + }, + "0006": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1", + "compute_graph.const_ifm_ddr_5", + "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0", + "compute_graph.l2l3_5_spill" + ], + "out": [ + "compute_graph.l2_6", + "compute_graph.l2l3_6_spill" + ] + }, + "buffer_iter": 9, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[1].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[1].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[1].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[1].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[1].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[1].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[1].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[1].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[1].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[1].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[1].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[1].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[1].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[1].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[1].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[1].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2560 + ], + "l1_pong": [ + "0x4000", + 2560 + ], + "l2": [ + [ + 2, + 0, + 114688, + 51200, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_5_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[1].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[1].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[1].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[1].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[1].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[1].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[1].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[1].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[1].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[1].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[1].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[1].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[1].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[1].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[1].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[1].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5400", + 2560 + ], + "l1_pong": [ + "0x1400", + 2560 + ], + "l2": [ + [ + 2, + 0, + 319488, + 51200, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1" + ], + "l3": { + "const_ifm_ddr_5": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.const_ifm_ddr_5" + ] + }, + "kernel_name": [ + "superkernel_add1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_6.out[0]", + "to": "compute_graph.l2l3_6_spill.in[0]" + }, + { + "from": "compute_graph.l2_6.out[1]", + "to": "compute_graph.l2l3_6_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.const_ifm_ddr_5.out[0]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.in[0]" + }, + { + "from": "compute_graph.const_ifm_ddr_5.out[1]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input5_0_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_5_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_5_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_6_port0.in[1]" + } + ], + "layer_name": "Sub_14", + "layer_object_name": "compute_graph.flexml_layers[1]", + "layer_order": 6, + "mlir_dag_id_map": { + "dag_layer": 6, + "mlir_layer": 6 + }, + "num_iter": 180, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[1].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[1].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[1].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[1].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[1].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[1].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[1].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[1].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[1].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[1].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[1].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[1].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[1].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[1].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[1].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[1].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xad80", + 640 + ], + "l1_pong": [ + "0xcd80", + 640 + ], + "l2": [ + [ + 0, + 0, + 0, + 204800, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_6" + ], + "l3": { + "l2l3_6_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_6_spill" + ] + }, + "super_iter": 9 + }, + "0007": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1", + "compute_graph.const_ifm_ddr_4", + "compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0", + "compute_graph.l2l3_6_spill" + ], + "out": [ + "compute_graph.l2_7", + "compute_graph.l2l3_7_spill" + ] + }, + "buffer_iter": 9, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[2].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[2].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[2].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[2].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[2].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[2].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[2].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[2].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[2].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[2].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[2].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[2].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[2].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[2].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[2].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[2].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2560 + ], + "l1_pong": [ + "0x4000", + 2560 + ], + "l2": [ + [ + 2, + 0, + 114688, + 51200, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_6_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[2].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[2].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[2].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[2].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[2].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[2].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[2].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[2].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[2].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[2].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[2].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[2].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[2].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[2].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[2].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[2].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5400", + 2560 + ], + "l1_pong": [ + "0x1400", + 2560 + ], + "l2": [ + [ + 2, + 0, + 319488, + 51200, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1" + ], + "l3": { + "const_ifm_ddr_4": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.const_ifm_ddr_4" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_7.out[0]", + "to": "compute_graph.l2l3_7_spill.in[0]" + }, + { + "from": "compute_graph.l2_7.out[1]", + "to": "compute_graph.l2l3_7_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.const_ifm_ddr_4.out[0]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.in[0]" + }, + { + "from": "compute_graph.const_ifm_ddr_4.out[1]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input4_0_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_6_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_6_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_6_for_layer_7_port0.in[1]" + } + ], + "layer_name": "Div_16", + "layer_object_name": "compute_graph.flexml_layers[2]", + "layer_order": 7, + "mlir_dag_id_map": { + "dag_layer": 7, + "mlir_layer": 7 + }, + "num_iter": 180, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[2].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[2].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[2].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[2].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[2].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[2].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[2].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[2].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[2].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[2].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[2].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[2].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[2].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[2].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[2].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[2].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xad80", + 640 + ], + "l1_pong": [ + "0xcd80", + 640 + ], + "l2": [ + [ + 0, + 0, + 0, + 204800, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_7" + ], + "l3": { + "l2l3_7_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_7_spill" + ] + }, + "super_iter": 9 + }, + "0008": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0", + "compute_graph.l2l3_7_spill" + ], + "out": [ + "compute_graph.l2_8", + "compute_graph.l2l3_8_spill" + ], + "wts": [ + "compute_graph.Layer_8_l2_wts", + "compute_graph.Layer_8_wts_ddr" + ] + }, + "buffer_iter": 3, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[3].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[3].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[3].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[3].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[3].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[3].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[3].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[3].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[3].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[3].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[3].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[3].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[3].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[3].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[3].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[3].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5280 + ], + "l1_pong": [ + "0xcd80", + 5280 + ], + "l2": [ + [ + 1, + 0, + 423936, + 156160, + 156160 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_7_spill" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_8.out[0]", + "to": "compute_graph.l2l3_8_spill.in[0]" + }, + { + "from": "compute_graph.l2_8.out[1]", + "to": "compute_graph.l2l3_8_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_7_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_7_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_7_for_layer_8_port0.in[1]" + } + ], + "layer_name": "Conv_17", + "layer_object_name": "compute_graph.flexml_layers[3]", + "layer_order": 8, + "mlir_dag_id_map": { + "dag_layer": 8, + "mlir_layer": 8 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[3].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[3].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[3].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[3].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[3].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[3].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[3].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[3].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[3].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[3].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[3].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[3].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[3].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[3].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[3].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[3].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_8" + ], + "l3": { + "l2l3_8_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_8_spill" + ] + }, + "super_iter": 3, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[3].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[3].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[3].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[3].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[3].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[3].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[3].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[3].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[3].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[3].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[3].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[3].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[3].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[3].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[3].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[3].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf6c0", + 608 + ], + "l1_pong": [ + "0xa940", + 608 + ], + "l2": [ + [ + 3, + 0, + 0, + 2432 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_8_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 0, + "size": 2432 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_8_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0009": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0", + "compute_graph.l2l3_8_spill" + ], + "out": [ + "compute_graph.l2_9", + "compute_graph.l2l3_9_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[4].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[4].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[4].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[4].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[4].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[4].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[4].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[4].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[4].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[4].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[4].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[4].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[4].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[4].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[4].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[4].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 5888 + ], + "l1_pong": [ + "0x4000", + 5888 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_8_spill" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_9.out[0]", + "to": "compute_graph.l2l3_9_spill.in[0]" + }, + { + "from": "compute_graph.l2_9.out[1]", + "to": "compute_graph.l2l3_9_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_8_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_8_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_9_port0.in[1]" + } + ], + "layer_name": "Add_19", + "layer_object_name": "compute_graph.flexml_layers[4]", + "layer_order": 9, + "mlir_dag_id_map": { + "dag_layer": 9, + "mlir_layer": 9 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[4].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[4].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[4].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[4].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[4].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[4].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[4].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[4].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[4].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[4].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[4].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[4].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[4].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[4].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[4].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[4].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa700", + 1472 + ], + "l1_pong": [ + "0xcd80", + 1472 + ], + "l2": [ + [ + 0, + 0, + 460800, + 471040, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_9" + ], + "l3": { + "l2l3_9_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_9_spill" + ] + }, + "super_iter": 1 + }, + "0010": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0", + "compute_graph.l2l3_9_spill" + ], + "out": [ + "compute_graph.l2_10", + "compute_graph.l2l3_10_spill" + ] + }, + "buffer_iter": 2, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[5].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[5].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[5].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[5].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[5].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[5].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[5].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[5].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[5].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[5].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[5].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[5].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[5].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[5].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[5].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[5].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 293888, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_9_spill" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_10.out[0]", + "to": "compute_graph.l2l3_10_spill.in[0]" + }, + { + "from": "compute_graph.l2_10.out[1]", + "to": "compute_graph.l2l3_10_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_9_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_9_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_9_for_layer_10_port0.in[1]" + } + ], + "layer_name": "Clip_22", + "layer_object_name": "compute_graph.flexml_layers[5]", + "layer_order": 10, + "mlir_dag_id_map": { + "dag_layer": 10, + "mlir_layer": 10 + }, + "num_iter": 16, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[5].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[5].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[5].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[5].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[5].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[5].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[5].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[5].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[5].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[5].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[5].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[5].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[5].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[5].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[5].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[5].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 245760, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_10" + ], + "l2_force_single_buffering": true, + "l3": { + "l2l3_10_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_10_spill" + ] + }, + "super_iter": 2 + }, + "0011": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0", + "compute_graph.l2l3_10_spill" + ], + "out": [ + "compute_graph.l2_11", + "compute_graph.l2l3_11_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[6].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[6].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[6].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[6].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[6].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[6].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[6].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[6].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[6].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[6].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[6].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[6].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[6].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[6].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[6].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[6].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 5888 + ], + "l1_pong": [ + "0x4000", + 5888 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_10_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_11.out[0]", + "to": "compute_graph.l2l3_11_spill.in[0]" + }, + { + "from": "compute_graph.l2_11.out[1]", + "to": "compute_graph.l2l3_11_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_10_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_10_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_10_for_layer_11_port0.in[1]" + } + ], + "layer_name": "Div_24", + "layer_object_name": "compute_graph.flexml_layers[6]", + "layer_order": 11, + "mlir_dag_id_map": { + "dag_layer": 11, + "mlir_layer": 11 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[6].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[6].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[6].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[6].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[6].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[6].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[6].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[6].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[6].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[6].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[6].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[6].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[6].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[6].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[6].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[6].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa700", + 1472 + ], + "l1_pong": [ + "0xcd80", + 1472 + ], + "l2": [ + [ + 0, + 0, + 460800, + 471040, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_11" + ], + "l3": { + "l2l3_11_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_11_spill" + ] + }, + "super_iter": 1 + }, + "0012": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0", + "compute_graph.l2l3_8_spill", + "compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1", + "compute_graph.l2l3_11_spill" + ], + "out": [ + "compute_graph.l2_12", + "compute_graph.l2l3_12_spill" + ] + }, + "buffer_iter": 3, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[7].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[7].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[7].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[7].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[7].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[7].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[7].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[7].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[7].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[7].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[7].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[7].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[7].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[7].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[7].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[7].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 4096 + ], + "l1_pong": [ + "0x4000", + 4096 + ], + "l2": [ + [ + 2, + 0, + 217088, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_8_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[7].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[7].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[7].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[7].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[7].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[7].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[7].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[7].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[7].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[7].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[7].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[7].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[7].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[7].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[7].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[7].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x6000", + 4096 + ], + "l1_pong": [ + "0x2000", + 4096 + ], + "l2": [ + [ + 1, + 0, + 434176, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_11_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_12.out[0]", + "to": "compute_graph.l2l3_12_spill.in[0]" + }, + { + "from": "compute_graph.l2_12.out[1]", + "to": "compute_graph.l2l3_12_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_11_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_11_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_11_for_layer_12_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_8_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_8_spill.out[3]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_8_for_layer_12_port0.in[1]" + } + ], + "layer_name": "Mul_25", + "layer_object_name": "compute_graph.flexml_layers[7]", + "layer_order": 12, + "mlir_dag_id_map": { + "dag_layer": 12, + "mlir_layer": 12 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[7].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[7].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[7].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[7].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[7].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[7].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[7].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[7].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[7].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[7].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[7].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[7].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[7].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[7].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[7].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[7].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaa80", + 1024 + ], + "l1_pong": [ + "0xcd80", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_12" + ], + "l3": { + "l2l3_12_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_12_spill" + ] + }, + "super_iter": 3 + }, + "0013": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0", + "compute_graph.l2l3_12_spill" + ], + "out": [ + "compute_graph.l2_13", + "compute_graph.l2l3_13_spill" + ], + "wts": [ + "compute_graph.Layer_13_l2_wts", + "compute_graph.Layer_13_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[8].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[8].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[8].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[8].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[8].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[8].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[8].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[8].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[8].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[8].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[8].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[8].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[8].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[8].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[8].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[8].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5120 + ], + "l1_pong": [ + "0xcd80", + 5120 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_12_spill" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_13.out[0]", + "to": "compute_graph.l2l3_13_spill.in[0]" + }, + { + "from": "compute_graph.l2_13.out[1]", + "to": "compute_graph.l2l3_13_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_12_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_12_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_13_port0.in[1]" + } + ], + "layer_name": "Conv_26", + "layer_object_name": "compute_graph.flexml_layers[8]", + "layer_order": 13, + "mlir_dag_id_map": { + "dag_layer": 13, + "mlir_layer": 13 + }, + "num_iter": 40, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[8].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[8].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[8].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[8].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[8].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[8].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[8].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[8].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[8].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[8].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[8].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[8].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[8].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[8].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[8].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[8].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 460800, + 491520, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_13" + ], + "l3": { + "l2l3_13_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_13_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[8].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[8].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[8].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[8].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[8].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[8].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[8].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[8].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[8].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[8].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[8].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[8].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[8].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[8].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[8].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[8].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf580", + 128 + ], + "l1_pong": [ + "0xa800", + 128 + ], + "l2": [ + [ + 3, + 0, + 523264, + 512 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_13_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 4864, + "size": 512 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_13_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0014": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1", + "compute_graph.l2l3_12_spill", + "compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0", + "compute_graph.l2l3_13_spill" + ], + "out": [ + "compute_graph.l2_14", + "compute_graph.l2l3_14_spill", + "compute_graph.spill_L3_Concat_Buffer_layer_275" + ], + "wts": [ + "compute_graph.Layer_14_l2_wts", + "compute_graph.Layer_14_wts_ddr" + ] + }, + "buffer_iter": 2, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[9].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[9].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[9].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[9].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[9].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[9].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[9].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[9].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[9].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[9].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[9].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[9].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[9].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[9].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[9].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[9].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4096 + ], + "l1_pong": [ + "0xcd80", + 4096 + ], + "l2": [ + [ + 0, + 0, + 491520, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_13_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[9].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[9].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[9].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[9].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[9].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[9].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[9].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[9].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[9].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[9].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[9].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[9].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[9].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[9].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[9].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[9].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 2048 + ], + "l1_pong": [ + "0x6000", + 2048 + ], + "l2": [ + [ + 0, + 0, + 0, + 245760, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_12_spill" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_14.out[0]", + "to": "compute_graph.l2l3_14_spill.in[0]" + }, + { + "from": "compute_graph.l2_14.out[1]", + "to": "compute_graph.l2l3_14_spill.in[1]" + }, + { + "from": "compute_graph.l2_14.out[2]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_275.in[2]" + }, + { + "from": "compute_graph.l2_14.out[3]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_275.in[3]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_12_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_12_spill.out[3]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_12_for_layer_14_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_13_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_13_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_13_for_layer_14_port0.in[1]" + } + ], + "layer_name": "Conv_28", + "layer_object_name": "compute_graph.flexml_layers[9]", + "layer_order": 14, + "mlir_dag_id_map": { + "dag_layer": 14, + "mlir_layer": 14 + }, + "num_iter": 60, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[9].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[9].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[9].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[9].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[9].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[9].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[9].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[9].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[9].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[9].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[9].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[9].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[9].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[9].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[9].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[9].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 512 + ], + "l1_pong": [ + "0x4000", + 512 + ], + "l2": [ + [ + 2, + 0, + 32768, + 245760, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_14" + ], + "l2_force_single_buffering": true, + "l3": { + "l2l3_14_spill": { + "size": 230400 + }, + "spill_L3_Concat_Buffer_layer_275": { + "buffer_bounds": [ + 1, + 16, + 45, + 160 + ], + "concat_offset": [ + 0, + 40, + 0, + 0 + ], + "size": 921600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_14_spill", + "compute_graph.spill_L3_Concat_Buffer_layer_275" + ] + }, + "super_iter": 2, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[9].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[9].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[9].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[9].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[9].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[9].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[9].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[9].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[9].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[9].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[9].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[9].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[9].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[9].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[9].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[9].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xed80", + 544 + ], + "l1_pong": [ + "0xa000", + 544 + ], + "l2": [ + [ + 3, + 0, + 0, + 2176 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_14_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 5888, + "size": 2176 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_14_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0015": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0", + "compute_graph.l2l3_14_spill" + ], + "out": [ + "compute_graph.l2_15", + "compute_graph.l2l3_15_spill" + ], + "wts": [ + "compute_graph.Layer_15_l2_wts", + "compute_graph.Layer_15_wts_ddr" + ] + }, + "buffer_iter": 5, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[10].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[10].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[10].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[10].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[10].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[10].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[10].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[10].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[10].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[10].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[10].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[10].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[10].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[10].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[10].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[10].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4608 + ], + "l1_pong": [ + "0xcd80", + 4608 + ], + "l2": [ + [ + 2, + 0, + 339968, + 46080, + 46080 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_14_spill" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_15.out[0]", + "to": "compute_graph.l2l3_15_spill.in[0]" + }, + { + "from": "compute_graph.l2_15.out[1]", + "to": "compute_graph.l2l3_15_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_14_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_14_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_14_for_layer_15_port0.in[1]" + } + ], + "layer_name": "Conv_30", + "layer_object_name": "compute_graph.flexml_layers[10]", + "layer_order": 15, + "mlir_dag_id_map": { + "dag_layer": 15, + "mlir_layer": 15 + }, + "num_iter": 50, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[10].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[10].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[10].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[10].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[10].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[10].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[10].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[10].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[10].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[10].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[10].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[10].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[10].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[10].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[10].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[10].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1152 + ], + "l1_pong": [ + "0x4000", + 1152 + ], + "l2": [ + [ + 0, + 0, + 0, + 184320, + 184320 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_15" + ], + "l3": { + "l2l3_15_spill": { + "size": 921600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_15_spill" + ] + }, + "super_iter": 5, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[10].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[10].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[10].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[10].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[10].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[10].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[10].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[10].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[10].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[10].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[10].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[10].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[10].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[10].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[10].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[10].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf180", + 1088 + ], + "l1_pong": [ + "0xa400", + 1088 + ], + "l2": [ + [ + 3, + 0, + 515584, + 4352 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_15_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 10240, + "size": 4352 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_15_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0016": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0", + "compute_graph.l2l3_15_spill" + ], + "out": [ + "compute_graph.l2_16", + "compute_graph.l2l3_16_spill" + ], + "wts": [ + "compute_graph.Layer_16_l2_wts", + "compute_graph.Layer_16_wts_ddr" + ] + }, + "buffer_iter": 6, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[11].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[11].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[11].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[11].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[11].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[11].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[11].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[11].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[11].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[11].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[11].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[11].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[11].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[11].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[11].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[11].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5376 + ], + "l1_pong": [ + "0xcd80", + 5376 + ], + "l2": [ + [ + 0, + 0, + 0, + 174080, + 174080 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_15_spill" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_16.out[0]", + "to": "compute_graph.l2l3_16_spill.in[0]" + }, + { + "from": "compute_graph.l2_16.out[1]", + "to": "compute_graph.l2l3_16_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_15_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_15_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_15_for_layer_16_port0.in[1]" + } + ], + "layer_name": "Conv_32", + "layer_object_name": "compute_graph.flexml_layers[11]", + "layer_order": 16, + "mlir_dag_id_map": { + "dag_layer": 16, + "mlir_layer": 16 + }, + "num_iter": 84, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[11].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[11].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[11].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[11].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[11].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[11].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[11].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[11].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[11].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[11].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[11].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[11].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[11].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[11].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[11].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[11].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 192 + ], + "l1_pong": [ + "0x4000", + 192 + ], + "l2": [ + [ + 2, + 0, + 352256, + 43008, + 40960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_16" + ], + "l3": { + "l2l3_16_spill": { + "size": 245760 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_16_spill" + ] + }, + "super_iter": 6, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[11].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[11].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[11].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[11].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[11].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[11].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[11].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[11].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[11].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[11].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[11].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[11].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[11].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[11].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[11].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[11].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf780", + 128 + ], + "l1_pong": [ + "0xaa00", + 128 + ], + "l2": [ + [ + 3, + 0, + 0, + 1024 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_16_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 18944, + "size": 1024 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_16_wts_ddr" + ] + }, + "wts_iter": 6 + }, + "0017": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_16_spill", + "compute_graph.l2l3_scratch_0_17_spill" + ], + "out": [ + "compute_graph.l2l3_17_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_16_spill": { + "size": 245760 + }, + "l2l3_scratch_0_17_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_16_spill", + "compute_graph.l2l3_scratch_0_17_spill" + ] + }, + "kernel_name": [ + "BufferUnpadAdf" + ], + "layer_name": "Generated-#60", + "layer_object_name": "compute_graph.templated_graph_17", + "layer_order": 17, + "mlir_dag_id_map": { + "dag_layer": 17, + "mlir_layer": 17 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_17_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_17_spill" + ] + }, + "templated_graph": 1 + }, + "0018": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0", + "compute_graph.l2l3_17_spill" + ], + "out": [ + "compute_graph.l2_18", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818" + ], + "wts": [ + "compute_graph.Layer_18_l2_wts", + "compute_graph.Layer_18_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[12].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[12].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[12].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[12].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[12].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[12].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[12].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[12].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[12].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[12].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[12].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[12].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[12].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[12].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[12].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[12].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4096 + ], + "l1_pong": [ + "0xcd80", + 4096 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_17_spill" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_18.out[4]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.in[0]" + }, + { + "from": "compute_graph.l2_18.out[5]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_17_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_17_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_17_for_layer_18_port0.in[1]" + } + ], + "layer_name": "Conv_34", + "layer_object_name": "compute_graph.flexml_layers[12]", + "layer_order": 18, + "mlir_dag_id_map": { + "dag_layer": 18, + "mlir_layer": 18 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[12].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[12].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[12].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[12].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[12].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[12].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[12].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[12].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[12].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[12].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[12].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[12].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[12].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[12].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[12].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[12].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 1, + 0, + 522240, + 245760, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_18" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_1818": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[12].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[12].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[12].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[12].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[12].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[12].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[12].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[12].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[12].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[12].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[12].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[12].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[12].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[12].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[12].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[12].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xed80", + 1088 + ], + "l1_pong": [ + "0xa000", + 1088 + ], + "l2": [ + [ + 3, + 0, + 515584, + 4352 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_18_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 20992, + "size": 4352 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_18_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0019": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_18" + ], + "out": [ + "compute_graph.l2_19", + "compute_graph.l2l3_19_spill" + ], + "wts": [ + "compute_graph.Layer_19_l2_wts", + "compute_graph.Layer_19_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[13].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[13].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[13].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[13].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[13].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[13].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[13].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[13].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[13].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[13].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[13].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[13].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[13].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[13].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[13].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[13].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4096 + ], + "l1_pong": [ + "0xcd80", + 4096 + ], + "l2": [ + [ + 1, + 0, + 522240, + 245760, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_18" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_18", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_19.out[0]", + "to": "compute_graph.l2l3_19_spill.in[0]" + }, + { + "from": "compute_graph.l2_19.out[1]", + "to": "compute_graph.l2l3_19_spill.in[1]" + } + ], + "layer_name": "Conv_35", + "layer_object_name": "compute_graph.flexml_layers[13]", + "layer_order": 19, + "mlir_dag_id_map": { + "dag_layer": 19, + "mlir_layer": 19 + }, + "num_iter": 18, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[13].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[13].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[13].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[13].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[13].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[13].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[13].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[13].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[13].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[13].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[13].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[13].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[13].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[13].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[13].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[13].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 0, + 0, + 0, + 442368, + 259200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_19" + ], + "l3": { + "l2l3_19_spill": { + "size": 259200 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_19_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[13].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[13].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[13].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[13].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[13].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[13].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[13].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[13].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[13].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[13].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[13].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[13].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[13].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[13].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[13].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[13].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xed80", + 1632 + ], + "l1_pong": [ + "0xa000", + 1632 + ], + "l2": [ + [ + 3, + 0, + 0, + 6528 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_19_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 29696, + "size": 6528 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_19_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0020": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0", + "compute_graph.l2l3_19_spill" + ], + "out": [ + "compute_graph.l2_20", + "compute_graph.l2l3_20_spill" + ], + "wts": [ + "compute_graph.Layer_20_l2_wts", + "compute_graph.Layer_20_wts_ddr" + ] + }, + "buffer_iter": 2, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[14].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[14].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[14].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[14].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[14].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[14].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[14].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[14].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[14].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[14].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[14].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[14].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[14].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[14].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[14].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[14].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5120 + ], + "l1_pong": [ + "0xcd80", + 5120 + ], + "l2": [ + [ + 2, + 0, + 236288, + 144000, + 144000 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_19_spill" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_20.out[0]", + "to": "compute_graph.l2l3_20_spill.in[0]" + }, + { + "from": "compute_graph.l2_20.out[1]", + "to": "compute_graph.l2l3_20_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_19_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_19_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_19_for_layer_20_port0.in[1]" + } + ], + "layer_name": "Conv_37", + "layer_object_name": "compute_graph.flexml_layers[14]", + "layer_order": 20, + "mlir_dag_id_map": { + "dag_layer": 20, + "mlir_layer": 20 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[14].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[14].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[14].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[14].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[14].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[14].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[14].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[14].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[14].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[14].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[14].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[14].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[14].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[14].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[14].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[14].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 0, + 184320, + 132480 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_20" + ], + "l2_force_single_buffering": true, + "l3": { + "l2l3_20_spill": { + "size": 264960 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_20_spill" + ] + }, + "super_iter": 2, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[14].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[14].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[14].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[14].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[14].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[14].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[14].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[14].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[14].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[14].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[14].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[14].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[14].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[14].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[14].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[14].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf580", + 128 + ], + "l1_pong": [ + "0xa800", + 128 + ], + "l2": [ + [ + 3, + 0, + 521216, + 1536 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_20_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 42752, + "size": 1536 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_20_wts_ddr" + ] + }, + "wts_iter": 2 + }, + "0021": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_20_spill", + "compute_graph.l2l3_scratch_0_21_spill" + ], + "out": [ + "compute_graph.l2l3_21_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_20_spill": { + "size": 264960 + }, + "l2l3_scratch_0_21_spill": { + "size": 518400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_20_spill", + "compute_graph.l2l3_scratch_0_21_spill" + ] + }, + "kernel_name": [ + "BufferUnpadAdf" + ], + "layer_name": "Generated-#62", + "layer_object_name": "compute_graph.templated_graph_21", + "layer_order": 21, + "mlir_dag_id_map": { + "dag_layer": 21, + "mlir_layer": 21 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_21_spill": { + "size": 259200 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_21_spill" + ] + }, + "templated_graph": 1 + }, + "0022": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818", + "compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0", + "compute_graph.l2l3_21_spill" + ], + "out": [ + "compute_graph.l2_22", + "compute_graph.spill_L3_Concat_Buffer_layer_256" + ], + "wts": [ + "compute_graph.Layer_22_l2_wts", + "compute_graph.Layer_22_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[15].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[15].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[15].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[15].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[15].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[15].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[15].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[15].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[15].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[15].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[15].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[15].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[15].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[15].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[15].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[15].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4608 + ], + "l1_pong": [ + "0xcd80", + 4608 + ], + "l2": [ + [ + 0, + 0, + 491520, + 259200, + 259200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_21_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[15].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[15].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[15].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[15].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[15].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[15].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[15].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[15].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[15].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[15].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[15].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[15].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[15].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[15].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[15].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[15].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 4096 + ], + "l1_pong": [ + "0x6000", + 4096 + ], + "l2": [ + [ + 0, + 0, + 0, + 245760, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_22.out[4]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_256.in[2]" + }, + { + "from": "compute_graph.l2_22.out[5]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_256.in[3]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_1818.in[1]" + }, + { + "from": "compute_graph.l2l3_21_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_21_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_21_for_layer_22_port0.in[1]" + } + ], + "layer_name": "Conv_39", + "layer_object_name": "compute_graph.flexml_layers[15]", + "layer_order": 22, + "mlir_dag_id_map": { + "dag_layer": 22, + "mlir_layer": 22 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[15].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[15].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[15].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[15].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[15].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[15].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[15].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[15].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[15].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[15].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[15].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[15].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[15].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[15].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[15].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[15].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 245760, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_22" + ], + "l3": { + "spill_L3_Concat_Buffer_layer_256": { + "buffer_bounds": [ + 1, + 24, + 45, + 80 + ], + "concat_offset": [ + 0, + 80, + 0, + 0 + ], + "size": 403200 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_256" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[15].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[15].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[15].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[15].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[15].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[15].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[15].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[15].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[15].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[15].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[15].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[15].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[15].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[15].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[15].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[15].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf180", + 1216 + ], + "l1_pong": [ + "0xa400", + 1216 + ], + "l2": [ + [ + 3, + 0, + 0, + 4864 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_22_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 45824, + "size": 4864 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_22_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0023": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_22" + ], + "out": [ + "compute_graph.l2_23" + ], + "wts": [ + "compute_graph.Layer_23_l2_wts", + "compute_graph.Layer_23_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[16].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[16].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[16].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[16].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[16].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[16].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[16].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[16].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[16].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[16].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[16].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[16].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[16].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[16].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[16].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[16].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4096 + ], + "l1_pong": [ + "0xcd80", + 4096 + ], + "l2": [ + [ + 0, + 0, + 0, + 245760, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_22" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_22", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_41", + "layer_object_name": "compute_graph.flexml_layers[16]", + "layer_order": 23, + "mlir_dag_id_map": { + "dag_layer": 23, + "mlir_layer": 23 + }, + "num_iter": 18, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[16].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[16].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[16].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[16].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[16].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[16].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[16].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[16].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[16].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[16].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[16].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[16].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[16].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[16].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[16].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[16].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 1, + 0, + 163840, + 442368, + 259200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_23" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[16].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[16].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[16].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[16].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[16].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[16].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[16].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[16].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[16].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[16].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[16].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[16].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[16].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[16].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[16].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[16].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xed80", + 1632 + ], + "l1_pong": [ + "0xa000", + 1632 + ], + "l2": [ + [ + 3, + 0, + 511232, + 6528 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_23_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 55552, + "size": 6528 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_23_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0024": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_23" + ], + "out": [ + "compute_graph.l2_24", + "compute_graph.l2l3_24_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424" + ], + "wts": [ + "compute_graph.Layer_24_l2_wts", + "compute_graph.Layer_24_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[17].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[17].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[17].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[17].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[17].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[17].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[17].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[17].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[17].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[17].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[17].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[17].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[17].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[17].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[17].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[17].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5120 + ], + "l1_pong": [ + "0xcd80", + 5120 + ], + "l2": [ + [ + 1, + 0, + 163840, + 442368, + 259200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_23" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_23", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_24.out[0]", + "to": "compute_graph.l2l3_24_spill.in[0]" + }, + { + "from": "compute_graph.l2_24.out[1]", + "to": "compute_graph.l2l3_24_spill.in[1]" + }, + { + "from": "compute_graph.l2_24.out[2]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.in[0]" + }, + { + "from": "compute_graph.l2_24.out[3]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.in[1]" + } + ], + "layer_name": "Conv_43", + "layer_object_name": "compute_graph.flexml_layers[17]", + "layer_order": 24, + "mlir_dag_id_map": { + "dag_layer": 24, + "mlir_layer": 24 + }, + "num_iter": 45, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[17].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[17].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[17].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[17].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[17].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[17].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[17].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[17].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[17].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[17].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[17].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[17].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[17].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[17].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[17].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[17].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 128 + ], + "l1_pong": [ + "0x4000", + 128 + ], + "l2": [ + [ + 0, + 0, + 0, + 92160, + 66240 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_24" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_2424": { + "size": 66240 + }, + "l2l3_24_spill": { + "size": 66240 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_24_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[17].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[17].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[17].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[17].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[17].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[17].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[17].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[17].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[17].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[17].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[17].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[17].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[17].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[17].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[17].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[17].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf580", + 352 + ], + "l1_pong": [ + "0xa800", + 352 + ], + "l2": [ + [ + 3, + 0, + 0, + 4224 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_24_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 68608, + "size": 4224 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_24_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0025": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_24_spill" + ], + "out": [ + "compute_graph.l2l3_25_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_24_spill": { + "size": 66240 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_24_spill" + ] + }, + "kernel_name": [ + "Transpose4dAdf" + ], + "layer_name": "Generated-#6", + "layer_object_name": "compute_graph.templated_graph_25", + "layer_order": 25, + "mlir_dag_id_map": { + "dag_layer": 25, + "mlir_layer": 25 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_25_spill": { + "size": 66240 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_25_spill" + ] + }, + "templated_graph": 1 + }, + "0026": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0", + "compute_graph.l2l3_25_spill" + ], + "out": [ + "compute_graph.l2_26" + ] + }, + "buffer_iter": 1, + "depth_iter": 15, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[18].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[18].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[18].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[18].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[18].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[18].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[18].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[18].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[18].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[18].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[18].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[18].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[18].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[18].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[18].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[18].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 8192 + ], + "l1_pong": [ + "0x4000", + 8192 + ], + "l2": [ + [ + 0, + 0, + 0, + 66240, + 66240 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_25_spill" + ] + }, + "kernel_name": [ + "superkernel_reduce_mean_c8" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_25_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_25_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_25_for_layer_26_port0.in[1]" + } + ], + "layer_name": "Generated-#8", + "layer_object_name": "compute_graph.flexml_layers[18]", + "layer_order": 26, + "mlir_dag_id_map": { + "dag_layer": 26, + "mlir_layer": 26 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[18].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[18].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[18].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[18].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[18].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[18].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[18].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[18].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[18].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[18].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[18].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[18].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[18].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[18].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[18].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[18].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 32 + ], + "l1_pong": [ + "0xcd80", + 32 + ], + "l2": [ + [ + 2, + 0, + 523264, + 512, + 72 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_26" + ] + }, + "super_iter": 1 + }, + "0027": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_26" + ], + "out": [ + "compute_graph.l2_27" + ], + "wts": [ + "compute_graph.Layer_27_l2_wts", + "compute_graph.Layer_27_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[19].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[19].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[19].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[19].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[19].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[19].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[19].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[19].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[19].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[19].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[19].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[19].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[19].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[19].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[19].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[19].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 1152 + ], + "l1_pong": [ + "0xcd80", + 1152 + ], + "l2": [ + [ + 2, + 0, + 523264, + 512, + 72 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_26" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_26", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_46", + "layer_object_name": "compute_graph.flexml_layers[19]", + "layer_order": 27, + "mlir_dag_id_map": { + "dag_layer": 27, + "mlir_layer": 27 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[19].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[19].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[19].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[19].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[19].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[19].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[19].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[19].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[19].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[19].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[19].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[19].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[19].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[19].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[19].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[19].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 0, + 0, + 0, + 4096, + 24 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_27" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[19].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[19].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[19].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[19].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[19].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[19].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[19].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[19].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[19].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[19].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[19].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[19].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[19].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[19].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[19].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[19].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xd680", + 1216 + ], + "l1_pong": [ + "0x8900", + 1216 + ], + "l2": [ + [ + 3, + 0, + 0, + 4864 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_27_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 77056, + "size": 4864 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_27_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0028": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_27" + ], + "out": [ + "compute_graph.l2_28" + ], + "wts": [ + "compute_graph.Layer_28_l2_wts", + "compute_graph.Layer_28_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[20].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[20].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[20].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[20].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[20].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[20].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[20].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[20].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[20].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[20].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[20].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[20].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[20].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[20].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[20].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[20].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 4096, + 24 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_27" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_27", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_48", + "layer_object_name": "compute_graph.flexml_layers[20]", + "layer_order": 28, + "mlir_dag_id_map": { + "dag_layer": 28, + "mlir_layer": 28 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[20].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[20].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[20].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[20].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[20].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[20].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[20].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[20].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[20].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[20].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[20].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[20].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[20].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[20].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[20].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[20].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 72 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_28" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[20].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[20].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[20].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[20].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[20].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[20].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[20].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[20].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[20].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[20].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[20].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[20].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[20].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[20].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[20].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[20].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xd180", + 2176 + ], + "l1_pong": [ + "0x8400", + 2176 + ], + "l2": [ + [ + 3, + 0, + 506880, + 8704 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_28_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 86784, + "size": 8704 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_28_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0029": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_28" + ], + "out": [ + "compute_graph.l2_29" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[21].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[21].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[21].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[21].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[21].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[21].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[21].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[21].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[21].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[21].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[21].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[21].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[21].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[21].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[21].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[21].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 72 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_28" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_28", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_50", + "layer_object_name": "compute_graph.flexml_layers[21]", + "layer_order": 29, + "mlir_dag_id_map": { + "dag_layer": 29, + "mlir_layer": 29 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[21].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[21].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[21].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[21].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[21].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[21].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[21].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[21].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[21].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[21].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[21].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[21].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[21].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[21].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[21].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[21].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaf80", + 384 + ], + "l1_pong": [ + "0xcd80", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 72 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_29" + ] + }, + "super_iter": 1 + }, + "0030": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_29" + ], + "out": [ + "compute_graph.l2_30" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[22].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[22].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[22].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[22].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[22].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[22].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[22].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[22].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[22].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[22].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[22].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[22].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[22].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[22].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[22].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[22].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 72 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_29" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_29", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_53", + "layer_object_name": "compute_graph.flexml_layers[22]", + "layer_order": 30, + "mlir_dag_id_map": { + "dag_layer": 30, + "mlir_layer": 30 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[22].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[22].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[22].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[22].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[22].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[22].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[22].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[22].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[22].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[22].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[22].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[22].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[22].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[22].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[22].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[22].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 256 + ], + "l1_pong": [ + "0xcd80", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 72 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_30" + ] + }, + "super_iter": 1 + }, + "0031": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_30" + ], + "out": [ + "compute_graph.l2_31", + "compute_graph.l2l3_31_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[23].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[23].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[23].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[23].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[23].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[23].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[23].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[23].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[23].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[23].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[23].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[23].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[23].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[23].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[23].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[23].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 72 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_30" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_30", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_31.out[0]", + "to": "compute_graph.l2l3_31_spill.in[0]" + }, + { + "from": "compute_graph.l2_31.out[1]", + "to": "compute_graph.l2l3_31_spill.in[1]" + } + ], + "layer_name": "Div_55", + "layer_object_name": "compute_graph.flexml_layers[23]", + "layer_order": 31, + "mlir_dag_id_map": { + "dag_layer": 31, + "mlir_layer": 31 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[23].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[23].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[23].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[23].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[23].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[23].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[23].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[23].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[23].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[23].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[23].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[23].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[23].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[23].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[23].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[23].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xae80", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 8192, + 72 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_31" + ], + "l3": { + "l2l3_31_spill": { + "size": 72 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_31_spill" + ] + }, + "super_iter": 1 + }, + "0032": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_31_spill", + "compute_graph.l2l3_scratch_0_32_spill", + "compute_graph.l2l3_scratch_1_32_spill" + ], + "out": [ + "compute_graph.l2l3_32_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_31_spill": { + "size": 72 + }, + "l2l3_scratch_0_32_spill": { + "size": 132480 + }, + "l2l3_scratch_1_32_spill": { + "size": 132480 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_31_spill", + "compute_graph.l2l3_scratch_0_32_spill", + "compute_graph.l2l3_scratch_1_32_spill" + ] + }, + "kernel_name": [ + "TileAdf" + ], + "layer_name": "Generated-#10", + "layer_object_name": "compute_graph.templated_graph_32", + "layer_order": 32, + "mlir_dag_id_map": { + "dag_layer": 32, + "mlir_layer": 32 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_32_spill": { + "size": 66240 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_32_spill" + ] + }, + "templated_graph": 1 + }, + "0033": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0", + "compute_graph.l2l3_32_spill", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424" + ], + "out": [ + "compute_graph.l2_33" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[24].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[24].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[24].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[24].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[24].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[24].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[24].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[24].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[24].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[24].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[24].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[24].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[24].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[24].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[24].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[24].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 66240, + 66240 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_32_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[24].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[24].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[24].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[24].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[24].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[24].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[24].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[24].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[24].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[24].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[24].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[24].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[24].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[24].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[24].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[24].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 132480, + 92160, + 66240 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_2424.in[1]" + }, + { + "from": "compute_graph.l2l3_32_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_32_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_32_for_layer_33_port0.in[1]" + } + ], + "layer_name": "Mul_56", + "layer_object_name": "compute_graph.flexml_layers[24]", + "layer_order": 33, + "mlir_dag_id_map": { + "dag_layer": 33, + "mlir_layer": 33 + }, + "num_iter": 6, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[24].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[24].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[24].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[24].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[24].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[24].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[24].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[24].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[24].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[24].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[24].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[24].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[24].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[24].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[24].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[24].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 339968, + 92160, + 66240 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_33" + ] + }, + "super_iter": 1 + }, + "0034": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_33" + ], + "out": [ + "compute_graph.l2_34", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434" + ], + "wts": [ + "compute_graph.Layer_34_l2_wts", + "compute_graph.Layer_34_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[25].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[25].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[25].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[25].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[25].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[25].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[25].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[25].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[25].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[25].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[25].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[25].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[25].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[25].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[25].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[25].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3456 + ], + "l1_pong": [ + "0xcd80", + 3456 + ], + "l2": [ + [ + 2, + 0, + 339968, + 92160, + 66240 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_33" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_33", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_34.out[4]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.in[0]" + }, + { + "from": "compute_graph.l2_34.out[5]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.in[1]" + } + ], + "layer_name": "Conv_57", + "layer_object_name": "compute_graph.flexml_layers[25]", + "layer_order": 34, + "mlir_dag_id_map": { + "dag_layer": 34, + "mlir_layer": 34 + }, + "num_iter": 5, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[25].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[25].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[25].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[25].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[25].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[25].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[25].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[25].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[25].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[25].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[25].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[25].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[25].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[25].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[25].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[25].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_34" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_3434": { + "size": 36800 + } + }, + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[25].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[25].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[25].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[25].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[25].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[25].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[25].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[25].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[25].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[25].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[25].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[25].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[25].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[25].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[25].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[25].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe880", + 1216 + ], + "l1_pong": [ + "0x9b00", + 1216 + ], + "l2": [ + [ + 3, + 0, + 0, + 4864 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_34_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 104192, + "size": 4864 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_34_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0035": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_34" + ], + "out": [ + "compute_graph.l2_35" + ], + "wts": [ + "compute_graph.Layer_35_l2_wts", + "compute_graph.Layer_35_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[26].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[26].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[26].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[26].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[26].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[26].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[26].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[26].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[26].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[26].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[26].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[26].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[26].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[26].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[26].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[26].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4096 + ], + "l1_pong": [ + "0xcd80", + 4096 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_34" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_34", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_58", + "layer_object_name": "compute_graph.flexml_layers[26]", + "layer_order": 35, + "mlir_dag_id_map": { + "dag_layer": 35, + "mlir_layer": 35 + }, + "num_iter": 6, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[26].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[26].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[26].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[26].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[26].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[26].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[26].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[26].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[26].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[26].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[26].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[26].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[26].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[26].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[26].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[26].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 2, + 0, + 131072, + 196608, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_35" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[26].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[26].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[26].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[26].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[26].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[26].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[26].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[26].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[26].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[26].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[26].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[26].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[26].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[26].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[26].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[26].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xed80", + 2176 + ], + "l1_pong": [ + "0xa000", + 2176 + ], + "l2": [ + [ + 3, + 0, + 506880, + 8704 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_35_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 113920, + "size": 8704 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_35_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0036": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_35" + ], + "out": [ + "compute_graph.l2_36", + "compute_graph.l2l3_36_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636" + ], + "wts": [ + "compute_graph.Layer_36_l2_wts", + "compute_graph.Layer_36_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[27].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[27].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[27].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[27].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[27].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[27].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[27].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[27].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[27].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[27].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[27].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[27].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[27].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[27].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[27].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[27].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3840 + ], + "l1_pong": [ + "0xcd80", + 3840 + ], + "l2": [ + [ + 2, + 0, + 131072, + 196608, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_35" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_35", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_36.out[0]", + "to": "compute_graph.l2l3_36_spill.in[0]" + }, + { + "from": "compute_graph.l2_36.out[1]", + "to": "compute_graph.l2l3_36_spill.in[1]" + }, + { + "from": "compute_graph.l2_36.out[2]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.in[0]" + }, + { + "from": "compute_graph.l2_36.out[3]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.in[1]" + } + ], + "layer_name": "Conv_60", + "layer_object_name": "compute_graph.flexml_layers[27]", + "layer_order": 36, + "mlir_dag_id_map": { + "dag_layer": 36, + "mlir_layer": 36 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[27].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[27].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[27].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[27].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[27].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[27].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[27].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[27].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[27].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[27].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[27].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[27].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[27].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[27].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[27].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[27].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 384 + ], + "l1_pong": [ + "0x4000", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_36" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_3636": { + "size": 110400 + }, + "l2l3_36_spill": { + "size": 110400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_36_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[27].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[27].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[27].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[27].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[27].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[27].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[27].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[27].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[27].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[27].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[27].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[27].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[27].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[27].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[27].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[27].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xeb80", + 352 + ], + "l1_pong": [ + "0x9e00", + 352 + ], + "l2": [ + [ + 3, + 0, + 0, + 5632 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_36_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 131328, + "size": 5632 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_36_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0037": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_36_spill" + ], + "out": [ + "compute_graph.l2l3_37_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_36_spill": { + "size": 110400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_36_spill" + ] + }, + "kernel_name": [ + "Transpose4dAdf" + ], + "layer_name": "Generated-#12", + "layer_object_name": "compute_graph.templated_graph_37", + "layer_order": 37, + "mlir_dag_id_map": { + "dag_layer": 37, + "mlir_layer": 37 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_37_spill": { + "size": 110400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_37_spill" + ] + }, + "templated_graph": 1 + }, + "0038": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0", + "compute_graph.l2l3_37_spill" + ], + "out": [ + "compute_graph.l2_38" + ] + }, + "buffer_iter": 1, + "depth_iter": 15, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[28].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[28].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[28].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[28].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[28].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[28].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[28].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[28].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[28].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[28].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[28].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[28].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[28].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[28].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[28].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[28].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 8192 + ], + "l1_pong": [ + "0x4000", + 8192 + ], + "l2": [ + [ + 0, + 0, + 0, + 110400, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_37_spill" + ] + }, + "kernel_name": [ + "superkernel_reduce_mean_c8" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_37_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_37_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_37_for_layer_38_port0.in[1]" + } + ], + "layer_name": "Generated-#14", + "layer_object_name": "compute_graph.flexml_layers[28]", + "layer_order": 38, + "mlir_dag_id_map": { + "dag_layer": 38, + "mlir_layer": 38 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[28].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[28].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[28].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[28].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[28].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[28].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[28].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[28].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[28].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[28].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[28].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[28].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[28].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[28].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[28].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[28].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 32 + ], + "l1_pong": [ + "0xcd80", + 32 + ], + "l2": [ + [ + 2, + 0, + 523264, + 512, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_38" + ] + }, + "super_iter": 1 + }, + "0039": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_38" + ], + "out": [ + "compute_graph.l2_39" + ], + "wts": [ + "compute_graph.Layer_39_l2_wts", + "compute_graph.Layer_39_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[29].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[29].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[29].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[29].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[29].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[29].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[29].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[29].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[29].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[29].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[29].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[29].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[29].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[29].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[29].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[29].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 2, + 0, + 523264, + 512, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_38" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_38", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_63", + "layer_object_name": "compute_graph.flexml_layers[29]", + "layer_order": 39, + "mlir_dag_id_map": { + "dag_layer": 39, + "mlir_layer": 39 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[29].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[29].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[29].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[29].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[29].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[29].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[29].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[29].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[29].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[29].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[29].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[29].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[29].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[29].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[29].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[29].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 0, + 0, + 0, + 4096, + 32 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_39" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[29].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[29].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[29].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[29].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[29].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[29].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[29].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[29].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[29].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[29].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[29].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[29].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[29].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[29].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[29].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[29].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xdc80", + 1984 + ], + "l1_pong": [ + "0x8f00", + 1984 + ], + "l2": [ + [ + 3, + 0, + 0, + 7936 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_39_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 142592, + "size": 7936 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_39_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0040": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_39" + ], + "out": [ + "compute_graph.l2_40" + ], + "wts": [ + "compute_graph.Layer_40_l2_wts", + "compute_graph.Layer_40_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[30].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[30].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[30].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[30].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[30].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[30].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[30].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[30].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[30].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[30].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[30].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[30].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[30].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[30].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[30].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[30].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 4096, + 32 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_39" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_39", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_65", + "layer_object_name": "compute_graph.flexml_layers[30]", + "layer_order": 40, + "mlir_dag_id_map": { + "dag_layer": 40, + "mlir_layer": 40 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[30].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[30].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[30].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[30].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[30].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[30].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[30].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[30].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[30].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[30].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[30].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[30].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[30].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[30].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[30].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[30].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_40" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[30].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[30].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[30].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[30].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[30].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[30].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[30].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[30].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[30].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[30].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[30].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[30].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[30].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[30].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[30].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[30].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xd180", + 2176 + ], + "l1_pong": [ + "0x8400", + 2176 + ], + "l2": [ + [ + 3, + 0, + 506880, + 8704 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_40_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 158464, + "size": 8704 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_40_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0041": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_40" + ], + "out": [ + "compute_graph.l2_41" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[31].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[31].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[31].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[31].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[31].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[31].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[31].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[31].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[31].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[31].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[31].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[31].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[31].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[31].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[31].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[31].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_40" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_40", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_67", + "layer_object_name": "compute_graph.flexml_layers[31]", + "layer_order": 41, + "mlir_dag_id_map": { + "dag_layer": 41, + "mlir_layer": 41 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[31].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[31].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[31].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[31].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[31].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[31].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[31].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[31].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[31].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[31].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[31].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[31].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[31].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[31].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[31].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[31].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaf80", + 384 + ], + "l1_pong": [ + "0xcd80", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_41" + ] + }, + "super_iter": 1 + }, + "0042": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_41" + ], + "out": [ + "compute_graph.l2_42" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[32].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[32].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[32].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[32].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[32].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[32].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[32].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[32].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[32].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[32].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[32].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[32].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[32].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[32].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[32].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[32].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_41" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_41", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_70", + "layer_object_name": "compute_graph.flexml_layers[32]", + "layer_order": 42, + "mlir_dag_id_map": { + "dag_layer": 42, + "mlir_layer": 42 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[32].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[32].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[32].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[32].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[32].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[32].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[32].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[32].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[32].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[32].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[32].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[32].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[32].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[32].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[32].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[32].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 256 + ], + "l1_pong": [ + "0xcd80", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_42" + ] + }, + "super_iter": 1 + }, + "0043": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_42" + ], + "out": [ + "compute_graph.l2_43", + "compute_graph.l2l3_43_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[33].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[33].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[33].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[33].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[33].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[33].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[33].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[33].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[33].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[33].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[33].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[33].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[33].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[33].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[33].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[33].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_42" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_42", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_43.out[0]", + "to": "compute_graph.l2l3_43_spill.in[0]" + }, + { + "from": "compute_graph.l2_43.out[1]", + "to": "compute_graph.l2l3_43_spill.in[1]" + } + ], + "layer_name": "Div_72", + "layer_object_name": "compute_graph.flexml_layers[33]", + "layer_order": 43, + "mlir_dag_id_map": { + "dag_layer": 43, + "mlir_layer": 43 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[33].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[33].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[33].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[33].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[33].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[33].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[33].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[33].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[33].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[33].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[33].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[33].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[33].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[33].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[33].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[33].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xae80", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 8192, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_43" + ], + "l3": { + "l2l3_43_spill": { + "size": 120 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_43_spill" + ] + }, + "super_iter": 1 + }, + "0044": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_43_spill", + "compute_graph.l2l3_scratch_0_44_spill", + "compute_graph.l2l3_scratch_1_44_spill" + ], + "out": [ + "compute_graph.l2l3_44_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_43_spill": { + "size": 120 + }, + "l2l3_scratch_0_44_spill": { + "size": 220800 + }, + "l2l3_scratch_1_44_spill": { + "size": 220800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_43_spill", + "compute_graph.l2l3_scratch_0_44_spill", + "compute_graph.l2l3_scratch_1_44_spill" + ] + }, + "kernel_name": [ + "TileAdf" + ], + "layer_name": "Generated-#16", + "layer_object_name": "compute_graph.templated_graph_44", + "layer_order": 44, + "mlir_dag_id_map": { + "dag_layer": 44, + "mlir_layer": 44 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_44_spill": { + "size": 110400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_44_spill" + ] + }, + "templated_graph": 1 + }, + "0045": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0", + "compute_graph.l2l3_44_spill", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636" + ], + "out": [ + "compute_graph.l2_45" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[34].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[34].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[34].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[34].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[34].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[34].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[34].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[34].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[34].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[34].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[34].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[34].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[34].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[34].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[34].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[34].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 110400, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_44_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[34].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[34].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[34].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[34].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[34].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[34].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[34].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[34].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[34].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[34].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[34].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[34].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[34].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[34].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[34].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[34].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 220800, + 122880, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_3636.in[1]" + }, + { + "from": "compute_graph.l2l3_44_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_44_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_44_for_layer_45_port0.in[1]" + } + ], + "layer_name": "Mul_73", + "layer_object_name": "compute_graph.flexml_layers[34]", + "layer_order": 45, + "mlir_dag_id_map": { + "dag_layer": 45, + "mlir_layer": 45 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[34].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[34].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[34].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[34].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[34].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[34].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[34].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[34].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[34].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[34].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[34].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[34].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[34].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[34].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[34].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[34].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_45" + ] + }, + "super_iter": 1 + }, + "0046": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_45", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434" + ], + "out": [ + "compute_graph.l2_46", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646" + ], + "wts": [ + "compute_graph.Layer_46_l2_wts", + "compute_graph.Layer_46_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[35].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[35].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[35].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[35].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[35].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[35].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[35].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[35].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[35].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[35].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[35].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[35].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[35].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[35].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[35].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[35].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2560 + ], + "l1_pong": [ + "0xcd80", + 2560 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_45" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[35].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[35].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[35].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[35].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[35].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[35].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[35].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[35].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[35].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[35].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[35].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[35].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[35].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[35].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[35].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[35].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 2560 + ], + "l1_pong": [ + "0x6000", + 2560 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_45", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_46.out[4]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.in[0]" + }, + { + "from": "compute_graph.l2_46.out[5]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_3434.in[1]" + } + ], + "layer_name": "Conv_74", + "layer_object_name": "compute_graph.flexml_layers[35]", + "layer_order": 46, + "mlir_dag_id_map": { + "dag_layer": 46, + "mlir_layer": 46 + }, + "num_iter": 12, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[35].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[35].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[35].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[35].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[35].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[35].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[35].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[35].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[35].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[35].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[35].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[35].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[35].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[35].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[35].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[35].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 640 + ], + "l1_pong": [ + "0x4000", + 640 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_46" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_4646": { + "size": 36800 + } + }, + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[35].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[35].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[35].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[35].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[35].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[35].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[35].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[35].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[35].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[35].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[35].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[35].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[35].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[35].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[35].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[35].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe180", + 1088 + ], + "l1_pong": [ + "0x9400", + 1088 + ], + "l2": [ + [ + 3, + 0, + 0, + 8704 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_46_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 175872, + "size": 8704 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_46_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0047": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_46" + ], + "out": [ + "compute_graph.l2_47" + ], + "wts": [ + "compute_graph.Layer_47_l2_wts", + "compute_graph.Layer_47_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[36].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[36].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[36].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[36].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[36].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[36].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[36].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[36].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[36].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[36].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[36].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[36].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[36].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[36].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[36].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[36].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4096 + ], + "l1_pong": [ + "0xcd80", + 4096 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_46" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_46", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_76", + "layer_object_name": "compute_graph.flexml_layers[36]", + "layer_order": 47, + "mlir_dag_id_map": { + "dag_layer": 47, + "mlir_layer": 47 + }, + "num_iter": 6, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[36].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[36].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[36].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[36].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[36].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[36].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[36].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[36].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[36].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[36].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[36].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[36].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[36].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[36].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[36].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[36].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 2, + 0, + 131072, + 196608, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_47" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[36].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[36].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[36].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[36].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[36].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[36].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[36].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[36].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[36].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[36].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[36].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[36].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[36].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[36].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[36].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[36].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xed80", + 2176 + ], + "l1_pong": [ + "0xa000", + 2176 + ], + "l2": [ + [ + 3, + 0, + 506880, + 8704 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_47_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 193280, + "size": 8704 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_47_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0048": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_47" + ], + "out": [ + "compute_graph.l2_48", + "compute_graph.l2l3_48_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848" + ], + "wts": [ + "compute_graph.Layer_48_l2_wts", + "compute_graph.Layer_48_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[37].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[37].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[37].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[37].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[37].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[37].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[37].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[37].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[37].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[37].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[37].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[37].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[37].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[37].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[37].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[37].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3840 + ], + "l1_pong": [ + "0xcd80", + 3840 + ], + "l2": [ + [ + 2, + 0, + 131072, + 196608, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_47" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_47", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_48.out[0]", + "to": "compute_graph.l2l3_48_spill.in[0]" + }, + { + "from": "compute_graph.l2_48.out[1]", + "to": "compute_graph.l2l3_48_spill.in[1]" + }, + { + "from": "compute_graph.l2_48.out[2]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.in[0]" + }, + { + "from": "compute_graph.l2_48.out[3]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.in[1]" + } + ], + "layer_name": "Conv_78", + "layer_object_name": "compute_graph.flexml_layers[37]", + "layer_order": 48, + "mlir_dag_id_map": { + "dag_layer": 48, + "mlir_layer": 48 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[37].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[37].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[37].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[37].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[37].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[37].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[37].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[37].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[37].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[37].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[37].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[37].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[37].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[37].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[37].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[37].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 384 + ], + "l1_pong": [ + "0x4000", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_48" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_4848": { + "size": 110400 + }, + "l2l3_48_spill": { + "size": 110400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_48_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[37].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[37].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[37].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[37].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[37].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[37].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[37].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[37].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[37].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[37].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[37].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[37].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[37].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[37].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[37].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[37].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xeb80", + 352 + ], + "l1_pong": [ + "0x9e00", + 352 + ], + "l2": [ + [ + 3, + 0, + 0, + 5632 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_48_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 210688, + "size": 5632 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_48_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0049": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_48_spill" + ], + "out": [ + "compute_graph.l2l3_49_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_48_spill": { + "size": 110400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_48_spill" + ] + }, + "kernel_name": [ + "Transpose4dAdf" + ], + "layer_name": "Generated-#18", + "layer_object_name": "compute_graph.templated_graph_49", + "layer_order": 49, + "mlir_dag_id_map": { + "dag_layer": 49, + "mlir_layer": 49 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_49_spill": { + "size": 110400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_49_spill" + ] + }, + "templated_graph": 1 + }, + "0050": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0", + "compute_graph.l2l3_49_spill" + ], + "out": [ + "compute_graph.l2_50" + ] + }, + "buffer_iter": 1, + "depth_iter": 15, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[38].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[38].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[38].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[38].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[38].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[38].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[38].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[38].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[38].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[38].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[38].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[38].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[38].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[38].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[38].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[38].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 8192 + ], + "l1_pong": [ + "0x4000", + 8192 + ], + "l2": [ + [ + 0, + 0, + 0, + 110400, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_49_spill" + ] + }, + "kernel_name": [ + "superkernel_reduce_mean_c8" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_49_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_49_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_49_for_layer_50_port0.in[1]" + } + ], + "layer_name": "Generated-#20", + "layer_object_name": "compute_graph.flexml_layers[38]", + "layer_order": 50, + "mlir_dag_id_map": { + "dag_layer": 50, + "mlir_layer": 50 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[38].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[38].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[38].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[38].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[38].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[38].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[38].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[38].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[38].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[38].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[38].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[38].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[38].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[38].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[38].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[38].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 32 + ], + "l1_pong": [ + "0xcd80", + 32 + ], + "l2": [ + [ + 2, + 0, + 523264, + 512, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_50" + ] + }, + "super_iter": 1 + }, + "0051": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_50" + ], + "out": [ + "compute_graph.l2_51" + ], + "wts": [ + "compute_graph.Layer_51_l2_wts", + "compute_graph.Layer_51_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[39].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[39].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[39].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[39].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[39].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[39].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[39].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[39].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[39].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[39].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[39].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[39].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[39].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[39].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[39].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[39].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 2, + 0, + 523264, + 512, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_50" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_50", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_81", + "layer_object_name": "compute_graph.flexml_layers[39]", + "layer_order": 51, + "mlir_dag_id_map": { + "dag_layer": 51, + "mlir_layer": 51 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[39].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[39].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[39].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[39].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[39].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[39].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[39].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[39].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[39].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[39].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[39].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[39].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[39].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[39].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[39].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[39].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 0, + 0, + 0, + 4096, + 32 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_51" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[39].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[39].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[39].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[39].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[39].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[39].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[39].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[39].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[39].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[39].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[39].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[39].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[39].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[39].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[39].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[39].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xdc80", + 1984 + ], + "l1_pong": [ + "0x8f00", + 1984 + ], + "l2": [ + [ + 3, + 0, + 0, + 7936 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_51_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 221952, + "size": 7936 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_51_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0052": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_51" + ], + "out": [ + "compute_graph.l2_52" + ], + "wts": [ + "compute_graph.Layer_52_l2_wts", + "compute_graph.Layer_52_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[40].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[40].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[40].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[40].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[40].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[40].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[40].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[40].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[40].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[40].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[40].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[40].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[40].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[40].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[40].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[40].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 4096, + 32 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_51" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_51", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_83", + "layer_object_name": "compute_graph.flexml_layers[40]", + "layer_order": 52, + "mlir_dag_id_map": { + "dag_layer": 52, + "mlir_layer": 52 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[40].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[40].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[40].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[40].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[40].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[40].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[40].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[40].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[40].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[40].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[40].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[40].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[40].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[40].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[40].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[40].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_52" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[40].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[40].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[40].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[40].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[40].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[40].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[40].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[40].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[40].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[40].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[40].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[40].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[40].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[40].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[40].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[40].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xd180", + 2176 + ], + "l1_pong": [ + "0x8400", + 2176 + ], + "l2": [ + [ + 3, + 0, + 506880, + 8704 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_52_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 237824, + "size": 8704 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_52_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0053": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_52" + ], + "out": [ + "compute_graph.l2_53" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[41].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[41].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[41].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[41].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[41].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[41].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[41].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[41].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[41].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[41].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[41].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[41].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[41].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[41].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[41].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[41].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_52" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_52", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_85", + "layer_object_name": "compute_graph.flexml_layers[41]", + "layer_order": 53, + "mlir_dag_id_map": { + "dag_layer": 53, + "mlir_layer": 53 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[41].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[41].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[41].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[41].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[41].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[41].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[41].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[41].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[41].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[41].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[41].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[41].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[41].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[41].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[41].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[41].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaf80", + 384 + ], + "l1_pong": [ + "0xcd80", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_53" + ] + }, + "super_iter": 1 + }, + "0054": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_53" + ], + "out": [ + "compute_graph.l2_54" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[42].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[42].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[42].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[42].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[42].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[42].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[42].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[42].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[42].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[42].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[42].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[42].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[42].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[42].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[42].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[42].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_53" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_53", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_88", + "layer_object_name": "compute_graph.flexml_layers[42]", + "layer_order": 54, + "mlir_dag_id_map": { + "dag_layer": 54, + "mlir_layer": 54 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[42].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[42].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[42].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[42].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[42].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[42].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[42].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[42].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[42].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[42].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[42].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[42].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[42].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[42].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[42].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[42].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 256 + ], + "l1_pong": [ + "0xcd80", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_54" + ] + }, + "super_iter": 1 + }, + "0055": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_54" + ], + "out": [ + "compute_graph.l2_55", + "compute_graph.l2l3_55_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[43].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[43].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[43].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[43].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[43].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[43].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[43].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[43].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[43].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[43].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[43].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[43].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[43].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[43].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[43].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[43].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_54" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_54", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_55.out[0]", + "to": "compute_graph.l2l3_55_spill.in[0]" + }, + { + "from": "compute_graph.l2_55.out[1]", + "to": "compute_graph.l2l3_55_spill.in[1]" + } + ], + "layer_name": "Div_90", + "layer_object_name": "compute_graph.flexml_layers[43]", + "layer_order": 55, + "mlir_dag_id_map": { + "dag_layer": 55, + "mlir_layer": 55 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[43].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[43].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[43].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[43].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[43].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[43].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[43].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[43].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[43].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[43].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[43].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[43].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[43].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[43].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[43].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[43].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xae80", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 8192, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_55" + ], + "l3": { + "l2l3_55_spill": { + "size": 120 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_55_spill" + ] + }, + "super_iter": 1 + }, + "0056": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_55_spill", + "compute_graph.l2l3_scratch_0_56_spill", + "compute_graph.l2l3_scratch_1_56_spill" + ], + "out": [ + "compute_graph.l2l3_56_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_55_spill": { + "size": 120 + }, + "l2l3_scratch_0_56_spill": { + "size": 220800 + }, + "l2l3_scratch_1_56_spill": { + "size": 220800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_55_spill", + "compute_graph.l2l3_scratch_0_56_spill", + "compute_graph.l2l3_scratch_1_56_spill" + ] + }, + "kernel_name": [ + "TileAdf" + ], + "layer_name": "Generated-#22", + "layer_object_name": "compute_graph.templated_graph_56", + "layer_order": 56, + "mlir_dag_id_map": { + "dag_layer": 56, + "mlir_layer": 56 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_56_spill": { + "size": 110400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_56_spill" + ] + }, + "templated_graph": 1 + }, + "0057": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0", + "compute_graph.l2l3_56_spill", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848" + ], + "out": [ + "compute_graph.l2_57" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[44].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[44].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[44].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[44].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[44].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[44].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[44].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[44].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[44].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[44].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[44].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[44].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[44].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[44].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[44].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[44].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 110400, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_56_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[44].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[44].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[44].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[44].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[44].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[44].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[44].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[44].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[44].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[44].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[44].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[44].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[44].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[44].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[44].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[44].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 220800, + 122880, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_4848.in[1]" + }, + { + "from": "compute_graph.l2l3_56_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_56_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_56_for_layer_57_port0.in[1]" + } + ], + "layer_name": "Mul_91", + "layer_object_name": "compute_graph.flexml_layers[44]", + "layer_order": 57, + "mlir_dag_id_map": { + "dag_layer": 57, + "mlir_layer": 57 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[44].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[44].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[44].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[44].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[44].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[44].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[44].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[44].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[44].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[44].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[44].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[44].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[44].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[44].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[44].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[44].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_57" + ] + }, + "super_iter": 1 + }, + "0058": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_57", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646" + ], + "out": [ + "compute_graph.l2_58", + "compute_graph.spill_L3_Concat_Buffer_layer_236" + ], + "wts": [ + "compute_graph.Layer_58_l2_wts", + "compute_graph.Layer_58_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[45].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[45].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[45].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[45].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[45].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[45].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[45].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[45].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[45].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[45].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[45].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[45].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[45].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[45].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[45].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[45].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2560 + ], + "l1_pong": [ + "0xcd80", + 2560 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 110400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_57" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[45].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[45].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[45].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[45].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[45].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[45].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[45].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[45].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[45].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[45].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[45].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[45].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[45].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[45].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[45].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[45].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 2560 + ], + "l1_pong": [ + "0x6000", + 2560 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_57", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_58.out[4]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_236.in[2]" + }, + { + "from": "compute_graph.l2_58.out[5]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_236.in[3]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_4646.in[1]" + } + ], + "layer_name": "Conv_92", + "layer_object_name": "compute_graph.flexml_layers[45]", + "layer_order": 58, + "mlir_dag_id_map": { + "dag_layer": 58, + "mlir_layer": 58 + }, + "num_iter": 12, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[45].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[45].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[45].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[45].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[45].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[45].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[45].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[45].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[45].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[45].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[45].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[45].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[45].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[45].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[45].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[45].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 640 + ], + "l1_pong": [ + "0x4000", + 640 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_58" + ], + "l3": { + "spill_L3_Concat_Buffer_layer_236": { + "buffer_bounds": [ + 1, + 40, + 23, + 40 + ], + "concat_offset": [ + 0, + 128, + 0, + 0 + ], + "size": 161920 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_236" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[45].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[45].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[45].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[45].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[45].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[45].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[45].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[45].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[45].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[45].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[45].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[45].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[45].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[45].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[45].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[45].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe180", + 1088 + ], + "l1_pong": [ + "0x9400", + 1088 + ], + "l2": [ + [ + 3, + 0, + 0, + 8704 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_58_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 255232, + "size": 8704 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_58_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0059": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_58" + ], + "out": [ + "compute_graph.l2_59", + "compute_graph.l2l3_59_spill" + ], + "wts": [ + "compute_graph.Layer_59_l2_wts", + "compute_graph.Layer_59_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[46].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[46].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[46].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[46].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[46].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[46].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[46].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[46].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[46].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[46].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[46].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[46].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[46].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[46].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[46].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[46].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4096 + ], + "l1_pong": [ + "0xcd80", + 4096 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_58" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_58", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_59.out[4]", + "to": "compute_graph.l2l3_59_spill.in[0]" + }, + { + "from": "compute_graph.l2_59.out[5]", + "to": "compute_graph.l2l3_59_spill.in[1]" + } + ], + "layer_name": "Conv_94", + "layer_object_name": "compute_graph.flexml_layers[46]", + "layer_order": 59, + "mlir_dag_id_map": { + "dag_layer": 59, + "mlir_layer": 59 + }, + "num_iter": 12, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[46].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[46].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[46].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[46].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[46].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[46].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[46].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[46].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[46].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[46].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[46].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[46].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[46].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[46].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[46].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[46].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 1, + 0, + 262144, + 393216, + 220800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_59" + ], + "l3": { + "l2l3_59_spill": { + "size": 220800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_59_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[46].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[46].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[46].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[46].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[46].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[46].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[46].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[46].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[46].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[46].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[46].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[46].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[46].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[46].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[46].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[46].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xed80", + 2176 + ], + "l1_pong": [ + "0xa000", + 2176 + ], + "l2": [ + [ + 3, + 0, + 489472, + 17408 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_59_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 272640, + "size": 17408 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_59_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0060": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_59" + ], + "out": [ + "compute_graph.l2_60" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[47].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[47].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[47].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[47].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[47].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[47].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[47].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[47].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[47].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[47].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[47].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[47].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[47].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[47].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[47].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[47].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 1, + 0, + 262144, + 393216, + 220800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_59" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_59", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_96", + "layer_object_name": "compute_graph.flexml_layers[47]", + "layer_order": 60, + "mlir_dag_id_map": { + "dag_layer": 60, + "mlir_layer": 60 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[47].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[47].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[47].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[47].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[47].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[47].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[47].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[47].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[47].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[47].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[47].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[47].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[47].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[47].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[47].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[47].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 245760, + 220800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_60" + ] + }, + "super_iter": 1 + }, + "0061": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_60" + ], + "out": [ + "compute_graph.l2_61" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[48].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[48].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[48].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[48].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[48].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[48].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[48].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[48].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[48].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[48].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[48].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[48].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[48].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[48].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[48].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[48].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 0, + 245760, + 220800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_60" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_60", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_99", + "layer_object_name": "compute_graph.flexml_layers[48]", + "layer_order": 61, + "mlir_dag_id_map": { + "dag_layer": 61, + "mlir_layer": 61 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[48].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[48].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[48].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[48].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[48].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[48].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[48].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[48].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[48].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[48].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[48].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[48].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[48].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[48].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[48].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[48].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 2, + 0, + 32768, + 245760, + 220800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_61" + ] + }, + "super_iter": 1 + }, + "0062": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_61" + ], + "out": [ + "compute_graph.l2_62", + "compute_graph.l2l3_62_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[49].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[49].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[49].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[49].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[49].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[49].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[49].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[49].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[49].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[49].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[49].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[49].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[49].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[49].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[49].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[49].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 32768, + 245760, + 220800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_61" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_61", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_62.out[0]", + "to": "compute_graph.l2l3_62_spill.in[0]" + }, + { + "from": "compute_graph.l2_62.out[1]", + "to": "compute_graph.l2l3_62_spill.in[1]" + } + ], + "layer_name": "Div_101", + "layer_object_name": "compute_graph.flexml_layers[49]", + "layer_order": 62, + "mlir_dag_id_map": { + "dag_layer": 62, + "mlir_layer": 62 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[49].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[49].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[49].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[49].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[49].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[49].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[49].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[49].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[49].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[49].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[49].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[49].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[49].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[49].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[49].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[49].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 245760, + 220800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_62" + ], + "l3": { + "l2l3_62_spill": { + "size": 220800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_62_spill" + ] + }, + "super_iter": 1 + }, + "0063": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0", + "compute_graph.l2l3_59_spill", + "compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1", + "compute_graph.l2l3_62_spill" + ], + "out": [ + "compute_graph.l2_63", + "compute_graph.l2l3_63_spill" + ] + }, + "buffer_iter": 2, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[50].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[50].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[50].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[50].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[50].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[50].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[50].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[50].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[50].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[50].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[50].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[50].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[50].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[50].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[50].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[50].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 293888, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_59_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[50].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[50].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[50].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[50].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[50].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[50].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[50].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[50].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[50].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[50].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[50].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[50].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[50].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[50].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[50].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[50].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 2, + 0, + 63488, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_62_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_63.out[0]", + "to": "compute_graph.l2l3_63_spill.in[0]" + }, + { + "from": "compute_graph.l2_63.out[1]", + "to": "compute_graph.l2l3_63_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_59_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_59_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_59_for_layer_63_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_62_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_62_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_62_for_layer_63_port1.in[1]" + } + ], + "layer_name": "Mul_102", + "layer_object_name": "compute_graph.flexml_layers[50]", + "layer_order": 63, + "mlir_dag_id_map": { + "dag_layer": 63, + "mlir_layer": 63 + }, + "num_iter": 16, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[50].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[50].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[50].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[50].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[50].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[50].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[50].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[50].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[50].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[50].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[50].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[50].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[50].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[50].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[50].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[50].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_63" + ], + "l2_force_single_buffering": true, + "l3": { + "l2l3_63_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_63_spill" + ] + }, + "super_iter": 2 + }, + "0064": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_63_spill", + "compute_graph.l2l3_scratch_0_64_spill" + ], + "out": [ + "compute_graph.l2l3_64_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_63_spill": { + "size": 230400 + }, + "l2l3_scratch_0_64_spill": { + "size": 441600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_63_spill", + "compute_graph.l2l3_scratch_0_64_spill" + ] + }, + "kernel_name": [ + "BufferUnpadAdf" + ], + "layer_name": "Generated-#64", + "layer_object_name": "compute_graph.templated_graph_64", + "layer_order": 64, + "mlir_dag_id_map": { + "dag_layer": 64, + "mlir_layer": 64 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_64_spill": { + "size": 220800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_64_spill" + ] + }, + "templated_graph": 1 + }, + "0065": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0", + "compute_graph.l2l3_64_spill" + ], + "out": [ + "compute_graph.l2_65" + ], + "wts": [ + "compute_graph.Layer_65_l2_wts", + "compute_graph.Layer_65_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[51].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[51].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[51].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[51].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[51].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[51].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[51].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[51].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[51].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[51].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[51].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[51].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[51].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[51].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[51].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[51].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3840 + ], + "l1_pong": [ + "0xcd80", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 220800, + 220800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_64_spill" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_64_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_64_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_64_for_layer_65_port0.in[1]" + } + ], + "layer_name": "Conv_103", + "layer_object_name": "compute_graph.flexml_layers[51]", + "layer_order": 65, + "mlir_dag_id_map": { + "dag_layer": 65, + "mlir_layer": 65 + }, + "num_iter": 40, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[51].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[51].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[51].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[51].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[51].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[51].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[51].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[51].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[51].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[51].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[51].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[51].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[51].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[51].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[51].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[51].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 128 + ], + "l1_pong": [ + "0x4000", + 128 + ], + "l2": [ + [ + 2, + 0, + 360448, + 81920, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_65" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[51].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[51].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[51].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[51].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[51].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[51].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[51].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[51].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[51].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[51].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[51].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[51].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[51].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[51].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[51].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[51].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xeb80", + 128 + ], + "l1_pong": [ + "0x9e00", + 128 + ], + "l2": [ + [ + 3, + 0, + 0, + 4096 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_65_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 307456, + "size": 4096 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_65_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0066": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_65" + ], + "out": [ + "compute_graph.l2_66" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[52].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[52].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[52].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[52].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[52].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[52].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[52].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[52].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[52].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[52].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[52].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[52].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[52].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[52].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[52].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[52].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 360448, + 81920, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_65" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_65", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_105", + "layer_object_name": "compute_graph.flexml_layers[52]", + "layer_order": 66, + "mlir_dag_id_map": { + "dag_layer": 66, + "mlir_layer": 66 + }, + "num_iter": 2, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[52].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[52].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[52].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[52].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[52].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[52].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[52].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[52].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[52].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[52].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[52].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[52].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[52].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[52].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[52].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[52].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_66" + ] + }, + "super_iter": 1 + }, + "0067": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_66" + ], + "out": [ + "compute_graph.l2_67" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[53].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[53].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[53].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[53].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[53].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[53].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[53].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[53].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[53].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[53].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[53].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[53].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[53].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[53].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[53].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[53].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_66" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_66", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_108", + "layer_object_name": "compute_graph.flexml_layers[53]", + "layer_order": 67, + "mlir_dag_id_map": { + "dag_layer": 67, + "mlir_layer": 67 + }, + "num_iter": 2, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[53].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[53].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[53].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[53].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[53].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[53].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[53].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[53].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[53].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[53].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[53].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[53].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[53].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[53].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[53].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[53].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 2, + 0, + 237568, + 61440, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_67" + ] + }, + "super_iter": 1 + }, + "0068": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_67" + ], + "out": [ + "compute_graph.l2_68" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[54].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[54].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[54].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[54].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[54].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[54].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[54].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[54].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[54].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[54].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[54].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[54].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[54].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[54].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[54].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[54].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 237568, + 61440, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_67" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_67", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_110", + "layer_object_name": "compute_graph.flexml_layers[54]", + "layer_order": 68, + "mlir_dag_id_map": { + "dag_layer": 68, + "mlir_layer": 68 + }, + "num_iter": 2, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[54].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[54].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[54].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[54].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[54].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[54].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[54].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[54].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[54].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[54].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[54].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[54].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[54].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[54].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[54].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[54].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_68" + ] + }, + "super_iter": 1 + }, + "0069": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_65", + "compute_graph.l2_68" + ], + "out": [ + "compute_graph.l2_69" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[55].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[55].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[55].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[55].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[55].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[55].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[55].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[55].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[55].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[55].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[55].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[55].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[55].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[55].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[55].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[55].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 360448, + 81920, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_65" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[55].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[55].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[55].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[55].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[55].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[55].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[55].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[55].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[55].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[55].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[55].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[55].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[55].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[55].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[55].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[55].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_68" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_65", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_68", + "dest_port": 2, + "src_offset": 0 + } + ], + "layer_name": "Mul_111", + "layer_object_name": "compute_graph.flexml_layers[55]", + "layer_order": 69, + "mlir_dag_id_map": { + "dag_layer": 69, + "mlir_layer": 69 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[55].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[55].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[55].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[55].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[55].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[55].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[55].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[55].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[55].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[55].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[55].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[55].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[55].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[55].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[55].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[55].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 122880, + 61440, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_69" + ] + }, + "super_iter": 1 + }, + "0070": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_69" + ], + "out": [ + "compute_graph.l2_70" + ], + "wts": [ + "compute_graph.Layer_70_l2_wts", + "compute_graph.Layer_70_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 3, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[56].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[56].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[56].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[56].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[56].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[56].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[56].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[56].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[56].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[56].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[56].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[56].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[56].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[56].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[56].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[56].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2560 + ], + "l1_pong": [ + "0xcd80", + 2560 + ], + "l2": [ + [ + 0, + 0, + 122880, + 61440, + 57600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_69" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_69", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_112", + "layer_object_name": "compute_graph.flexml_layers[56]", + "layer_order": 70, + "mlir_dag_id_map": { + "dag_layer": 70, + "mlir_layer": 70 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[56].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[56].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[56].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[56].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[56].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[56].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[56].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[56].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[56].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[56].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[56].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[56].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[56].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[56].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[56].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[56].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_70" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[56].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[56].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[56].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[56].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[56].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[56].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[56].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[56].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[56].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[56].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[56].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[56].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[56].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[56].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[56].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[56].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe180", + 2016 + ], + "l1_pong": [ + "0x9400", + 2016 + ], + "l2": [ + [ + 3, + 0, + 475904, + 24192 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_70_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 315648, + "size": 24192 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_70_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0071": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_70" + ], + "out": [ + "compute_graph.l2_71" + ], + "wts": [ + "compute_graph.Layer_71_l2_wts", + "compute_graph.Layer_71_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[57].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[57].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[57].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[57].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[57].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[57].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[57].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[57].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[57].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[57].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[57].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[57].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[57].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[57].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[57].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[57].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2560 + ], + "l1_pong": [ + "0xcd80", + 2560 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_70" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_70", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_113", + "layer_object_name": "compute_graph.flexml_layers[57]", + "layer_order": 71, + "mlir_dag_id_map": { + "dag_layer": 71, + "mlir_layer": 71 + }, + "num_iter": 6, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[57].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[57].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[57].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[57].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[57].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[57].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[57].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[57].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[57].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[57].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[57].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[57].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[57].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[57].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[57].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[57].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 73728, + 98304, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_71" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[57].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[57].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[57].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[57].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[57].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[57].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[57].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[57].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[57].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[57].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[57].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[57].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[57].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[57].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[57].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[57].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe180", + 2688 + ], + "l1_pong": [ + "0x9400", + 2688 + ], + "l2": [ + [ + 3, + 0, + 0, + 21504 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_71_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 364032, + "size": 21504 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_71_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0072": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_71" + ], + "out": [ + "compute_graph.l2_72" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[58].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[58].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[58].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[58].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[58].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[58].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[58].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[58].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[58].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[58].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[58].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[58].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[58].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[58].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[58].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[58].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 73728, + 98304, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_71" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_71", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_115", + "layer_object_name": "compute_graph.flexml_layers[58]", + "layer_order": 72, + "mlir_dag_id_map": { + "dag_layer": 72, + "mlir_layer": 72 + }, + "num_iter": 2, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[58].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[58].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[58].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[58].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[58].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[58].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[58].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[58].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[58].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[58].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[58].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[58].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[58].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[58].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[58].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[58].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 270336, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_72" + ] + }, + "super_iter": 1 + }, + "0073": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_72" + ], + "out": [ + "compute_graph.l2_73" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[59].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[59].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[59].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[59].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[59].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[59].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[59].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[59].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[59].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[59].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[59].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[59].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[59].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[59].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[59].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[59].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 270336, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_72" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_72", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_118", + "layer_object_name": "compute_graph.flexml_layers[59]", + "layer_order": 73, + "mlir_dag_id_map": { + "dag_layer": 73, + "mlir_layer": 73 + }, + "num_iter": 2, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[59].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[59].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[59].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[59].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[59].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[59].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[59].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[59].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[59].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[59].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[59].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[59].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[59].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[59].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[59].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[59].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_73" + ] + }, + "super_iter": 1 + }, + "0074": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_73" + ], + "out": [ + "compute_graph.l2_74" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[60].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[60].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[60].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[60].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[60].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[60].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[60].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[60].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[60].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[60].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[60].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[60].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[60].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[60].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[60].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[60].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_73" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_73", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_120", + "layer_object_name": "compute_graph.flexml_layers[60]", + "layer_order": 74, + "mlir_dag_id_map": { + "dag_layer": 74, + "mlir_layer": 74 + }, + "num_iter": 2, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[60].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[60].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[60].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[60].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[60].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[60].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[60].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[60].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[60].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[60].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[60].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[60].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[60].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[60].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[60].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[60].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 270336, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_74" + ] + }, + "super_iter": 1 + }, + "0075": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_74", + "compute_graph.l2_71" + ], + "out": [ + "compute_graph.l2_75" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[61].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[61].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[61].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[61].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[61].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[61].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[61].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[61].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[61].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[61].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[61].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[61].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[61].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[61].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[61].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[61].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73728, + 98304, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_71" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[61].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[61].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[61].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[61].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[61].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[61].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[61].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[61].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[61].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[61].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[61].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[61].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[61].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[61].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[61].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[61].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 270336, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_74" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_71", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_74", + "dest_port": 2, + "src_offset": 0 + } + ], + "layer_name": "Mul_121", + "layer_object_name": "compute_graph.flexml_layers[61]", + "layer_order": 75, + "mlir_dag_id_map": { + "dag_layer": 75, + "mlir_layer": 75 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[61].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[61].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[61].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[61].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[61].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[61].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[61].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[61].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[61].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[61].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[61].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[61].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[61].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[61].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[61].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[61].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_75" + ] + }, + "super_iter": 1 + }, + "0076": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_75" + ], + "out": [ + "compute_graph.l2_76" + ], + "wts": [ + "compute_graph.Layer_76_l2_wts", + "compute_graph.Layer_76_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[62].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[62].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[62].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[62].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[62].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[62].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[62].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[62].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[62].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[62].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[62].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[62].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[62].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[62].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[62].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[62].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5376 + ], + "l1_pong": [ + "0xcd80", + 5376 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_75" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_75", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_122", + "layer_object_name": "compute_graph.flexml_layers[62]", + "layer_order": 76, + "mlir_dag_id_map": { + "dag_layer": 76, + "mlir_layer": 76 + }, + "num_iter": 7, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[62].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[62].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[62].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[62].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[62].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[62].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[62].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[62].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[62].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[62].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[62].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[62].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[62].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[62].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[62].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[62].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 73728, + 86016, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_76" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[62].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[62].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[62].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[62].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[62].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[62].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[62].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[62].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[62].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[62].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[62].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[62].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[62].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[62].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[62].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[62].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf780", + 128 + ], + "l1_pong": [ + "0xaa00", + 128 + ], + "l2": [ + [ + 3, + 0, + 517120, + 3584 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_76_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 407040, + "size": 3584 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_76_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0077": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_76" + ], + "out": [ + "compute_graph.l2_77" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[63].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[63].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[63].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[63].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[63].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[63].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[63].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[63].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[63].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[63].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[63].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[63].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[63].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[63].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[63].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[63].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 73728, + 86016, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_76" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_76", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_124", + "layer_object_name": "compute_graph.flexml_layers[63]", + "layer_order": 77, + "mlir_dag_id_map": { + "dag_layer": 77, + "mlir_layer": 77 + }, + "num_iter": 2, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[63].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[63].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[63].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[63].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[63].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[63].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[63].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[63].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[63].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[63].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[63].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[63].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[63].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[63].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[63].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[63].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 245760, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_77" + ] + }, + "super_iter": 1 + }, + "0078": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_77" + ], + "out": [ + "compute_graph.l2_78" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[64].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[64].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[64].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[64].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[64].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[64].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[64].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[64].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[64].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[64].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[64].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[64].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[64].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[64].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[64].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[64].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 245760, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_77" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_77", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_127", + "layer_object_name": "compute_graph.flexml_layers[64]", + "layer_order": 78, + "mlir_dag_id_map": { + "dag_layer": 78, + "mlir_layer": 78 + }, + "num_iter": 2, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[64].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[64].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[64].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[64].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[64].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[64].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[64].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[64].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[64].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[64].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[64].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[64].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[64].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[64].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[64].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[64].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_78" + ] + }, + "super_iter": 1 + }, + "0079": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_78" + ], + "out": [ + "compute_graph.l2_79" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[65].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[65].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[65].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[65].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[65].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[65].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[65].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[65].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[65].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[65].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[65].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[65].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[65].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[65].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[65].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[65].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_78" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_78", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_129", + "layer_object_name": "compute_graph.flexml_layers[65]", + "layer_order": 79, + "mlir_dag_id_map": { + "dag_layer": 79, + "mlir_layer": 79 + }, + "num_iter": 2, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[65].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[65].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[65].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[65].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[65].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[65].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[65].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[65].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[65].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[65].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[65].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[65].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[65].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[65].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[65].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[65].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 245760, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_79" + ] + }, + "super_iter": 1 + }, + "0080": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_79", + "compute_graph.l2_76" + ], + "out": [ + "compute_graph.l2_80" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[66].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[66].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[66].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[66].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[66].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[66].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[66].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[66].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[66].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[66].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[66].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[66].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[66].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[66].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[66].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[66].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73728, + 86016, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_76" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[66].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[66].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[66].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[66].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[66].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[66].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[66].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[66].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[66].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[66].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[66].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[66].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[66].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[66].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[66].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[66].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 245760, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_79" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_76", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_79", + "dest_port": 2, + "src_offset": 0 + } + ], + "layer_name": "Mul_130", + "layer_object_name": "compute_graph.flexml_layers[66]", + "layer_order": 80, + "mlir_dag_id_map": { + "dag_layer": 80, + "mlir_layer": 80 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[66].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[66].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[66].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[66].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[66].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[66].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[66].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[66].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[66].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[66].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[66].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[66].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[66].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[66].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[66].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[66].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_80" + ] + }, + "super_iter": 1 + }, + "0081": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_70", + "compute_graph.l2_80" + ], + "out": [ + "compute_graph.l2_81" + ], + "wts": [ + "compute_graph.Layer_81_l2_wts", + "compute_graph.Layer_81_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[67].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[67].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[67].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[67].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[67].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[67].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[67].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[67].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[67].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[67].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[67].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[67].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[67].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[67].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[67].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[67].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3328 + ], + "l1_pong": [ + "0xcd80", + 3328 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 48000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_80" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[67].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[67].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[67].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[67].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[67].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[67].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[67].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[67].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[67].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[67].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[67].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[67].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[67].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[67].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[67].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[67].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 3072 + ], + "l1_pong": [ + "0x6000", + 3072 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_70" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_70", + "dest_port": 2, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_80", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_131", + "layer_object_name": "compute_graph.flexml_layers[67]", + "layer_order": 81, + "mlir_dag_id_map": { + "dag_layer": 81, + "mlir_layer": 81 + }, + "num_iter": 6, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[67].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[67].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[67].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[67].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[67].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[67].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[67].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[67].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[67].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[67].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[67].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[67].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[67].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[67].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[67].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[67].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_81" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[67].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[67].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[67].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[67].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[67].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[67].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[67].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[67].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[67].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[67].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[67].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[67].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[67].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[67].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[67].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[67].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe780", + 2592 + ], + "l1_pong": [ + "0x9a00", + 2592 + ], + "l2": [ + [ + 3, + 0, + 0, + 20736 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_81_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 414208, + "size": 20736 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_81_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0082": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_81" + ], + "out": [ + "compute_graph.l2_82" + ], + "wts": [ + "compute_graph.Layer_82_l2_wts", + "compute_graph.Layer_82_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[68].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[68].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[68].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[68].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[68].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[68].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[68].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[68].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[68].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[68].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[68].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[68].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[68].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[68].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[68].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[68].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_81" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_81", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_133", + "layer_object_name": "compute_graph.flexml_layers[68]", + "layer_order": 82, + "mlir_dag_id_map": { + "dag_layer": 82, + "mlir_layer": 82 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[68].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[68].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[68].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[68].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[68].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[68].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[68].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[68].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[68].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[68].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[68].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[68].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[68].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[68].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[68].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[68].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1152 + ], + "l1_pong": [ + "0x4000", + 1152 + ], + "l2": [ + [ + 0, + 0, + 73728, + 55296, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_82" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[68].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[68].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[68].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[68].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[68].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[68].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[68].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[68].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[68].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[68].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[68].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[68].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[68].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[68].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[68].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[68].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xdc80", + 4032 + ], + "l1_pong": [ + "0x8f00", + 4032 + ], + "l2": [ + [ + 3, + 0, + 492032, + 16128 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_82_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 455680, + "size": 16128 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_82_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0083": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_82" + ], + "out": [ + "compute_graph.l2_83" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[69].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[69].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[69].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[69].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[69].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[69].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[69].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[69].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[69].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[69].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[69].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[69].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[69].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[69].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[69].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[69].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73728, + 55296, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_82" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_82", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_135", + "layer_object_name": "compute_graph.flexml_layers[69]", + "layer_order": 83, + "mlir_dag_id_map": { + "dag_layer": 83, + "mlir_layer": 83 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[69].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[69].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[69].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[69].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[69].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[69].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[69].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[69].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[69].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[69].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[69].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[69].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[69].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[69].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[69].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[69].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 184320, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_83" + ] + }, + "super_iter": 1 + }, + "0084": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_83" + ], + "out": [ + "compute_graph.l2_84" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[70].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[70].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[70].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[70].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[70].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[70].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[70].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[70].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[70].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[70].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[70].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[70].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[70].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[70].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[70].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[70].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 184320, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_83" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_83", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_138", + "layer_object_name": "compute_graph.flexml_layers[70]", + "layer_order": 84, + "mlir_dag_id_map": { + "dag_layer": 84, + "mlir_layer": 84 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[70].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[70].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[70].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[70].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[70].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[70].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[70].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[70].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[70].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[70].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[70].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[70].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[70].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[70].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[70].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[70].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_84" + ] + }, + "super_iter": 1 + }, + "0085": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_84" + ], + "out": [ + "compute_graph.l2_85" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[71].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[71].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[71].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[71].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[71].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[71].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[71].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[71].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[71].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[71].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[71].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[71].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[71].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[71].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[71].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[71].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_84" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_84", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_140", + "layer_object_name": "compute_graph.flexml_layers[71]", + "layer_order": 85, + "mlir_dag_id_map": { + "dag_layer": 85, + "mlir_layer": 85 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[71].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[71].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[71].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[71].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[71].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[71].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[71].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[71].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[71].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[71].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[71].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[71].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[71].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[71].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[71].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[71].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 184320, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_85" + ] + }, + "super_iter": 1 + }, + "0086": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_85", + "compute_graph.l2_82" + ], + "out": [ + "compute_graph.l2_86" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[72].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[72].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[72].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[72].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[72].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[72].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[72].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[72].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[72].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[72].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[72].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[72].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[72].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[72].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[72].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[72].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73728, + 55296, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_82" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[72].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[72].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[72].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[72].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[72].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[72].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[72].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[72].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[72].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[72].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[72].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[72].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[72].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[72].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[72].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[72].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 184320, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_85" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_82", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_85", + "dest_port": 2, + "src_offset": 0 + } + ], + "layer_name": "Mul_141", + "layer_object_name": "compute_graph.flexml_layers[72]", + "layer_order": 86, + "mlir_dag_id_map": { + "dag_layer": 86, + "mlir_layer": 86 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[72].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[72].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[72].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[72].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[72].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[72].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[72].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[72].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[72].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[72].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[72].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[72].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[72].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[72].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[72].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[72].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_86" + ] + }, + "super_iter": 1 + }, + "0087": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_86" + ], + "out": [ + "compute_graph.l2_87" + ], + "wts": [ + "compute_graph.Layer_87_l2_wts", + "compute_graph.Layer_87_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[73].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[73].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[73].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[73].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[73].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[73].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[73].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[73].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[73].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[73].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[73].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[73].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[73].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[73].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[73].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[73].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5376 + ], + "l1_pong": [ + "0xcd80", + 5376 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_86" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_86", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_142", + "layer_object_name": "compute_graph.flexml_layers[73]", + "layer_order": 87, + "mlir_dag_id_map": { + "dag_layer": 87, + "mlir_layer": 87 + }, + "num_iter": 6, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[73].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[73].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[73].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[73].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[73].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[73].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[73].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[73].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[73].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[73].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[73].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[73].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[73].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[73].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[73].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[73].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 73728, + 73728, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_87" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[73].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[73].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[73].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[73].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[73].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[73].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[73].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[73].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[73].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[73].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[73].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[73].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[73].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[73].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[73].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[73].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf780", + 128 + ], + "l1_pong": [ + "0xaa00", + 128 + ], + "l2": [ + [ + 3, + 0, + 0, + 3072 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_87_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 487936, + "size": 3072 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_87_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0088": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_87" + ], + "out": [ + "compute_graph.l2_88" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[74].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[74].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[74].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[74].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[74].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[74].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[74].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[74].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[74].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[74].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[74].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[74].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[74].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[74].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[74].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[74].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73728, + 73728, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_87" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_87", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_144", + "layer_object_name": "compute_graph.flexml_layers[74]", + "layer_order": 88, + "mlir_dag_id_map": { + "dag_layer": 88, + "mlir_layer": 88 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[74].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[74].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[74].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[74].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[74].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[74].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[74].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[74].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[74].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[74].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[74].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[74].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[74].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[74].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[74].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[74].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 221184, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_88" + ] + }, + "super_iter": 1 + }, + "0089": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_88" + ], + "out": [ + "compute_graph.l2_89" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[75].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[75].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[75].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[75].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[75].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[75].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[75].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[75].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[75].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[75].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[75].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[75].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[75].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[75].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[75].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[75].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 221184, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_88" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_88", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_147", + "layer_object_name": "compute_graph.flexml_layers[75]", + "layer_order": 89, + "mlir_dag_id_map": { + "dag_layer": 89, + "mlir_layer": 89 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[75].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[75].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[75].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[75].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[75].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[75].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[75].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[75].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[75].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[75].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[75].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[75].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[75].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[75].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[75].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[75].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_89" + ] + }, + "super_iter": 1 + }, + "0090": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_89" + ], + "out": [ + "compute_graph.l2_90" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[76].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[76].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[76].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[76].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[76].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[76].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[76].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[76].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[76].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[76].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[76].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[76].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[76].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[76].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[76].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[76].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_89" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_89", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_149", + "layer_object_name": "compute_graph.flexml_layers[76]", + "layer_order": 90, + "mlir_dag_id_map": { + "dag_layer": 90, + "mlir_layer": 90 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[76].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[76].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[76].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[76].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[76].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[76].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[76].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[76].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[76].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[76].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[76].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[76].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[76].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[76].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[76].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[76].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 221184, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_90" + ] + }, + "super_iter": 1 + }, + "0091": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_90", + "compute_graph.l2_87" + ], + "out": [ + "compute_graph.l2_91" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[77].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[77].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[77].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[77].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[77].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[77].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[77].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[77].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[77].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[77].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[77].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[77].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[77].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[77].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[77].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[77].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73728, + 73728, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_87" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[77].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[77].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[77].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[77].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[77].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[77].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[77].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[77].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[77].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[77].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[77].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[77].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[77].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[77].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[77].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[77].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 221184, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_90" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_87", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_90", + "dest_port": 2, + "src_offset": 0 + } + ], + "layer_name": "Mul_150", + "layer_object_name": "compute_graph.flexml_layers[77]", + "layer_order": 91, + "mlir_dag_id_map": { + "dag_layer": 91, + "mlir_layer": 91 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[77].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[77].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[77].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[77].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[77].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[77].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[77].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[77].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[77].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[77].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[77].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[77].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[77].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[77].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[77].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[77].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_91" + ] + }, + "super_iter": 1 + }, + "0092": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_81", + "compute_graph.l2_91" + ], + "out": [ + "compute_graph.l2_92" + ], + "wts": [ + "compute_graph.Layer_92_l2_wts", + "compute_graph.Layer_92_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[78].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[78].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[78].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[78].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[78].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[78].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[78].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[78].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[78].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[78].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[78].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[78].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[78].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[78].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[78].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[78].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3072 + ], + "l1_pong": [ + "0xcd80", + 3072 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_91" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[78].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[78].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[78].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[78].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[78].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[78].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[78].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[78].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[78].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[78].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[78].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[78].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[78].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[78].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[78].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[78].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 3072 + ], + "l1_pong": [ + "0x6000", + 3072 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_81" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_81", + "dest_port": 2, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_91", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_151", + "layer_object_name": "compute_graph.flexml_layers[78]", + "layer_order": 92, + "mlir_dag_id_map": { + "dag_layer": 92, + "mlir_layer": 92 + }, + "num_iter": 6, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[78].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[78].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[78].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[78].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[78].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[78].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[78].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[78].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[78].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[78].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[78].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[78].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[78].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[78].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[78].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[78].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_92" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[78].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[78].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[78].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[78].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[78].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[78].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[78].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[78].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[78].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[78].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[78].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[78].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[78].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[78].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[78].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[78].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe580", + 2400 + ], + "l1_pong": [ + "0x9800", + 2400 + ], + "l2": [ + [ + 3, + 0, + 485888, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_92_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 494080, + "size": 19200 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_92_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0093": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_92" + ], + "out": [ + "compute_graph.l2_93" + ], + "wts": [ + "compute_graph.Layer_93_l2_wts", + "compute_graph.Layer_93_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[79].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[79].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[79].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[79].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[79].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[79].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[79].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[79].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[79].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[79].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[79].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[79].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[79].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[79].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[79].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[79].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_92" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_92", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_153", + "layer_object_name": "compute_graph.flexml_layers[79]", + "layer_order": 93, + "mlir_dag_id_map": { + "dag_layer": 93, + "mlir_layer": 93 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[79].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[79].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[79].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[79].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[79].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[79].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[79].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[79].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[79].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[79].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[79].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[79].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[79].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[79].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[79].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[79].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1152 + ], + "l1_pong": [ + "0x4000", + 1152 + ], + "l2": [ + [ + 0, + 0, + 73728, + 55296, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_93" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[79].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[79].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[79].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[79].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[79].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[79].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[79].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[79].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[79].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[79].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[79].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[79].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[79].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[79].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[79].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[79].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xdc80", + 4032 + ], + "l1_pong": [ + "0x8f00", + 4032 + ], + "l2": [ + [ + 3, + 0, + 0, + 16128 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_93_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 532480, + "size": 16128 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_93_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0094": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_93" + ], + "out": [ + "compute_graph.l2_94" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[80].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[80].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[80].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[80].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[80].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[80].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[80].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[80].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[80].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[80].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[80].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[80].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[80].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[80].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[80].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[80].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73728, + 55296, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_93" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_93", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_155", + "layer_object_name": "compute_graph.flexml_layers[80]", + "layer_order": 94, + "mlir_dag_id_map": { + "dag_layer": 94, + "mlir_layer": 94 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[80].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[80].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[80].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[80].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[80].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[80].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[80].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[80].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[80].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[80].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[80].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[80].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[80].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[80].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[80].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[80].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 184320, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_94" + ] + }, + "super_iter": 1 + }, + "0095": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_94" + ], + "out": [ + "compute_graph.l2_95" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[81].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[81].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[81].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[81].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[81].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[81].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[81].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[81].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[81].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[81].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[81].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[81].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[81].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[81].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[81].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[81].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 184320, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_94" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_94", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_158", + "layer_object_name": "compute_graph.flexml_layers[81]", + "layer_order": 95, + "mlir_dag_id_map": { + "dag_layer": 95, + "mlir_layer": 95 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[81].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[81].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[81].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[81].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[81].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[81].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[81].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[81].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[81].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[81].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[81].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[81].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[81].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[81].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[81].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[81].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_95" + ] + }, + "super_iter": 1 + }, + "0096": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_95" + ], + "out": [ + "compute_graph.l2_96" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[82].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[82].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[82].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[82].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[82].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[82].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[82].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[82].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[82].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[82].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[82].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[82].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[82].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[82].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[82].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[82].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_95" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_95", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_160", + "layer_object_name": "compute_graph.flexml_layers[82]", + "layer_order": 96, + "mlir_dag_id_map": { + "dag_layer": 96, + "mlir_layer": 96 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[82].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[82].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[82].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[82].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[82].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[82].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[82].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[82].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[82].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[82].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[82].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[82].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[82].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[82].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[82].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[82].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 184320, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_96" + ] + }, + "super_iter": 1 + }, + "0097": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_96", + "compute_graph.l2_93" + ], + "out": [ + "compute_graph.l2_97" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[83].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[83].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[83].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[83].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[83].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[83].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[83].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[83].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[83].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[83].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[83].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[83].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[83].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[83].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[83].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[83].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73728, + 55296, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_93" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[83].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[83].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[83].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[83].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[83].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[83].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[83].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[83].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[83].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[83].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[83].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[83].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[83].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[83].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[83].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[83].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 184320, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_96" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_93", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_96", + "dest_port": 2, + "src_offset": 0 + } + ], + "layer_name": "Mul_161", + "layer_object_name": "compute_graph.flexml_layers[83]", + "layer_order": 97, + "mlir_dag_id_map": { + "dag_layer": 97, + "mlir_layer": 97 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[83].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[83].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[83].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[83].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[83].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[83].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[83].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[83].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[83].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[83].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[83].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[83].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[83].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[83].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[83].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[83].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_97" + ] + }, + "super_iter": 1 + }, + "0098": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_97" + ], + "out": [ + "compute_graph.l2_98" + ], + "wts": [ + "compute_graph.Layer_98_l2_wts", + "compute_graph.Layer_98_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[84].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[84].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[84].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[84].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[84].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[84].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[84].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[84].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[84].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[84].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[84].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[84].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[84].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[84].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[84].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[84].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5376 + ], + "l1_pong": [ + "0xcd80", + 5376 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_97" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_97", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_162", + "layer_object_name": "compute_graph.flexml_layers[84]", + "layer_order": 98, + "mlir_dag_id_map": { + "dag_layer": 98, + "mlir_layer": 98 + }, + "num_iter": 6, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[84].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[84].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[84].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[84].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[84].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[84].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[84].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[84].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[84].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[84].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[84].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[84].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[84].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[84].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[84].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[84].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 73728, + 73728, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_98" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[84].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[84].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[84].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[84].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[84].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[84].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[84].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[84].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[84].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[84].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[84].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[84].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[84].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[84].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[84].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[84].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf780", + 128 + ], + "l1_pong": [ + "0xaa00", + 128 + ], + "l2": [ + [ + 3, + 0, + 518144, + 3072 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_98_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 564736, + "size": 3072 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_98_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0099": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_98" + ], + "out": [ + "compute_graph.l2_99" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[85].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[85].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[85].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[85].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[85].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[85].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[85].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[85].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[85].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[85].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[85].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[85].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[85].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[85].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[85].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[85].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73728, + 73728, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_98" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_98", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_164", + "layer_object_name": "compute_graph.flexml_layers[85]", + "layer_order": 99, + "mlir_dag_id_map": { + "dag_layer": 99, + "mlir_layer": 99 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[85].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[85].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[85].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[85].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[85].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[85].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[85].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[85].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[85].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[85].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[85].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[85].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[85].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[85].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[85].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[85].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 221184, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_99" + ] + }, + "super_iter": 1 + }, + "0100": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_99" + ], + "out": [ + "compute_graph.l2_100" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[86].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[86].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[86].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[86].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[86].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[86].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[86].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[86].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[86].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[86].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[86].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[86].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[86].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[86].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[86].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[86].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 221184, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_99" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_99", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_167", + "layer_object_name": "compute_graph.flexml_layers[86]", + "layer_order": 100, + "mlir_dag_id_map": { + "dag_layer": 100, + "mlir_layer": 100 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[86].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[86].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[86].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[86].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[86].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[86].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[86].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[86].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[86].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[86].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[86].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[86].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[86].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[86].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[86].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[86].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_100" + ] + }, + "super_iter": 1 + }, + "0101": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_100" + ], + "out": [ + "compute_graph.l2_101" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[87].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[87].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[87].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[87].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[87].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[87].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[87].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[87].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[87].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[87].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[87].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[87].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[87].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[87].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[87].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[87].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_100" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_100", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_169", + "layer_object_name": "compute_graph.flexml_layers[87]", + "layer_order": 101, + "mlir_dag_id_map": { + "dag_layer": 101, + "mlir_layer": 101 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[87].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[87].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[87].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[87].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[87].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[87].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[87].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[87].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[87].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[87].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[87].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[87].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[87].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[87].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[87].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[87].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 221184, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_101" + ] + }, + "super_iter": 1 + }, + "0102": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_101", + "compute_graph.l2_98" + ], + "out": [ + "compute_graph.l2_102" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[88].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[88].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[88].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[88].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[88].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[88].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[88].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[88].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[88].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[88].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[88].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[88].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[88].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[88].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[88].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[88].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73728, + 73728, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_98" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[88].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[88].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[88].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[88].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[88].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[88].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[88].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[88].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[88].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[88].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[88].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[88].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[88].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[88].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[88].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[88].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 221184, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_101" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_101", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_98", + "dest_port": 0, + "src_offset": 4 + } + ], + "layer_name": "Mul_170", + "layer_object_name": "compute_graph.flexml_layers[88]", + "layer_order": 102, + "mlir_dag_id_map": { + "dag_layer": 102, + "mlir_layer": 102 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[88].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[88].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[88].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[88].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[88].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[88].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[88].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[88].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[88].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[88].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[88].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[88].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[88].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[88].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[88].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[88].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_102" + ] + }, + "super_iter": 1 + }, + "0103": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_92", + "compute_graph.l2_102" + ], + "out": [ + "compute_graph.l2_103" + ], + "wts": [ + "compute_graph.Layer_103_l2_wts", + "compute_graph.Layer_103_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[89].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[89].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[89].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[89].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[89].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[89].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[89].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[89].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[89].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[89].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[89].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[89].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[89].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[89].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[89].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[89].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3072 + ], + "l1_pong": [ + "0xcd80", + 3072 + ], + "l2": [ + [ + 2, + 0, + 432128, + 46080, + 44160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_102" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[89].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[89].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[89].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[89].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[89].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[89].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[89].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[89].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[89].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[89].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[89].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[89].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[89].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[89].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[89].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[89].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 3072 + ], + "l1_pong": [ + "0x6000", + 3072 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_92" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_102", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_92", + "dest_port": 2, + "src_offset": 4 + } + ], + "layer_name": "Conv_171", + "layer_object_name": "compute_graph.flexml_layers[89]", + "layer_order": 103, + "mlir_dag_id_map": { + "dag_layer": 103, + "mlir_layer": 103 + }, + "num_iter": 6, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[89].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[89].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[89].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[89].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[89].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[89].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[89].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[89].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[89].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[89].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[89].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[89].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[89].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[89].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[89].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[89].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_103" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[89].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[89].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[89].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[89].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[89].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[89].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[89].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[89].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[89].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[89].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[89].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[89].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[89].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[89].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[89].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[89].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe580", + 2400 + ], + "l1_pong": [ + "0x9800", + 2400 + ], + "l2": [ + [ + 3, + 0, + 0, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_103_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 570880, + "size": 19200 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_103_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0104": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_103" + ], + "out": [ + "compute_graph.l2_104" + ], + "wts": [ + "compute_graph.Layer_104_l2_wts", + "compute_graph.Layer_104_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[90].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[90].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[90].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[90].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[90].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[90].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[90].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[90].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[90].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[90].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[90].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[90].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[90].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[90].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[90].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[90].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2560 + ], + "l1_pong": [ + "0xcd80", + 2560 + ], + "l2": [ + [ + 0, + 0, + 0, + 36864, + 19200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_103" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_103", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_173", + "layer_object_name": "compute_graph.flexml_layers[90]", + "layer_order": 104, + "mlir_dag_id_map": { + "dag_layer": 104, + "mlir_layer": 104 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[90].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[90].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[90].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[90].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[90].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[90].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[90].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[90].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[90].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[90].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[90].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[90].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[90].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[90].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[90].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[90].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1280 + ], + "l1_pong": [ + "0x4000", + 1280 + ], + "l2": [ + [ + 2, + 0, + 155648, + 184320, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_104" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[90].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[90].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[90].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[90].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[90].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[90].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[90].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[90].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[90].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[90].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[90].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[90].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[90].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[90].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[90].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[90].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe180", + 3360 + ], + "l1_pong": [ + "0x9400", + 3360 + ], + "l2": [ + [ + 3, + 0, + 443648, + 40320 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_104_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 609280, + "size": 40320 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_104_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0105": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_104" + ], + "out": [ + "compute_graph.l2_105" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[91].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[91].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[91].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[91].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[91].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[91].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[91].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[91].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[91].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[91].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[91].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[91].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[91].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[91].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[91].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[91].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 155648, + 184320, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_104" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_104", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_175", + "layer_object_name": "compute_graph.flexml_layers[91]", + "layer_order": 105, + "mlir_dag_id_map": { + "dag_layer": 105, + "mlir_layer": 105 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[91].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[91].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[91].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[91].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[91].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[91].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[91].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[91].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[91].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[91].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[91].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[91].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[91].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[91].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[91].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[91].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_105" + ] + }, + "super_iter": 1 + }, + "0106": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_105" + ], + "out": [ + "compute_graph.l2_106" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[92].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[92].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[92].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[92].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[92].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[92].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[92].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[92].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[92].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[92].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[92].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[92].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[92].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[92].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[92].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[92].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_105" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_105", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_178", + "layer_object_name": "compute_graph.flexml_layers[92]", + "layer_order": 106, + "mlir_dag_id_map": { + "dag_layer": 106, + "mlir_layer": 106 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[92].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[92].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[92].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[92].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[92].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[92].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[92].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[92].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[92].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[92].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[92].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[92].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[92].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[92].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[92].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[92].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 1, + 0, + 434176, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_106" + ] + }, + "super_iter": 1 + }, + "0107": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_106" + ], + "out": [ + "compute_graph.l2_107" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[93].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[93].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[93].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[93].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[93].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[93].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[93].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[93].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[93].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[93].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[93].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[93].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[93].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[93].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[93].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[93].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 1, + 0, + 434176, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_106" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_106", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_180", + "layer_object_name": "compute_graph.flexml_layers[93]", + "layer_order": 107, + "mlir_dag_id_map": { + "dag_layer": 107, + "mlir_layer": 107 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[93].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[93].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[93].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[93].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[93].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[93].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[93].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[93].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[93].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[93].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[93].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[93].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[93].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[93].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[93].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[93].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_107" + ] + }, + "super_iter": 1 + }, + "0108": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_104", + "compute_graph.l2_107" + ], + "out": [ + "compute_graph.l2_108" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[94].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[94].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[94].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[94].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[94].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[94].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[94].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[94].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[94].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[94].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[94].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[94].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[94].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[94].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[94].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[94].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 155648, + 184320, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_104" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[94].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[94].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[94].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[94].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[94].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[94].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[94].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[94].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[94].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[94].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[94].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[94].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[94].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[94].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[94].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[94].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_107" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_104", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_107", + "dest_port": 2, + "src_offset": 0 + } + ], + "layer_name": "Mul_181", + "layer_object_name": "compute_graph.flexml_layers[94]", + "layer_order": 108, + "mlir_dag_id_map": { + "dag_layer": 108, + "mlir_layer": 108 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[94].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[94].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[94].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[94].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[94].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[94].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[94].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[94].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[94].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[94].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[94].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[94].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[94].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[94].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[94].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[94].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 245760, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_108" + ] + }, + "super_iter": 1 + }, + "0109": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_108" + ], + "out": [ + "compute_graph.l2_109" + ], + "wts": [ + "compute_graph.Layer_109_l2_wts", + "compute_graph.Layer_109_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[95].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[95].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[95].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[95].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[95].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[95].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[95].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[95].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[95].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[95].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[95].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[95].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[95].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[95].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[95].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[95].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5376 + ], + "l1_pong": [ + "0xcd80", + 5376 + ], + "l2": [ + [ + 0, + 0, + 245760, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_108" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_108", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_182", + "layer_object_name": "compute_graph.flexml_layers[95]", + "layer_order": 109, + "mlir_dag_id_map": { + "dag_layer": 109, + "mlir_layer": 109 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[95].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[95].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[95].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[95].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[95].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[95].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[95].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[95].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[95].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[95].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[95].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[95].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[95].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[95].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[95].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[95].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 491520, + 184320, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_109" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[95].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[95].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[95].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[95].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[95].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[95].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[95].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[95].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[95].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[95].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[95].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[95].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[95].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[95].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[95].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[95].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf780", + 128 + ], + "l1_pong": [ + "0xaa00", + 128 + ], + "l2": [ + [ + 3, + 0, + 0, + 7680 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_109_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 689920, + "size": 7680 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_109_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0110": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_109" + ], + "out": [ + "compute_graph.l2_110" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[96].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[96].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[96].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[96].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[96].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[96].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[96].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[96].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[96].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[96].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[96].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[96].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[96].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[96].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[96].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[96].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 491520, + 184320, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_109" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_109", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_184", + "layer_object_name": "compute_graph.flexml_layers[96]", + "layer_order": 110, + "mlir_dag_id_map": { + "dag_layer": 110, + "mlir_layer": 110 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[96].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[96].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[96].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[96].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[96].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[96].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[96].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[96].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[96].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[96].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[96].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[96].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[96].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[96].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[96].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[96].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_110" + ] + }, + "super_iter": 1 + }, + "0111": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_110" + ], + "out": [ + "compute_graph.l2_111" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[97].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[97].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[97].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[97].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[97].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[97].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[97].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[97].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[97].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[97].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[97].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[97].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[97].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[97].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[97].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[97].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_110" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_110", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_187", + "layer_object_name": "compute_graph.flexml_layers[97]", + "layer_order": 111, + "mlir_dag_id_map": { + "dag_layer": 111, + "mlir_layer": 111 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[97].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[97].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[97].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[97].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[97].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[97].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[97].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[97].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[97].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[97].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[97].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[97].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[97].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[97].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[97].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[97].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_111" + ] + }, + "super_iter": 1 + }, + "0112": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_111" + ], + "out": [ + "compute_graph.l2_112" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[98].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[98].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[98].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[98].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[98].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[98].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[98].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[98].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[98].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[98].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[98].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[98].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[98].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[98].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[98].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[98].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_111" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_111", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_189", + "layer_object_name": "compute_graph.flexml_layers[98]", + "layer_order": 112, + "mlir_dag_id_map": { + "dag_layer": 112, + "mlir_layer": 112 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[98].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[98].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[98].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[98].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[98].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[98].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[98].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[98].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[98].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[98].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[98].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[98].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[98].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[98].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[98].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[98].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_112" + ] + }, + "super_iter": 1 + }, + "0113": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_109", + "compute_graph.l2_112" + ], + "out": [ + "compute_graph.l2_113", + "compute_graph.l2l3_113_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[99].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[99].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[99].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[99].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[99].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[99].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[99].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[99].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[99].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[99].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[99].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[99].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[99].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[99].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[99].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[99].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 491520, + 184320, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_109" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[99].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[99].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[99].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[99].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[99].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[99].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[99].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[99].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[99].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[99].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[99].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[99].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[99].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[99].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[99].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[99].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_112" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_109", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_112", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_113.out[0]", + "to": "compute_graph.l2l3_113_spill.in[0]" + }, + { + "from": "compute_graph.l2_113.out[1]", + "to": "compute_graph.l2l3_113_spill.in[1]" + }, + { + "from": "compute_graph.l2_113.out[2]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.in[0]" + }, + { + "from": "compute_graph.l2_113.out[3]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.in[1]" + } + ], + "layer_name": "Mul_190", + "layer_object_name": "compute_graph.flexml_layers[99]", + "layer_order": 113, + "mlir_dag_id_map": { + "dag_layer": 113, + "mlir_layer": 113 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[99].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[99].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[99].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[99].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[99].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[99].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[99].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[99].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[99].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[99].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[99].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[99].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[99].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[99].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[99].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[99].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 245760, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_113" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_113113": { + "size": 115200 + }, + "l2l3_113_spill": { + "size": 115200 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_113_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113" + ] + }, + "super_iter": 1 + }, + "0114": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_113_spill" + ], + "out": [ + "compute_graph.l2l3_114_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_113_spill": { + "size": 115200 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_113_spill" + ] + }, + "kernel_name": [ + "Transpose4dAdf" + ], + "layer_name": "Generated-#24", + "layer_object_name": "compute_graph.templated_graph_114", + "layer_order": 114, + "mlir_dag_id_map": { + "dag_layer": 114, + "mlir_layer": 114 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_114_spill": { + "size": 115200 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_114_spill" + ] + }, + "templated_graph": 1 + }, + "0115": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0", + "compute_graph.l2l3_114_spill" + ], + "out": [ + "compute_graph.l2_115" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[100].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[100].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[100].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[100].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[100].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[100].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[100].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[100].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[100].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[100].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[100].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[100].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[100].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[100].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[100].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[100].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 0, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_114_spill" + ] + }, + "kernel_name": [ + "superkernel_reduce_mean_c8" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_114_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_114_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_114_for_layer_115_port0.in[1]" + } + ], + "layer_name": "Generated-#26", + "layer_object_name": "compute_graph.flexml_layers[100]", + "layer_order": 115, + "mlir_dag_id_map": { + "dag_layer": 115, + "mlir_layer": 115 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[100].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[100].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[100].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[100].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[100].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[100].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[100].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[100].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[100].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[100].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[100].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[100].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[100].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[100].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[100].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[100].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb000", + 40 + ], + "l1_pong": [ + "0xcd80", + 40 + ], + "l2": [ + [ + 2, + 0, + 520448, + 1920, + 480 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_115" + ] + }, + "super_iter": 1 + }, + "0116": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_115" + ], + "out": [ + "compute_graph.l2_116" + ], + "wts": [ + "compute_graph.Layer_116_l2_wts", + "compute_graph.Layer_116_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 4, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[101].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[101].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[101].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[101].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[101].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[101].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[101].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[101].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[101].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[101].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[101].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[101].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[101].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[101].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[101].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[101].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 520448, + 1920, + 480 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_115" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_115", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_192", + "layer_object_name": "compute_graph.flexml_layers[101]", + "layer_order": 116, + "mlir_dag_id_map": { + "dag_layer": 116, + "mlir_layer": 116 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[101].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[101].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[101].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[101].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[101].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[101].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[101].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[101].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[101].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[101].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[101].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[101].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[101].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[101].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[101].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[101].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 0, + 0, + 0, + 4096, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_116" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[101].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[101].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[101].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[101].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[101].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[101].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[101].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[101].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[101].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[101].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[101].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[101].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[101].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[101].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[101].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[101].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xd500", + 3968 + ], + "l1_pong": [ + "0x8780", + 3968 + ], + "l2": [ + [ + 3, + 0, + 0, + 63488 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_116_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 705280, + "size": 63488 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_116_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0117": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_116" + ], + "out": [ + "compute_graph.l2_117" + ], + "wts": [ + "compute_graph.Layer_117_l2_wts", + "compute_graph.Layer_117_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[102].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[102].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[102].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[102].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[102].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[102].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[102].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[102].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[102].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[102].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[102].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[102].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[102].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[102].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[102].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[102].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 4096, + 120 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_116" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_116", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_194", + "layer_object_name": "compute_graph.flexml_layers[102]", + "layer_order": 117, + "mlir_dag_id_map": { + "dag_layer": 117, + "mlir_layer": 117 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[102].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[102].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[102].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[102].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[102].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[102].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[102].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[102].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[102].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[102].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[102].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[102].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[102].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[102].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[102].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[102].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 2, + 0, + 491520, + 16384, + 480 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_117" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[102].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[102].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[102].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[102].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[102].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[102].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[102].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[102].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[102].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[102].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[102].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[102].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[102].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[102].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[102].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[102].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xd500", + 3968 + ], + "l1_pong": [ + "0x8780", + 3968 + ], + "l2": [ + [ + 3, + 0, + 397312, + 63488 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_117_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 832256, + "size": 63488 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_117_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0118": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_117" + ], + "out": [ + "compute_graph.l2_118" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[103].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[103].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[103].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[103].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[103].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[103].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[103].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[103].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[103].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[103].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[103].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[103].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[103].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[103].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[103].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[103].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 2, + 0, + 491520, + 16384, + 480 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_117" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_117", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_196", + "layer_object_name": "compute_graph.flexml_layers[103]", + "layer_order": 118, + "mlir_dag_id_map": { + "dag_layer": 118, + "mlir_layer": 118 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[103].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[103].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[103].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[103].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[103].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[103].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[103].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[103].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[103].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[103].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[103].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[103].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[103].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[103].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[103].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[103].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaf80", + 384 + ], + "l1_pong": [ + "0xcd80", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 480 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_118" + ] + }, + "super_iter": 1 + }, + "0119": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_118" + ], + "out": [ + "compute_graph.l2_119" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[104].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[104].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[104].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[104].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[104].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[104].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[104].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[104].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[104].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[104].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[104].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[104].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[104].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[104].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[104].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[104].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 480 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_118" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_118", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_199", + "layer_object_name": "compute_graph.flexml_layers[104]", + "layer_order": 119, + "mlir_dag_id_map": { + "dag_layer": 119, + "mlir_layer": 119 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[104].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[104].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[104].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[104].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[104].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[104].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[104].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[104].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[104].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[104].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[104].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[104].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[104].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[104].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[104].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[104].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 256 + ], + "l1_pong": [ + "0xcd80", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 480 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_119" + ] + }, + "super_iter": 1 + }, + "0120": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_119" + ], + "out": [ + "compute_graph.l2_120", + "compute_graph.l2l3_120_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[105].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[105].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[105].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[105].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[105].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[105].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[105].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[105].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[105].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[105].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[105].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[105].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[105].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[105].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[105].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[105].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 480 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_119" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_119", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_120.out[0]", + "to": "compute_graph.l2l3_120_spill.in[0]" + }, + { + "from": "compute_graph.l2_120.out[1]", + "to": "compute_graph.l2l3_120_spill.in[1]" + } + ], + "layer_name": "Div_201", + "layer_object_name": "compute_graph.flexml_layers[105]", + "layer_order": 120, + "mlir_dag_id_map": { + "dag_layer": 120, + "mlir_layer": 120 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[105].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[105].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[105].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[105].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[105].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[105].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[105].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[105].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[105].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[105].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[105].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[105].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[105].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[105].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[105].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[105].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xae80", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 8192, + 480 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_120" + ], + "l3": { + "l2l3_120_spill": { + "size": 480 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_120_spill" + ] + }, + "super_iter": 1 + }, + "0121": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_120_spill", + "compute_graph.l2l3_scratch_0_121_spill", + "compute_graph.l2l3_scratch_1_121_spill" + ], + "out": [ + "compute_graph.l2l3_121_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_120_spill": { + "size": 480 + }, + "l2l3_scratch_0_121_spill": { + "size": 230400 + }, + "l2l3_scratch_1_121_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_120_spill", + "compute_graph.l2l3_scratch_0_121_spill", + "compute_graph.l2l3_scratch_1_121_spill" + ] + }, + "kernel_name": [ + "TileAdf" + ], + "layer_name": "Generated-#28", + "layer_object_name": "compute_graph.templated_graph_121", + "layer_order": 121, + "mlir_dag_id_map": { + "dag_layer": 121, + "mlir_layer": 121 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_121_spill": { + "size": 115200 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_121_spill" + ] + }, + "templated_graph": 1 + }, + "0122": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0", + "compute_graph.l2l3_121_spill", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113" + ], + "out": [ + "compute_graph.l2_122" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[106].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[106].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[106].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[106].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[106].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[106].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[106].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[106].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[106].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[106].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[106].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[106].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[106].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[106].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[106].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[106].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_121_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[106].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[106].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[106].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[106].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[106].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[106].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[106].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[106].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[106].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[106].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[106].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[106].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[106].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[106].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[106].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[106].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 230400, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_113113.in[1]" + }, + { + "from": "compute_graph.l2l3_121_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_121_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_121_for_layer_122_port0.in[1]" + } + ], + "layer_name": "Mul_202", + "layer_object_name": "compute_graph.flexml_layers[106]", + "layer_order": 122, + "mlir_dag_id_map": { + "dag_layer": 122, + "mlir_layer": 122 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[106].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[106].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[106].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[106].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[106].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[106].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[106].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[106].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[106].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[106].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[106].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[106].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[106].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[106].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[106].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[106].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_122" + ] + }, + "super_iter": 1 + }, + "0123": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_122" + ], + "out": [ + "compute_graph.l2_123", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123" + ], + "wts": [ + "compute_graph.Layer_123_l2_wts", + "compute_graph.Layer_123_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[107].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[107].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[107].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[107].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[107].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[107].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[107].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[107].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[107].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[107].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[107].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[107].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[107].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[107].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[107].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[107].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3072 + ], + "l1_pong": [ + "0xcd80", + 3072 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_122" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_122", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_123.out[4]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.in[0]" + }, + { + "from": "compute_graph.l2_123.out[5]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.in[1]" + } + ], + "layer_name": "Conv_203", + "layer_object_name": "compute_graph.flexml_layers[107]", + "layer_order": 123, + "mlir_dag_id_map": { + "dag_layer": 123, + "mlir_layer": 123 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[107].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[107].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[107].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[107].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[107].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[107].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[107].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[107].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[107].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[107].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[107].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[107].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[107].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[107].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[107].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[107].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 49152, + 26880 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_123" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_123123": { + "size": 26880 + } + }, + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[107].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[107].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[107].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[107].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[107].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[107].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[107].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[107].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[107].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[107].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[107].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[107].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[107].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[107].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[107].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[107].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe580", + 3200 + ], + "l1_pong": [ + "0x9800", + 3200 + ], + "l2": [ + [ + 3, + 0, + 0, + 64000 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_123_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 959232, + "size": 64000 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_123_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0124": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_123" + ], + "out": [ + "compute_graph.l2_124" + ], + "wts": [ + "compute_graph.Layer_124_l2_wts", + "compute_graph.Layer_124_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[108].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[108].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[108].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[108].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[108].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[108].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[108].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[108].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[108].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[108].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[108].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[108].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[108].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[108].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[108].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[108].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2048 + ], + "l1_pong": [ + "0xcd80", + 2048 + ], + "l2": [ + [ + 0, + 0, + 0, + 49152, + 26880 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_123" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_123", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_204", + "layer_object_name": "compute_graph.flexml_layers[108]", + "layer_order": 124, + "mlir_dag_id_map": { + "dag_layer": 124, + "mlir_layer": 124 + }, + "num_iter": 18, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[108].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[108].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[108].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[108].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[108].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[108].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[108].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[108].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[108].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[108].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[108].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[108].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[108].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[108].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[108].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[108].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1792 + ], + "l1_pong": [ + "0x4000", + 1792 + ], + "l2": [ + [ + 2, + 0, + 8192, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_124" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[108].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[108].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[108].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[108].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[108].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[108].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[108].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[108].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[108].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[108].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[108].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[108].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[108].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[108].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[108].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[108].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xdd80", + 3808 + ], + "l1_pong": [ + "0x9000", + 3808 + ], + "l2": [ + [ + 3, + 0, + 341504, + 91392 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_124_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 1087232, + "size": 91392 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_124_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0125": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_124" + ], + "out": [ + "compute_graph.l2_125" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[109].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[109].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[109].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[109].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[109].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[109].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[109].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[109].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[109].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[109].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[109].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[109].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[109].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[109].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[109].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[109].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 8192, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_124" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_124", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_206", + "layer_object_name": "compute_graph.flexml_layers[109]", + "layer_order": 125, + "mlir_dag_id_map": { + "dag_layer": 125, + "mlir_layer": 125 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[109].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[109].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[109].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[109].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[109].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[109].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[109].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[109].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[109].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[109].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[109].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[109].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[109].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[109].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[109].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[109].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_125" + ] + }, + "super_iter": 1 + }, + "0126": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_125" + ], + "out": [ + "compute_graph.l2_126" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[110].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[110].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[110].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[110].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[110].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[110].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[110].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[110].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[110].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[110].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[110].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[110].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[110].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[110].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[110].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[110].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_125" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_125", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_209", + "layer_object_name": "compute_graph.flexml_layers[110]", + "layer_order": 126, + "mlir_dag_id_map": { + "dag_layer": 126, + "mlir_layer": 126 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[110].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[110].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[110].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[110].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[110].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[110].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[110].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[110].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[110].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[110].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[110].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[110].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[110].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[110].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[110].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[110].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 1, + 0, + 194560, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_126" + ] + }, + "super_iter": 1 + }, + "0127": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_126" + ], + "out": [ + "compute_graph.l2_127" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[111].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[111].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[111].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[111].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[111].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[111].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[111].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[111].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[111].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[111].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[111].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[111].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[111].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[111].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[111].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[111].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 1, + 0, + 194560, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_126" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_126", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_211", + "layer_object_name": "compute_graph.flexml_layers[111]", + "layer_order": 127, + "mlir_dag_id_map": { + "dag_layer": 127, + "mlir_layer": 127 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[111].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[111].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[111].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[111].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[111].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[111].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[111].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[111].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[111].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[111].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[111].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[111].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[111].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[111].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[111].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[111].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_127" + ] + }, + "super_iter": 1 + }, + "0128": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_124", + "compute_graph.l2_127" + ], + "out": [ + "compute_graph.l2_128" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[112].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[112].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[112].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[112].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[112].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[112].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[112].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[112].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[112].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[112].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[112].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[112].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[112].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[112].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[112].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[112].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 8192, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_124" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[112].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[112].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[112].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[112].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[112].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[112].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[112].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[112].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[112].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[112].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[112].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[112].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[112].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[112].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[112].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[112].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_127" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_124", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_127", + "dest_port": 2, + "src_offset": 0 + } + ], + "layer_name": "Mul_212", + "layer_object_name": "compute_graph.flexml_layers[112]", + "layer_order": 128, + "mlir_dag_id_map": { + "dag_layer": 128, + "mlir_layer": 128 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[112].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[112].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[112].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[112].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[112].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[112].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[112].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[112].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[112].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[112].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[112].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[112].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[112].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[112].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[112].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[112].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 337920, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_128" + ] + }, + "super_iter": 1 + }, + "0129": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_128" + ], + "out": [ + "compute_graph.l2_129" + ], + "wts": [ + "compute_graph.Layer_129_l2_wts", + "compute_graph.Layer_129_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[113].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[113].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[113].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[113].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[113].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[113].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[113].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[113].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[113].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[113].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[113].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[113].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[113].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[113].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[113].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[113].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5376 + ], + "l1_pong": [ + "0xcd80", + 5376 + ], + "l2": [ + [ + 0, + 0, + 337920, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_128" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_128", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_213", + "layer_object_name": "compute_graph.flexml_layers[113]", + "layer_order": 129, + "mlir_dag_id_map": { + "dag_layer": 129, + "mlir_layer": 129 + }, + "num_iter": 21, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[113].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[113].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[113].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[113].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[113].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[113].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[113].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[113].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[113].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[113].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[113].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[113].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[113].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[113].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[113].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[113].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 1, + 0, + 151552, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_129" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[113].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[113].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[113].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[113].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[113].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[113].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[113].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[113].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[113].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[113].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[113].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[113].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[113].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[113].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[113].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[113].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf780", + 128 + ], + "l1_pong": [ + "0xaa00", + 128 + ], + "l2": [ + [ + 3, + 0, + 0, + 10752 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_129_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 1270016, + "size": 10752 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_129_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0130": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_129" + ], + "out": [ + "compute_graph.l2_130" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[114].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[114].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[114].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[114].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[114].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[114].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[114].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[114].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[114].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[114].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[114].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[114].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[114].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[114].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[114].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[114].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 1, + 0, + 151552, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_129" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_129", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_215", + "layer_object_name": "compute_graph.flexml_layers[114]", + "layer_order": 130, + "mlir_dag_id_map": { + "dag_layer": 130, + "mlir_layer": 130 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[114].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[114].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[114].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[114].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[114].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[114].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[114].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[114].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[114].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[114].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[114].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[114].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[114].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[114].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[114].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[114].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_130" + ] + }, + "super_iter": 1 + }, + "0131": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_130" + ], + "out": [ + "compute_graph.l2_131" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[115].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[115].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[115].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[115].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[115].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[115].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[115].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[115].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[115].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[115].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[115].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[115].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[115].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[115].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[115].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[115].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_130" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_130", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_218", + "layer_object_name": "compute_graph.flexml_layers[115]", + "layer_order": 131, + "mlir_dag_id_map": { + "dag_layer": 131, + "mlir_layer": 131 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[115].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[115].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[115].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[115].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[115].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[115].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[115].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[115].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[115].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[115].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[115].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[115].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[115].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[115].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[115].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[115].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 186368, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_131" + ] + }, + "super_iter": 1 + }, + "0132": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_131" + ], + "out": [ + "compute_graph.l2_132" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[116].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[116].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[116].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[116].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[116].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[116].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[116].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[116].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[116].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[116].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[116].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[116].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[116].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[116].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[116].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[116].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 186368, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_131" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_131", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_220", + "layer_object_name": "compute_graph.flexml_layers[116]", + "layer_order": 132, + "mlir_dag_id_map": { + "dag_layer": 132, + "mlir_layer": 132 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[116].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[116].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[116].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[116].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[116].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[116].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[116].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[116].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[116].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[116].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[116].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[116].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[116].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[116].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[116].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[116].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_132" + ] + }, + "super_iter": 1 + }, + "0133": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_129", + "compute_graph.l2_132" + ], + "out": [ + "compute_graph.l2_133", + "compute_graph.l2l3_133_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[117].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[117].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[117].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[117].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[117].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[117].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[117].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[117].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[117].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[117].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[117].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[117].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[117].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[117].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[117].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[117].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 1, + 0, + 151552, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_129" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[117].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[117].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[117].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[117].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[117].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[117].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[117].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[117].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[117].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[117].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[117].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[117].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[117].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[117].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[117].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[117].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_132" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_129", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_132", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_133.out[0]", + "to": "compute_graph.l2l3_133_spill.in[0]" + }, + { + "from": "compute_graph.l2_133.out[1]", + "to": "compute_graph.l2l3_133_spill.in[1]" + }, + { + "from": "compute_graph.l2_133.out[2]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.in[0]" + }, + { + "from": "compute_graph.l2_133.out[3]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.in[1]" + } + ], + "layer_name": "Mul_221", + "layer_object_name": "compute_graph.flexml_layers[117]", + "layer_order": 133, + "mlir_dag_id_map": { + "dag_layer": 133, + "mlir_layer": 133 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[117].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[117].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[117].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[117].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[117].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[117].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[117].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[117].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[117].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[117].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[117].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[117].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[117].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[117].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[117].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[117].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 337920, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_133" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_133133": { + "size": 161280 + }, + "l2l3_133_spill": { + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_133_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133" + ] + }, + "super_iter": 1 + }, + "0134": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_133_spill" + ], + "out": [ + "compute_graph.l2l3_134_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_133_spill": { + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_133_spill" + ] + }, + "kernel_name": [ + "Transpose4dAdf" + ], + "layer_name": "Generated-#30", + "layer_object_name": "compute_graph.templated_graph_134", + "layer_order": 134, + "mlir_dag_id_map": { + "dag_layer": 134, + "mlir_layer": 134 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_134_spill": { + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_134_spill" + ] + }, + "templated_graph": 1 + }, + "0135": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0", + "compute_graph.l2l3_134_spill" + ], + "out": [ + "compute_graph.l2_135" + ] + }, + "buffer_iter": 1, + "depth_iter": 8, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[118].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[118].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[118].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[118].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[118].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[118].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[118].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[118].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[118].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[118].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[118].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[118].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[118].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[118].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[118].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[118].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7168 + ], + "l1_pong": [ + "0x4000", + 7168 + ], + "l2": [ + [ + 0, + 0, + 0, + 161280, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_134_spill" + ] + }, + "kernel_name": [ + "superkernel_reduce_mean_c8" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_134_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_134_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_134_for_layer_135_port0.in[1]" + } + ], + "layer_name": "Generated-#32", + "layer_object_name": "compute_graph.flexml_layers[118]", + "layer_order": 135, + "mlir_dag_id_map": { + "dag_layer": 135, + "mlir_layer": 135 + }, + "num_iter": 24, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[118].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[118].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[118].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[118].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[118].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[118].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[118].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[118].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[118].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[118].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[118].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[118].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[118].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[118].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[118].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[118].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaf00", + 56 + ], + "l1_pong": [ + "0xcd80", + 56 + ], + "l2": [ + [ + 2, + 0, + 518912, + 2688, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_135" + ] + }, + "super_iter": 1 + }, + "0136": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_135" + ], + "out": [ + "compute_graph.l2_136" + ], + "wts": [ + "compute_graph.Layer_136_l2_wts", + "compute_graph.Layer_136_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 7, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[119].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[119].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[119].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[119].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[119].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[119].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[119].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[119].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[119].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[119].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[119].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[119].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[119].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[119].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[119].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[119].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 768 + ], + "l1_pong": [ + "0xcd80", + 768 + ], + "l2": [ + [ + 2, + 0, + 518912, + 2688, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_135" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_135", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_223", + "layer_object_name": "compute_graph.flexml_layers[119]", + "layer_order": 136, + "mlir_dag_id_map": { + "dag_layer": 136, + "mlir_layer": 136 + }, + "num_iter": 7, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[119].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[119].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[119].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[119].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[119].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[119].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[119].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[119].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[119].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[119].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[119].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[119].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[119].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[119].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[119].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[119].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 384 + ], + "l1_pong": [ + "0x4000", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 168 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_136" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[119].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[119].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[119].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[119].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[119].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[119].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[119].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[119].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[119].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[119].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[119].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[119].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[119].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[119].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[119].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[119].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xd380", + 4800 + ], + "l1_pong": [ + "0x8600", + 4800 + ], + "l2": [ + [ + 3, + 0, + 0, + 134400 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_136_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 1291520, + "size": 134400 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_136_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0137": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_136" + ], + "out": [ + "compute_graph.l2_137" + ], + "wts": [ + "compute_graph.Layer_137_l2_wts", + "compute_graph.Layer_137_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[120].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[120].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[120].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[120].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[120].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[120].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[120].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[120].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[120].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[120].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[120].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[120].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[120].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[120].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[120].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[120].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2688 + ], + "l1_pong": [ + "0xcd80", + 2688 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 168 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_136" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_136", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_225", + "layer_object_name": "compute_graph.flexml_layers[120]", + "layer_order": 137, + "mlir_dag_id_map": { + "dag_layer": 137, + "mlir_layer": 137 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[120].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[120].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[120].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[120].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[120].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[120].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[120].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[120].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[120].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[120].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[120].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[120].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[120].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[120].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[120].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[120].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 2, + 0, + 434176, + 45056, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_137" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[120].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[120].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[120].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[120].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[120].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[120].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[120].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[120].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[120].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[120].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[120].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[120].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[120].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[120].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[120].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[120].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe280", + 2752 + ], + "l1_pong": [ + "0x9500", + 2752 + ], + "l2": [ + [ + 3, + 0, + 282112, + 121088 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_137_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 1560320, + "size": 121088 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_137_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0138": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_137" + ], + "out": [ + "compute_graph.l2_138" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[121].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[121].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[121].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[121].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[121].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[121].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[121].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[121].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[121].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[121].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[121].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[121].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[121].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[121].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[121].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[121].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 2, + 0, + 434176, + 45056, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_137" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_137", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_227", + "layer_object_name": "compute_graph.flexml_layers[121]", + "layer_order": 138, + "mlir_dag_id_map": { + "dag_layer": 138, + "mlir_layer": 138 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[121].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[121].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[121].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[121].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[121].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[121].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[121].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[121].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[121].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[121].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[121].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[121].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[121].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[121].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[121].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[121].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaf80", + 384 + ], + "l1_pong": [ + "0xcd80", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_138" + ] + }, + "super_iter": 1 + }, + "0139": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_138" + ], + "out": [ + "compute_graph.l2_139" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[122].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[122].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[122].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[122].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[122].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[122].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[122].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[122].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[122].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[122].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[122].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[122].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[122].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[122].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[122].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[122].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_138" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_138", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_230", + "layer_object_name": "compute_graph.flexml_layers[122]", + "layer_order": 139, + "mlir_dag_id_map": { + "dag_layer": 139, + "mlir_layer": 139 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[122].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[122].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[122].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[122].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[122].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[122].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[122].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[122].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[122].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[122].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[122].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[122].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[122].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[122].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[122].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[122].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 256 + ], + "l1_pong": [ + "0xcd80", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_139" + ] + }, + "super_iter": 1 + }, + "0140": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_139" + ], + "out": [ + "compute_graph.l2_140", + "compute_graph.l2l3_140_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[123].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[123].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[123].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[123].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[123].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[123].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[123].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[123].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[123].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[123].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[123].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[123].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[123].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[123].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[123].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[123].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_139" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_139", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_140.out[0]", + "to": "compute_graph.l2l3_140_spill.in[0]" + }, + { + "from": "compute_graph.l2_140.out[1]", + "to": "compute_graph.l2l3_140_spill.in[1]" + } + ], + "layer_name": "Div_232", + "layer_object_name": "compute_graph.flexml_layers[123]", + "layer_order": 140, + "mlir_dag_id_map": { + "dag_layer": 140, + "mlir_layer": 140 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[123].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[123].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[123].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[123].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[123].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[123].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[123].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[123].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[123].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[123].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[123].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[123].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[123].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[123].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[123].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[123].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xae80", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 8192, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_140" + ], + "l3": { + "l2l3_140_spill": { + "size": 672 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_140_spill" + ] + }, + "super_iter": 1 + }, + "0141": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_140_spill", + "compute_graph.l2l3_scratch_0_141_spill", + "compute_graph.l2l3_scratch_1_141_spill" + ], + "out": [ + "compute_graph.l2l3_141_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_140_spill": { + "size": 672 + }, + "l2l3_scratch_0_141_spill": { + "size": 322560 + }, + "l2l3_scratch_1_141_spill": { + "size": 322560 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_140_spill", + "compute_graph.l2l3_scratch_0_141_spill", + "compute_graph.l2l3_scratch_1_141_spill" + ] + }, + "kernel_name": [ + "TileAdf" + ], + "layer_name": "Generated-#34", + "layer_object_name": "compute_graph.templated_graph_141", + "layer_order": 141, + "mlir_dag_id_map": { + "dag_layer": 141, + "mlir_layer": 141 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_141_spill": { + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_141_spill" + ] + }, + "templated_graph": 1 + }, + "0142": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0", + "compute_graph.l2l3_141_spill", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133" + ], + "out": [ + "compute_graph.l2_142" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[124].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[124].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[124].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[124].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[124].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[124].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[124].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[124].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[124].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[124].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[124].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[124].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[124].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[124].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[124].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[124].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 161280, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_141_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[124].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[124].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[124].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[124].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[124].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[124].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[124].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[124].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[124].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[124].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[124].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[124].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[124].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[124].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[124].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[124].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 322560, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_133133.in[1]" + }, + { + "from": "compute_graph.l2l3_141_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_141_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_141_for_layer_142_port0.in[1]" + } + ], + "layer_name": "Mul_233", + "layer_object_name": "compute_graph.flexml_layers[124]", + "layer_order": 142, + "mlir_dag_id_map": { + "dag_layer": 142, + "mlir_layer": 142 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[124].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[124].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[124].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[124].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[124].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[124].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[124].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[124].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[124].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[124].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[124].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[124].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[124].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[124].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[124].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[124].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 186368, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_142" + ] + }, + "super_iter": 1 + }, + "0143": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_142", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123" + ], + "out": [ + "compute_graph.l2_143" + ], + "wts": [ + "compute_graph.Layer_143_l2_wts", + "compute_graph.Layer_143_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 7, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[125].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[125].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[125].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[125].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[125].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[125].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[125].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[125].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[125].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[125].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[125].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[125].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[125].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[125].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[125].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[125].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3072 + ], + "l1_pong": [ + "0xcd80", + 3072 + ], + "l2": [ + [ + 2, + 0, + 186368, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_142" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[125].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[125].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[125].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[125].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[125].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[125].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[125].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[125].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[125].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[125].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[125].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[125].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[125].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[125].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[125].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[125].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 4096 + ], + "l1_pong": [ + "0x6000", + 4096 + ], + "l2": [ + [ + 0, + 0, + 0, + 49152, + 26880 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_142", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_123123.in[1]" + } + ], + "layer_name": "Conv_234", + "layer_object_name": "compute_graph.flexml_layers[125]", + "layer_order": 143, + "mlir_dag_id_map": { + "dag_layer": 143, + "mlir_layer": 143 + }, + "num_iter": 21, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[125].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[125].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[125].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[125].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[125].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[125].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[125].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[125].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[125].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[125].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[125].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[125].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[125].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[125].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[125].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[125].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 49152, + 26880 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_143" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[125].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[125].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[125].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[125].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[125].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[125].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[125].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[125].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[125].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[125].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[125].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[125].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[125].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[125].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[125].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[125].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe580", + 3200 + ], + "l1_pong": [ + "0x9800", + 3200 + ], + "l2": [ + [ + 3, + 0, + 0, + 89600 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_143_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 1802496, + "size": 89600 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_143_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0144": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_143" + ], + "out": [ + "compute_graph.l2_144" + ], + "wts": [ + "compute_graph.Layer_144_l2_wts", + "compute_graph.Layer_144_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[126].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[126].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[126].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[126].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[126].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[126].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[126].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[126].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[126].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[126].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[126].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[126].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[126].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[126].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[126].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[126].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2048 + ], + "l1_pong": [ + "0xcd80", + 2048 + ], + "l2": [ + [ + 0, + 0, + 0, + 49152, + 26880 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_143" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_143", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_236", + "layer_object_name": "compute_graph.flexml_layers[126]", + "layer_order": 144, + "mlir_dag_id_map": { + "dag_layer": 144, + "mlir_layer": 144 + }, + "num_iter": 18, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[126].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[126].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[126].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[126].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[126].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[126].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[126].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[126].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[126].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[126].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[126].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[126].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[126].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[126].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[126].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[126].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1792 + ], + "l1_pong": [ + "0x4000", + 1792 + ], + "l2": [ + [ + 2, + 0, + 8192, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_144" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[126].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[126].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[126].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[126].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[126].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[126].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[126].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[126].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[126].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[126].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[126].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[126].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[126].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[126].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[126].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[126].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xdd80", + 3808 + ], + "l1_pong": [ + "0x9000", + 3808 + ], + "l2": [ + [ + 3, + 0, + 341504, + 91392 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_144_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 1981696, + "size": 91392 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_144_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0145": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_144" + ], + "out": [ + "compute_graph.l2_145" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[127].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[127].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[127].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[127].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[127].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[127].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[127].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[127].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[127].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[127].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[127].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[127].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[127].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[127].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[127].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[127].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 8192, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_144" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_144", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_238", + "layer_object_name": "compute_graph.flexml_layers[127]", + "layer_order": 145, + "mlir_dag_id_map": { + "dag_layer": 145, + "mlir_layer": 145 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[127].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[127].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[127].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[127].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[127].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[127].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[127].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[127].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[127].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[127].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[127].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[127].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[127].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[127].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[127].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[127].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_145" + ] + }, + "super_iter": 1 + }, + "0146": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_145" + ], + "out": [ + "compute_graph.l2_146" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[128].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[128].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[128].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[128].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[128].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[128].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[128].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[128].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[128].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[128].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[128].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[128].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[128].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[128].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[128].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[128].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_145" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_145", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_241", + "layer_object_name": "compute_graph.flexml_layers[128]", + "layer_order": 146, + "mlir_dag_id_map": { + "dag_layer": 146, + "mlir_layer": 146 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[128].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[128].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[128].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[128].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[128].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[128].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[128].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[128].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[128].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[128].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[128].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[128].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[128].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[128].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[128].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[128].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 1, + 0, + 194560, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_146" + ] + }, + "super_iter": 1 + }, + "0147": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_146" + ], + "out": [ + "compute_graph.l2_147" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[129].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[129].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[129].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[129].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[129].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[129].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[129].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[129].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[129].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[129].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[129].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[129].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[129].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[129].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[129].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[129].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 1, + 0, + 194560, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_146" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_146", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_243", + "layer_object_name": "compute_graph.flexml_layers[129]", + "layer_order": 147, + "mlir_dag_id_map": { + "dag_layer": 147, + "mlir_layer": 147 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[129].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[129].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[129].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[129].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[129].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[129].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[129].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[129].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[129].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[129].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[129].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[129].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[129].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[129].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[129].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[129].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_147" + ] + }, + "super_iter": 1 + }, + "0148": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_144", + "compute_graph.l2_147" + ], + "out": [ + "compute_graph.l2_148" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[130].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[130].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[130].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[130].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[130].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[130].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[130].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[130].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[130].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[130].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[130].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[130].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[130].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[130].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[130].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[130].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 8192, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_144" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[130].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[130].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[130].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[130].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[130].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[130].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[130].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[130].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[130].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[130].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[130].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[130].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[130].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[130].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[130].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[130].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_147" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_144", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_147", + "dest_port": 2, + "src_offset": 0 + } + ], + "layer_name": "Mul_244", + "layer_object_name": "compute_graph.flexml_layers[130]", + "layer_order": 148, + "mlir_dag_id_map": { + "dag_layer": 148, + "mlir_layer": 148 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[130].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[130].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[130].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[130].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[130].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[130].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[130].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[130].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[130].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[130].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[130].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[130].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[130].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[130].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[130].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[130].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 337920, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_148" + ] + }, + "super_iter": 1 + }, + "0149": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_148" + ], + "out": [ + "compute_graph.l2_149" + ], + "wts": [ + "compute_graph.Layer_149_l2_wts", + "compute_graph.Layer_149_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[131].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[131].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[131].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[131].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[131].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[131].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[131].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[131].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[131].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[131].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[131].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[131].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[131].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[131].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[131].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[131].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5120 + ], + "l1_pong": [ + "0xcd80", + 5120 + ], + "l2": [ + [ + 0, + 0, + 337920, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_148" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_148", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_245", + "layer_object_name": "compute_graph.flexml_layers[131]", + "layer_order": 149, + "mlir_dag_id_map": { + "dag_layer": 149, + "mlir_layer": 149 + }, + "num_iter": 126, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[131].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[131].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[131].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[131].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[131].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[131].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[131].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[131].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[131].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[131].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[131].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[131].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[131].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[131].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[131].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[131].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 128 + ], + "l1_pong": [ + "0x4000", + 128 + ], + "l2": [ + [ + 1, + 0, + 151552, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_149" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[131].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[131].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[131].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[131].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[131].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[131].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[131].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[131].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[131].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[131].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[131].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[131].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[131].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[131].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[131].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[131].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf580", + 896 + ], + "l1_pong": [ + "0xa800", + 896 + ], + "l2": [ + [ + 3, + 0, + 0, + 75264 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_149_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 2164480, + "size": 75264 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_149_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0150": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_149" + ], + "out": [ + "compute_graph.l2_150" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[132].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[132].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[132].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[132].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[132].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[132].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[132].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[132].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[132].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[132].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[132].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[132].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[132].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[132].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[132].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[132].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 1, + 0, + 151552, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_149" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_149", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_247", + "layer_object_name": "compute_graph.flexml_layers[132]", + "layer_order": 150, + "mlir_dag_id_map": { + "dag_layer": 150, + "mlir_layer": 150 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[132].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[132].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[132].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[132].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[132].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[132].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[132].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[132].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[132].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[132].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[132].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[132].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[132].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[132].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[132].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[132].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_150" + ] + }, + "super_iter": 1 + }, + "0151": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_150" + ], + "out": [ + "compute_graph.l2_151" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[133].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[133].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[133].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[133].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[133].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[133].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[133].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[133].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[133].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[133].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[133].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[133].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[133].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[133].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[133].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[133].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_150" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_150", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_250", + "layer_object_name": "compute_graph.flexml_layers[133]", + "layer_order": 151, + "mlir_dag_id_map": { + "dag_layer": 151, + "mlir_layer": 151 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[133].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[133].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[133].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[133].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[133].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[133].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[133].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[133].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[133].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[133].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[133].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[133].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[133].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[133].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[133].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[133].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 186368, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_151" + ] + }, + "super_iter": 1 + }, + "0152": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_151" + ], + "out": [ + "compute_graph.l2_152" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[134].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[134].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[134].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[134].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[134].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[134].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[134].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[134].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[134].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[134].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[134].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[134].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[134].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[134].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[134].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[134].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 186368, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_151" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_151", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Div_252", + "layer_object_name": "compute_graph.flexml_layers[134]", + "layer_order": 152, + "mlir_dag_id_map": { + "dag_layer": 152, + "mlir_layer": 152 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[134].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[134].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[134].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[134].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[134].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[134].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[134].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[134].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[134].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[134].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[134].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[134].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[134].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[134].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[134].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[134].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_152" + ] + }, + "super_iter": 1 + }, + "0153": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_149", + "compute_graph.l2_152" + ], + "out": [ + "compute_graph.l2_153", + "compute_graph.l2l3_153_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[135].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[135].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[135].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[135].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[135].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[135].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[135].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[135].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[135].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[135].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[135].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[135].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[135].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[135].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[135].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[135].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 1, + 0, + 151552, + 258048, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_149" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[135].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[135].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[135].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[135].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[135].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[135].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[135].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[135].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[135].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[135].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[135].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[135].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[135].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[135].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[135].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[135].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_152" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_149", + "dest_port": 0, + "src_offset": 4 + }, + { + "buff_obj": "compute_graph.l2_152", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_153.out[0]", + "to": "compute_graph.l2l3_153_spill.in[0]" + }, + { + "from": "compute_graph.l2_153.out[1]", + "to": "compute_graph.l2l3_153_spill.in[1]" + }, + { + "from": "compute_graph.l2_153.out[2]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.in[0]" + }, + { + "from": "compute_graph.l2_153.out[3]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.in[1]" + } + ], + "layer_name": "Mul_253", + "layer_object_name": "compute_graph.flexml_layers[135]", + "layer_order": 153, + "mlir_dag_id_map": { + "dag_layer": 153, + "mlir_layer": 153 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[135].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[135].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[135].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[135].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[135].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[135].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[135].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[135].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[135].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[135].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[135].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[135].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[135].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[135].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[135].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[135].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 337920, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_153" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_153153": { + "size": 161280 + }, + "l2l3_153_spill": { + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_153_spill", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153" + ] + }, + "super_iter": 1 + }, + "0154": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_153_spill" + ], + "out": [ + "compute_graph.l2l3_154_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_153_spill": { + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_153_spill" + ] + }, + "kernel_name": [ + "Transpose4dAdf" + ], + "layer_name": "Generated-#36", + "layer_object_name": "compute_graph.templated_graph_154", + "layer_order": 154, + "mlir_dag_id_map": { + "dag_layer": 154, + "mlir_layer": 154 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_154_spill": { + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_154_spill" + ] + }, + "templated_graph": 1 + }, + "0155": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0", + "compute_graph.l2l3_154_spill" + ], + "out": [ + "compute_graph.l2_155" + ] + }, + "buffer_iter": 1, + "depth_iter": 8, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[136].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[136].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[136].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[136].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[136].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[136].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[136].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[136].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[136].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[136].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[136].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[136].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[136].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[136].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[136].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[136].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7168 + ], + "l1_pong": [ + "0x4000", + 7168 + ], + "l2": [ + [ + 0, + 0, + 0, + 161280, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_154_spill" + ] + }, + "kernel_name": [ + "superkernel_reduce_mean_c8" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_154_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_154_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_154_for_layer_155_port0.in[1]" + } + ], + "layer_name": "Generated-#38", + "layer_object_name": "compute_graph.flexml_layers[136]", + "layer_order": 155, + "mlir_dag_id_map": { + "dag_layer": 155, + "mlir_layer": 155 + }, + "num_iter": 24, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[136].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[136].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[136].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[136].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[136].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[136].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[136].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[136].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[136].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[136].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[136].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[136].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[136].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[136].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[136].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[136].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaf00", + 56 + ], + "l1_pong": [ + "0xcd80", + 56 + ], + "l2": [ + [ + 2, + 0, + 518912, + 2688, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_155" + ] + }, + "super_iter": 1 + }, + "0156": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_155" + ], + "out": [ + "compute_graph.l2_156" + ], + "wts": [ + "compute_graph.Layer_156_l2_wts", + "compute_graph.Layer_156_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 7, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[137].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[137].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[137].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[137].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[137].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[137].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[137].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[137].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[137].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[137].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[137].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[137].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[137].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[137].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[137].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[137].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 768 + ], + "l1_pong": [ + "0xcd80", + 768 + ], + "l2": [ + [ + 2, + 0, + 518912, + 2688, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_155" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_155", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_255", + "layer_object_name": "compute_graph.flexml_layers[137]", + "layer_order": 156, + "mlir_dag_id_map": { + "dag_layer": 156, + "mlir_layer": 156 + }, + "num_iter": 7, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[137].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[137].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[137].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[137].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[137].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[137].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[137].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[137].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[137].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[137].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[137].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[137].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[137].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[137].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[137].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[137].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 384 + ], + "l1_pong": [ + "0x4000", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 168 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_156" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[137].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[137].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[137].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[137].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[137].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[137].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[137].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[137].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[137].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[137].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[137].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[137].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[137].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[137].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[137].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[137].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xd380", + 4800 + ], + "l1_pong": [ + "0x8600", + 4800 + ], + "l2": [ + [ + 3, + 0, + 0, + 134400 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_156_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 2315008, + "size": 134400 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_156_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0157": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_156" + ], + "out": [ + "compute_graph.l2_157" + ], + "wts": [ + "compute_graph.Layer_157_l2_wts", + "compute_graph.Layer_157_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[138].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[138].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[138].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[138].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[138].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[138].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[138].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[138].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[138].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[138].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[138].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[138].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[138].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[138].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[138].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[138].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2688 + ], + "l1_pong": [ + "0xcd80", + 2688 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 168 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_156" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_156", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_257", + "layer_object_name": "compute_graph.flexml_layers[138]", + "layer_order": 157, + "mlir_dag_id_map": { + "dag_layer": 157, + "mlir_layer": 157 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[138].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[138].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[138].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[138].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[138].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[138].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[138].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[138].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[138].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[138].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[138].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[138].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[138].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[138].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[138].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[138].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 2, + 0, + 434176, + 45056, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_157" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[138].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[138].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[138].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[138].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[138].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[138].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[138].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[138].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[138].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[138].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[138].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[138].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[138].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[138].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[138].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[138].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe280", + 2752 + ], + "l1_pong": [ + "0x9500", + 2752 + ], + "l2": [ + [ + 3, + 0, + 282112, + 121088 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_157_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 2583808, + "size": 121088 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_157_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0158": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_157" + ], + "out": [ + "compute_graph.l2_158" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[139].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[139].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[139].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[139].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[139].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[139].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[139].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[139].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[139].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[139].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[139].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[139].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[139].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[139].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[139].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[139].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 2, + 0, + 434176, + 45056, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_157" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_157", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_259", + "layer_object_name": "compute_graph.flexml_layers[139]", + "layer_order": 158, + "mlir_dag_id_map": { + "dag_layer": 158, + "mlir_layer": 158 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[139].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[139].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[139].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[139].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[139].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[139].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[139].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[139].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[139].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[139].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[139].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[139].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[139].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[139].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[139].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[139].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaf80", + 384 + ], + "l1_pong": [ + "0xcd80", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_158" + ] + }, + "super_iter": 1 + }, + "0159": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_158" + ], + "out": [ + "compute_graph.l2_159" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[140].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[140].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[140].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[140].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[140].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[140].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[140].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[140].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[140].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[140].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[140].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[140].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[140].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[140].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[140].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[140].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_158" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_158", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_262", + "layer_object_name": "compute_graph.flexml_layers[140]", + "layer_order": 159, + "mlir_dag_id_map": { + "dag_layer": 159, + "mlir_layer": 159 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[140].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[140].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[140].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[140].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[140].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[140].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[140].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[140].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[140].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[140].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[140].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[140].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[140].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[140].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[140].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[140].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 256 + ], + "l1_pong": [ + "0xcd80", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_159" + ] + }, + "super_iter": 1 + }, + "0160": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_159" + ], + "out": [ + "compute_graph.l2_160", + "compute_graph.l2l3_160_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[141].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[141].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[141].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[141].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[141].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[141].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[141].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[141].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[141].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[141].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[141].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[141].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[141].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[141].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[141].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[141].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_159" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_159", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_160.out[0]", + "to": "compute_graph.l2l3_160_spill.in[0]" + }, + { + "from": "compute_graph.l2_160.out[1]", + "to": "compute_graph.l2l3_160_spill.in[1]" + } + ], + "layer_name": "Div_264", + "layer_object_name": "compute_graph.flexml_layers[141]", + "layer_order": 160, + "mlir_dag_id_map": { + "dag_layer": 160, + "mlir_layer": 160 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[141].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[141].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[141].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[141].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[141].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[141].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[141].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[141].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[141].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[141].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[141].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[141].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[141].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[141].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[141].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[141].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xae80", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 8192, + 672 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_160" + ], + "l3": { + "l2l3_160_spill": { + "size": 672 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_160_spill" + ] + }, + "super_iter": 1 + }, + "0161": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_160_spill", + "compute_graph.l2l3_scratch_0_161_spill", + "compute_graph.l2l3_scratch_1_161_spill" + ], + "out": [ + "compute_graph.l2l3_161_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_160_spill": { + "size": 672 + }, + "l2l3_scratch_0_161_spill": { + "size": 322560 + }, + "l2l3_scratch_1_161_spill": { + "size": 322560 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_160_spill", + "compute_graph.l2l3_scratch_0_161_spill", + "compute_graph.l2l3_scratch_1_161_spill" + ] + }, + "kernel_name": [ + "TileAdf" + ], + "layer_name": "Generated-#40", + "layer_object_name": "compute_graph.templated_graph_161", + "layer_order": 161, + "mlir_dag_id_map": { + "dag_layer": 161, + "mlir_layer": 161 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_161_spill": { + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_161_spill" + ] + }, + "templated_graph": 1 + }, + "0162": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0", + "compute_graph.l2l3_161_spill", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153" + ], + "out": [ + "compute_graph.l2_162" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[142].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[142].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[142].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[142].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[142].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[142].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[142].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[142].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[142].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[142].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[142].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[142].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[142].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[142].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[142].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[142].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 161280, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_161_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[142].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[142].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[142].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[142].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[142].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[142].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[142].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[142].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[142].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[142].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[142].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[142].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[142].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[142].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[142].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[142].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 322560, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_153153.in[1]" + }, + { + "from": "compute_graph.l2l3_161_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_161_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_161_for_layer_162_port0.in[1]" + } + ], + "layer_name": "Mul_265", + "layer_object_name": "compute_graph.flexml_layers[142]", + "layer_order": 162, + "mlir_dag_id_map": { + "dag_layer": 162, + "mlir_layer": 162 + }, + "num_iter": 11, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[142].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[142].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[142].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[142].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[142].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[142].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[142].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[142].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[142].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[142].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[142].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[142].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[142].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[142].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[142].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[142].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 186368, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_162" + ] + }, + "super_iter": 1 + }, + "0163": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_162" + ], + "out": [ + "compute_graph.l2_163", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163" + ], + "wts": [ + "compute_graph.Layer_163_l2_wts", + "compute_graph.Layer_163_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 9, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[143].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[143].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[143].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[143].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[143].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[143].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[143].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[143].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[143].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[143].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[143].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[143].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[143].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[143].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[143].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[143].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2560 + ], + "l1_pong": [ + "0xcd80", + 2560 + ], + "l2": [ + [ + 2, + 0, + 186368, + 168960, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_162" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_162", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_163.out[4]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.in[0]" + }, + { + "from": "compute_graph.l2_163.out[5]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.in[1]" + } + ], + "layer_name": "Conv_266", + "layer_object_name": "compute_graph.flexml_layers[143]", + "layer_order": 163, + "mlir_dag_id_map": { + "dag_layer": 163, + "mlir_layer": 163 + }, + "num_iter": 27, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[143].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[143].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[143].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[143].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[143].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[143].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[143].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[143].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[143].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[143].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[143].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[143].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[143].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[143].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[143].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[143].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1280 + ], + "l1_pong": [ + "0x4000", + 1280 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 38400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_163" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_163163": { + "size": 38400 + } + }, + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[143].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[143].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[143].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[143].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[143].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[143].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[143].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[143].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[143].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[143].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[143].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[143].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[143].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[143].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[143].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[143].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe180", + 3360 + ], + "l1_pong": [ + "0x9400", + 3360 + ], + "l2": [ + [ + 3, + 0, + 0, + 120960 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_163_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 2825984, + "size": 120960 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_163_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0164": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_163" + ], + "out": [ + "compute_graph.l2_164", + "compute_graph.l2l3_164_spill" + ], + "wts": [ + "compute_graph.Layer_164_l2_wts", + "compute_graph.Layer_164_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[144].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[144].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[144].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[144].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[144].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[144].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[144].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[144].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[144].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[144].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[144].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[144].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[144].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[144].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[144].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[144].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2560 + ], + "l1_pong": [ + "0xcd80", + 2560 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 38400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_163" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_163", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_164.out[4]", + "to": "compute_graph.l2l3_164_spill.in[0]" + }, + { + "from": "compute_graph.l2_164.out[5]", + "to": "compute_graph.l2l3_164_spill.in[1]" + } + ], + "layer_name": "Conv_267", + "layer_object_name": "compute_graph.flexml_layers[144]", + "layer_order": 164, + "mlir_dag_id_map": { + "dag_layer": 164, + "mlir_layer": 164 + }, + "num_iter": 36, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[144].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[144].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[144].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[144].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[144].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[144].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[144].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[144].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[144].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[144].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[144].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[144].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[144].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[144].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[144].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[144].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1280 + ], + "l1_pong": [ + "0x4000", + 1280 + ], + "l2": [ + [ + 1, + 0, + 311296, + 368640, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_164" + ], + "l3": { + "l2l3_164_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_164_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[144].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[144].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[144].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[144].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[144].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[144].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[144].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[144].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[144].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[144].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[144].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[144].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[144].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[144].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[144].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[144].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe180", + 3360 + ], + "l1_pong": [ + "0x9400", + 3360 + ], + "l2": [ + [ + 3, + 0, + 201728, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_164_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 3067904, + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_164_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0165": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_164" + ], + "out": [ + "compute_graph.l2_165" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[145].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[145].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[145].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[145].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[145].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[145].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[145].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[145].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[145].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[145].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[145].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[145].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[145].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[145].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[145].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[145].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 1, + 0, + 311296, + 368640, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_164" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_164", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_269", + "layer_object_name": "compute_graph.flexml_layers[145]", + "layer_order": 165, + "mlir_dag_id_map": { + "dag_layer": 165, + "mlir_layer": 165 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[145].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[145].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[145].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[145].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[145].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[145].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[145].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[145].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[145].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[145].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[145].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[145].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[145].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[145].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[145].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[145].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_165" + ] + }, + "super_iter": 1 + }, + "0166": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_165" + ], + "out": [ + "compute_graph.l2_166" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[146].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[146].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[146].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[146].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[146].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[146].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[146].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[146].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[146].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[146].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[146].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[146].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[146].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[146].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[146].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[146].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_165" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_165", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_272", + "layer_object_name": "compute_graph.flexml_layers[146]", + "layer_order": 166, + "mlir_dag_id_map": { + "dag_layer": 166, + "mlir_layer": 166 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[146].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[146].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[146].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[146].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[146].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[146].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[146].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[146].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[146].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[146].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[146].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[146].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[146].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[146].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[146].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[146].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_166" + ] + }, + "super_iter": 1 + }, + "0167": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_166" + ], + "out": [ + "compute_graph.l2_167", + "compute_graph.l2l3_167_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[147].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[147].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[147].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[147].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[147].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[147].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[147].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[147].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[147].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[147].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[147].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[147].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[147].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[147].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[147].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[147].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_166" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_166", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_167.out[0]", + "to": "compute_graph.l2l3_167_spill.in[0]" + }, + { + "from": "compute_graph.l2_167.out[1]", + "to": "compute_graph.l2l3_167_spill.in[1]" + } + ], + "layer_name": "Div_274", + "layer_object_name": "compute_graph.flexml_layers[147]", + "layer_order": 167, + "mlir_dag_id_map": { + "dag_layer": 167, + "mlir_layer": 167 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[147].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[147].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[147].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[147].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[147].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[147].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[147].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[147].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[147].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[147].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[147].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[147].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[147].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[147].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[147].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[147].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_167" + ], + "l3": { + "l2l3_167_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_167_spill" + ] + }, + "super_iter": 1 + }, + "0168": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0", + "compute_graph.l2l3_164_spill", + "compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1", + "compute_graph.l2l3_167_spill" + ], + "out": [ + "compute_graph.l2_168", + "compute_graph.l2l3_168_spill" + ] + }, + "buffer_iter": 2, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[148].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[148].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[148].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[148].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[148].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[148].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[148].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[148].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[148].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[148].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[148].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[148].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[148].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[148].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[148].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[148].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 293888, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_164_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[148].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[148].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[148].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[148].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[148].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[148].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[148].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[148].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[148].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[148].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[148].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[148].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[148].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[148].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[148].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[148].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 2, + 0, + 63488, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_167_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_168.out[0]", + "to": "compute_graph.l2l3_168_spill.in[0]" + }, + { + "from": "compute_graph.l2_168.out[1]", + "to": "compute_graph.l2l3_168_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_164_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_164_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_164_for_layer_168_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_167_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_167_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_167_for_layer_168_port1.in[1]" + } + ], + "layer_name": "Mul_275", + "layer_object_name": "compute_graph.flexml_layers[148]", + "layer_order": 168, + "mlir_dag_id_map": { + "dag_layer": 168, + "mlir_layer": 168 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[148].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[148].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[148].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[148].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[148].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[148].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[148].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[148].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[148].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[148].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[148].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[148].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[148].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[148].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[148].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[148].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 153600, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_168" + ], + "l2_force_single_buffering": true, + "l3": { + "l2l3_168_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_168_spill" + ] + }, + "super_iter": 2 + }, + "0169": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0", + "compute_graph.l2l3_168_spill" + ], + "out": [ + "compute_graph.l2_169", + "compute_graph.l2l3_169_spill" + ], + "wts": [ + "compute_graph.Layer_169_l2_wts", + "compute_graph.Layer_169_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[149].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[149].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[149].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[149].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[149].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[149].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[149].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[149].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[149].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[149].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[149].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[149].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[149].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[149].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[149].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[149].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5120 + ], + "l1_pong": [ + "0xcd80", + 5120 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_168_spill" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_169.out[4]", + "to": "compute_graph.l2l3_169_spill.in[0]" + }, + { + "from": "compute_graph.l2_169.out[5]", + "to": "compute_graph.l2l3_169_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_168_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_168_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_168_for_layer_169_port0.in[1]" + } + ], + "layer_name": "Conv_276", + "layer_object_name": "compute_graph.flexml_layers[149]", + "layer_order": 169, + "mlir_dag_id_map": { + "dag_layer": 169, + "mlir_layer": 169 + }, + "num_iter": 180, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[149].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[149].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[149].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[149].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[149].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[149].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[149].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[149].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[149].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[149].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[149].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[149].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[149].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[149].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[149].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[149].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 128 + ], + "l1_pong": [ + "0x4000", + 128 + ], + "l2": [ + [ + 1, + 0, + 311296, + 368640, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_169" + ], + "l3": { + "l2l3_169_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_169_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[149].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[149].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[149].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[149].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[149].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[149].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[149].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[149].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[149].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[149].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[149].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[149].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[149].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[149].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[149].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[149].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf580", + 896 + ], + "l1_pong": [ + "0xa800", + 896 + ], + "l2": [ + [ + 3, + 0, + 0, + 107520 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_169_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 3390464, + "size": 107520 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_169_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0170": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_169" + ], + "out": [ + "compute_graph.l2_170" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[150].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[150].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[150].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[150].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[150].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[150].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[150].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[150].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[150].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[150].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[150].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[150].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[150].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[150].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[150].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[150].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 1, + 0, + 311296, + 368640, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_169" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_169", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_278", + "layer_object_name": "compute_graph.flexml_layers[150]", + "layer_order": 170, + "mlir_dag_id_map": { + "dag_layer": 170, + "mlir_layer": 170 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[150].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[150].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[150].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[150].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[150].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[150].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[150].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[150].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[150].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[150].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[150].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[150].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[150].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[150].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[150].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[150].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_170" + ] + }, + "super_iter": 1 + }, + "0171": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_170" + ], + "out": [ + "compute_graph.l2_171" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[151].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[151].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[151].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[151].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[151].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[151].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[151].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[151].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[151].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[151].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[151].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[151].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[151].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[151].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[151].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[151].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_170" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_170", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_281", + "layer_object_name": "compute_graph.flexml_layers[151]", + "layer_order": 171, + "mlir_dag_id_map": { + "dag_layer": 171, + "mlir_layer": 171 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[151].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[151].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[151].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[151].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[151].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[151].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[151].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[151].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[151].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[151].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[151].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[151].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[151].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[151].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[151].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[151].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_171" + ] + }, + "super_iter": 1 + }, + "0172": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_171" + ], + "out": [ + "compute_graph.l2_172", + "compute_graph.l2l3_172_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[152].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[152].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[152].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[152].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[152].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[152].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[152].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[152].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[152].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[152].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[152].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[152].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[152].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[152].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[152].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[152].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_171" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_171", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_172.out[0]", + "to": "compute_graph.l2l3_172_spill.in[0]" + }, + { + "from": "compute_graph.l2_172.out[1]", + "to": "compute_graph.l2l3_172_spill.in[1]" + } + ], + "layer_name": "Div_283", + "layer_object_name": "compute_graph.flexml_layers[152]", + "layer_order": 172, + "mlir_dag_id_map": { + "dag_layer": 172, + "mlir_layer": 172 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[152].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[152].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[152].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[152].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[152].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[152].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[152].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[152].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[152].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[152].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[152].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[152].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[152].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[152].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[152].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[152].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_172" + ], + "l3": { + "l2l3_172_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_172_spill" + ] + }, + "super_iter": 1 + }, + "0173": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0", + "compute_graph.l2l3_169_spill", + "compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1", + "compute_graph.l2l3_172_spill" + ], + "out": [ + "compute_graph.l2_173", + "compute_graph.l2l3_173_spill" + ] + }, + "buffer_iter": 2, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[153].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[153].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[153].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[153].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[153].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[153].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[153].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[153].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[153].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[153].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[153].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[153].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[153].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[153].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[153].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[153].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 293888, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_169_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[153].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[153].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[153].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[153].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[153].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[153].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[153].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[153].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[153].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[153].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[153].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[153].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[153].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[153].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[153].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[153].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 2, + 0, + 63488, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_172_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_173.out[0]", + "to": "compute_graph.l2l3_173_spill.in[0]" + }, + { + "from": "compute_graph.l2_173.out[1]", + "to": "compute_graph.l2l3_173_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_169_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_169_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_169_for_layer_173_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_172_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_172_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_172_for_layer_173_port1.in[1]" + } + ], + "layer_name": "Mul_284", + "layer_object_name": "compute_graph.flexml_layers[153]", + "layer_order": 173, + "mlir_dag_id_map": { + "dag_layer": 173, + "mlir_layer": 173 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[153].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[153].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[153].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[153].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[153].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[153].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[153].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[153].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[153].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[153].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[153].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[153].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[153].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[153].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[153].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[153].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 153600, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_173" + ], + "l2_force_single_buffering": true, + "l3": { + "l2l3_173_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_173_spill" + ] + }, + "super_iter": 2 + }, + "0174": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_173_spill" + ], + "out": [ + "compute_graph.l2l3_174_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_173_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_173_spill" + ] + }, + "kernel_name": [ + "Transpose4dAdf" + ], + "layer_name": "Generated-#42", + "layer_object_name": "compute_graph.templated_graph_174", + "layer_order": 174, + "mlir_dag_id_map": { + "dag_layer": 174, + "mlir_layer": 174 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_174_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_174_spill" + ] + }, + "templated_graph": 1 + }, + "0175": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0", + "compute_graph.l2l3_174_spill" + ], + "out": [ + "compute_graph.l2_175" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[154].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[154].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[154].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[154].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[154].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[154].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[154].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[154].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[154].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[154].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[154].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[154].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[154].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[154].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[154].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[154].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_174_spill" + ] + }, + "kernel_name": [ + "superkernel_reduce_mean_c8" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_174_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_174_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_174_for_layer_175_port0.in[1]" + } + ], + "layer_name": "Generated-#44", + "layer_object_name": "compute_graph.flexml_layers[154]", + "layer_order": 175, + "mlir_dag_id_map": { + "dag_layer": 175, + "mlir_layer": 175 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[154].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[154].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[154].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[154].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[154].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[154].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[154].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[154].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[154].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[154].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[154].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[154].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[154].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[154].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[154].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[154].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb000", + 40 + ], + "l1_pong": [ + "0xcd80", + 40 + ], + "l2": [ + [ + 2, + 0, + 516608, + 3840, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_175" + ] + }, + "super_iter": 1 + }, + "0176": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_175" + ], + "out": [ + "compute_graph.l2_176" + ], + "wts": [ + "compute_graph.Layer_176_l2_wts", + "compute_graph.Layer_176_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[155].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[155].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[155].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[155].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[155].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[155].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[155].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[155].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[155].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[155].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[155].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[155].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[155].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[155].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[155].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[155].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3072 + ], + "l1_pong": [ + "0xcd80", + 3072 + ], + "l2": [ + [ + 2, + 0, + 516608, + 3840, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_175" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_175", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_286", + "layer_object_name": "compute_graph.flexml_layers[155]", + "layer_order": 176, + "mlir_dag_id_map": { + "dag_layer": 176, + "mlir_layer": 176 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[155].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[155].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[155].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[155].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[155].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[155].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[155].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[155].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[155].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[155].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[155].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[155].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[155].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[155].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[155].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[155].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 0, + 0, + 0, + 16384, + 240 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_176" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[155].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[155].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[155].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[155].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[155].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[155].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[155].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[155].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[155].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[155].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[155].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[155].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[155].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[155].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[155].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[155].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe580", + 3136 + ], + "l1_pong": [ + "0x9800", + 3136 + ], + "l2": [ + [ + 3, + 0, + 0, + 250880 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_176_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 3605504, + "size": 250880 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_176_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0177": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_176" + ], + "out": [ + "compute_graph.l2_177" + ], + "wts": [ + "compute_graph.Layer_177_l2_wts", + "compute_graph.Layer_177_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 3, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[156].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[156].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[156].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[156].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[156].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[156].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[156].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[156].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[156].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[156].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[156].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[156].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[156].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[156].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[156].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[156].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 640 + ], + "l1_pong": [ + "0xcd80", + 640 + ], + "l2": [ + [ + 0, + 0, + 0, + 16384, + 240 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_176" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_176", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_288", + "layer_object_name": "compute_graph.flexml_layers[156]", + "layer_order": 177, + "mlir_dag_id_map": { + "dag_layer": 177, + "mlir_layer": 177 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[156].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[156].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[156].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[156].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[156].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[156].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[156].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[156].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[156].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[156].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[156].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[156].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[156].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[156].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[156].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[156].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 384 + ], + "l1_pong": [ + "0x4000", + 384 + ], + "l2": [ + [ + 2, + 0, + 462848, + 30720, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_177" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[156].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[156].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[156].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[156].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[156].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[156].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[156].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[156].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[156].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[156].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[156].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[156].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[156].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[156].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[156].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[156].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xd280", + 4032 + ], + "l1_pong": [ + "0x8500", + 4032 + ], + "l2": [ + [ + 3, + 0, + 40448, + 241920 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_177_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 4107264, + "size": 241920 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_177_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0178": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_177" + ], + "out": [ + "compute_graph.l2_178" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[157].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[157].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[157].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[157].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[157].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[157].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[157].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[157].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[157].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[157].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[157].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[157].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[157].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[157].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[157].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[157].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 2, + 0, + 462848, + 30720, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_177" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_177", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_290", + "layer_object_name": "compute_graph.flexml_layers[157]", + "layer_order": 178, + "mlir_dag_id_map": { + "dag_layer": 178, + "mlir_layer": 178 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[157].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[157].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[157].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[157].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[157].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[157].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[157].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[157].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[157].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[157].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[157].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[157].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[157].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[157].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[157].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[157].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaf80", + 384 + ], + "l1_pong": [ + "0xcd80", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_178" + ] + }, + "super_iter": 1 + }, + "0179": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_178" + ], + "out": [ + "compute_graph.l2_179" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[158].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[158].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[158].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[158].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[158].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[158].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[158].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[158].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[158].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[158].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[158].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[158].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[158].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[158].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[158].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[158].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_178" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_178", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_293", + "layer_object_name": "compute_graph.flexml_layers[158]", + "layer_order": 179, + "mlir_dag_id_map": { + "dag_layer": 179, + "mlir_layer": 179 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[158].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[158].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[158].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[158].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[158].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[158].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[158].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[158].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[158].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[158].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[158].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[158].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[158].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[158].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[158].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[158].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 256 + ], + "l1_pong": [ + "0xcd80", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_179" + ] + }, + "super_iter": 1 + }, + "0180": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_179" + ], + "out": [ + "compute_graph.l2_180", + "compute_graph.l2l3_180_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[159].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[159].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[159].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[159].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[159].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[159].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[159].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[159].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[159].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[159].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[159].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[159].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[159].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[159].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[159].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[159].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_179" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_179", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_180.out[0]", + "to": "compute_graph.l2l3_180_spill.in[0]" + }, + { + "from": "compute_graph.l2_180.out[1]", + "to": "compute_graph.l2l3_180_spill.in[1]" + } + ], + "layer_name": "Div_295", + "layer_object_name": "compute_graph.flexml_layers[159]", + "layer_order": 180, + "mlir_dag_id_map": { + "dag_layer": 180, + "mlir_layer": 180 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[159].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[159].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[159].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[159].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[159].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[159].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[159].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[159].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[159].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[159].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[159].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[159].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[159].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[159].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[159].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[159].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xae80", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 8192, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_180" + ], + "l3": { + "l2l3_180_spill": { + "size": 960 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_180_spill" + ] + }, + "super_iter": 1 + }, + "0181": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_180_spill", + "compute_graph.l2l3_scratch_0_181_spill", + "compute_graph.l2l3_scratch_1_181_spill" + ], + "out": [ + "compute_graph.l2l3_181_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_180_spill": { + "size": 960 + }, + "l2l3_scratch_0_181_spill": { + "size": 460800 + }, + "l2l3_scratch_1_181_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_180_spill", + "compute_graph.l2l3_scratch_0_181_spill", + "compute_graph.l2l3_scratch_1_181_spill" + ] + }, + "kernel_name": [ + "TileAdf" + ], + "layer_name": "Generated-#46", + "layer_object_name": "compute_graph.templated_graph_181", + "layer_order": 181, + "mlir_dag_id_map": { + "dag_layer": 181, + "mlir_layer": 181 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_181_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_181_spill" + ] + }, + "templated_graph": 1 + }, + "0182": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1", + "compute_graph.l2l3_173_spill", + "compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0", + "compute_graph.l2l3_181_spill" + ], + "out": [ + "compute_graph.l2_182" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[160].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[160].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[160].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[160].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[160].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[160].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[160].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[160].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[160].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[160].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[160].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[160].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[160].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[160].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[160].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[160].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 460800, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_181_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[160].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[160].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[160].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[160].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[160].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[160].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[160].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[160].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[160].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[160].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[160].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[160].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[160].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[160].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[160].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[160].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_173_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_173_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_173_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_173_for_layer_182_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_181_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_181_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_181_for_layer_182_port0.in[1]" + } + ], + "layer_name": "Mul_296", + "layer_object_name": "compute_graph.flexml_layers[160]", + "layer_order": 182, + "mlir_dag_id_map": { + "dag_layer": 182, + "mlir_layer": 182 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[160].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[160].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[160].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[160].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[160].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[160].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[160].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[160].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[160].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[160].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[160].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[160].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[160].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[160].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[160].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[160].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_182" + ] + }, + "super_iter": 1 + }, + "0183": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_182", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163" + ], + "out": [ + "compute_graph.l2_183", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183" + ], + "wts": [ + "compute_graph.Layer_183_l2_wts", + "compute_graph.Layer_183_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 9, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[161].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[161].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[161].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[161].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[161].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[161].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[161].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[161].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[161].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[161].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[161].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[161].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[161].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[161].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[161].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[161].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3584 + ], + "l1_pong": [ + "0xcd80", + 3584 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_182" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[161].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[161].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[161].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[161].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[161].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[161].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[161].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[161].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[161].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[161].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[161].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[161].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[161].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[161].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[161].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[161].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 3072 + ], + "l1_pong": [ + "0x6000", + 3072 + ], + "l2": [ + [ + 0, + 0, + 0, + 73728, + 38400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_182", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_183.out[4]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.in[0]" + }, + { + "from": "compute_graph.l2_183.out[5]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_163163.in[1]" + } + ], + "layer_name": "Conv_297", + "layer_object_name": "compute_graph.flexml_layers[161]", + "layer_order": 183, + "mlir_dag_id_map": { + "dag_layer": 183, + "mlir_layer": 183 + }, + "num_iter": 54, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[161].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[161].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[161].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[161].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[161].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[161].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[161].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[161].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[161].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[161].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[161].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[161].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[161].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[161].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[161].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[161].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 0, + 73728, + 38400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_183" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_183183": { + "size": 38400 + } + }, + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[161].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[161].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[161].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[161].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[161].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[161].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[161].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[161].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[161].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[161].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[161].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[161].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[161].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[161].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[161].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[161].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe980", + 2784 + ], + "l1_pong": [ + "0x9c00", + 2784 + ], + "l2": [ + [ + 3, + 0, + 0, + 200448 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_183_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 4591104, + "size": 200448 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_183_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0184": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_183" + ], + "out": [ + "compute_graph.l2_184", + "compute_graph.l2l3_184_spill" + ], + "wts": [ + "compute_graph.Layer_184_l2_wts", + "compute_graph.Layer_184_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[162].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[162].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[162].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[162].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[162].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[162].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[162].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[162].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[162].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[162].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[162].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[162].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[162].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[162].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[162].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[162].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2560 + ], + "l1_pong": [ + "0xcd80", + 2560 + ], + "l2": [ + [ + 0, + 0, + 0, + 73728, + 38400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_183" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_183", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_184.out[4]", + "to": "compute_graph.l2l3_184_spill.in[0]" + }, + { + "from": "compute_graph.l2_184.out[5]", + "to": "compute_graph.l2l3_184_spill.in[1]" + } + ], + "layer_name": "Conv_299", + "layer_object_name": "compute_graph.flexml_layers[162]", + "layer_order": 184, + "mlir_dag_id_map": { + "dag_layer": 184, + "mlir_layer": 184 + }, + "num_iter": 36, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[162].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[162].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[162].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[162].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[162].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[162].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[162].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[162].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[162].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[162].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[162].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[162].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[162].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[162].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[162].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[162].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1280 + ], + "l1_pong": [ + "0x4000", + 1280 + ], + "l2": [ + [ + 1, + 0, + 311296, + 368640, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_184" + ], + "l3": { + "l2l3_184_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_184_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[162].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[162].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[162].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[162].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[162].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[162].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[162].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[162].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[162].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[162].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[162].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[162].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[162].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[162].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[162].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[162].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe180", + 3360 + ], + "l1_pong": [ + "0x9400", + 3360 + ], + "l2": [ + [ + 3, + 0, + 201728, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_184_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 4992000, + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_184_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0185": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_184" + ], + "out": [ + "compute_graph.l2_185" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[163].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[163].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[163].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[163].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[163].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[163].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[163].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[163].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[163].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[163].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[163].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[163].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[163].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[163].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[163].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[163].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 1, + 0, + 311296, + 368640, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_184" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_184", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_301", + "layer_object_name": "compute_graph.flexml_layers[163]", + "layer_order": 185, + "mlir_dag_id_map": { + "dag_layer": 185, + "mlir_layer": 185 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[163].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[163].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[163].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[163].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[163].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[163].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[163].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[163].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[163].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[163].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[163].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[163].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[163].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[163].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[163].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[163].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_185" + ] + }, + "super_iter": 1 + }, + "0186": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_185" + ], + "out": [ + "compute_graph.l2_186" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[164].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[164].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[164].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[164].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[164].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[164].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[164].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[164].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[164].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[164].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[164].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[164].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[164].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[164].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[164].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[164].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_185" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_185", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_304", + "layer_object_name": "compute_graph.flexml_layers[164]", + "layer_order": 186, + "mlir_dag_id_map": { + "dag_layer": 186, + "mlir_layer": 186 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[164].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[164].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[164].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[164].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[164].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[164].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[164].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[164].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[164].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[164].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[164].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[164].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[164].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[164].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[164].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[164].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_186" + ] + }, + "super_iter": 1 + }, + "0187": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_186" + ], + "out": [ + "compute_graph.l2_187", + "compute_graph.l2l3_187_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[165].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[165].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[165].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[165].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[165].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[165].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[165].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[165].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[165].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[165].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[165].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[165].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[165].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[165].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[165].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[165].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_186" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_186", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_187.out[0]", + "to": "compute_graph.l2l3_187_spill.in[0]" + }, + { + "from": "compute_graph.l2_187.out[1]", + "to": "compute_graph.l2l3_187_spill.in[1]" + } + ], + "layer_name": "Div_306", + "layer_object_name": "compute_graph.flexml_layers[165]", + "layer_order": 187, + "mlir_dag_id_map": { + "dag_layer": 187, + "mlir_layer": 187 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[165].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[165].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[165].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[165].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[165].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[165].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[165].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[165].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[165].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[165].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[165].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[165].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[165].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[165].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[165].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[165].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_187" + ], + "l3": { + "l2l3_187_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_187_spill" + ] + }, + "super_iter": 1 + }, + "0188": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0", + "compute_graph.l2l3_184_spill", + "compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1", + "compute_graph.l2l3_187_spill" + ], + "out": [ + "compute_graph.l2_188", + "compute_graph.l2l3_188_spill" + ] + }, + "buffer_iter": 2, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[166].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[166].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[166].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[166].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[166].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[166].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[166].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[166].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[166].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[166].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[166].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[166].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[166].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[166].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[166].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[166].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 293888, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_184_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[166].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[166].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[166].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[166].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[166].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[166].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[166].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[166].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[166].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[166].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[166].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[166].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[166].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[166].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[166].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[166].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 2, + 0, + 63488, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_187_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_188.out[0]", + "to": "compute_graph.l2l3_188_spill.in[0]" + }, + { + "from": "compute_graph.l2_188.out[1]", + "to": "compute_graph.l2l3_188_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_184_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_184_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_184_for_layer_188_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_187_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_187_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_187_for_layer_188_port1.in[1]" + } + ], + "layer_name": "Mul_307", + "layer_object_name": "compute_graph.flexml_layers[166]", + "layer_order": 188, + "mlir_dag_id_map": { + "dag_layer": 188, + "mlir_layer": 188 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[166].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[166].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[166].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[166].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[166].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[166].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[166].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[166].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[166].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[166].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[166].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[166].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[166].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[166].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[166].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[166].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 153600, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_188" + ], + "l2_force_single_buffering": true, + "l3": { + "l2l3_188_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_188_spill" + ] + }, + "super_iter": 2 + }, + "0189": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0", + "compute_graph.l2l3_188_spill" + ], + "out": [ + "compute_graph.l2_189", + "compute_graph.l2l3_189_spill" + ], + "wts": [ + "compute_graph.Layer_189_l2_wts", + "compute_graph.Layer_189_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[167].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[167].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[167].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[167].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[167].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[167].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[167].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[167].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[167].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[167].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[167].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[167].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[167].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[167].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[167].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[167].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5120 + ], + "l1_pong": [ + "0xcd80", + 5120 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_188_spill" + ] + }, + "kernel_name": [ + "superkernel_conv2d_dwc" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_189.out[4]", + "to": "compute_graph.l2l3_189_spill.in[0]" + }, + { + "from": "compute_graph.l2_189.out[5]", + "to": "compute_graph.l2l3_189_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_188_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_188_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_188_for_layer_189_port0.in[1]" + } + ], + "layer_name": "Conv_308", + "layer_object_name": "compute_graph.flexml_layers[167]", + "layer_order": 189, + "mlir_dag_id_map": { + "dag_layer": 189, + "mlir_layer": 189 + }, + "num_iter": 180, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[167].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[167].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[167].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[167].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[167].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[167].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[167].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[167].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[167].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[167].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[167].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[167].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[167].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[167].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[167].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[167].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 128 + ], + "l1_pong": [ + "0x4000", + 128 + ], + "l2": [ + [ + 1, + 0, + 311296, + 368640, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_189" + ], + "l3": { + "l2l3_189_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_189_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[167].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[167].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[167].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[167].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[167].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[167].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[167].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[167].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[167].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[167].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[167].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[167].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[167].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[167].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[167].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[167].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xf580", + 896 + ], + "l1_pong": [ + "0xa800", + 896 + ], + "l2": [ + [ + 3, + 0, + 0, + 107520 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_189_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 5314560, + "size": 107520 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_189_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0190": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_189" + ], + "out": [ + "compute_graph.l2_190" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[168].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[168].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[168].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[168].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[168].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[168].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[168].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[168].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[168].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[168].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[168].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[168].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[168].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[168].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[168].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[168].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 1, + 0, + 311296, + 368640, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_189" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_189", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_310", + "layer_object_name": "compute_graph.flexml_layers[168]", + "layer_order": 190, + "mlir_dag_id_map": { + "dag_layer": 190, + "mlir_layer": 190 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[168].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[168].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[168].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[168].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[168].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[168].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[168].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[168].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[168].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[168].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[168].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[168].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[168].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[168].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[168].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[168].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_190" + ] + }, + "super_iter": 1 + }, + "0191": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_190" + ], + "out": [ + "compute_graph.l2_191" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[169].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[169].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[169].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[169].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[169].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[169].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[169].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[169].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[169].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[169].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[169].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[169].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[169].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[169].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[169].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[169].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_190" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_190", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_313", + "layer_object_name": "compute_graph.flexml_layers[169]", + "layer_order": 191, + "mlir_dag_id_map": { + "dag_layer": 191, + "mlir_layer": 191 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[169].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[169].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[169].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[169].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[169].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[169].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[169].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[169].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[169].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[169].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[169].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[169].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[169].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[169].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[169].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[169].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_191" + ] + }, + "super_iter": 1 + }, + "0192": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_191" + ], + "out": [ + "compute_graph.l2_192", + "compute_graph.l2l3_192_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[170].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[170].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[170].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[170].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[170].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[170].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[170].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[170].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[170].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[170].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[170].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[170].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[170].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[170].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[170].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[170].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_191" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_191", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_192.out[0]", + "to": "compute_graph.l2l3_192_spill.in[0]" + }, + { + "from": "compute_graph.l2_192.out[1]", + "to": "compute_graph.l2l3_192_spill.in[1]" + } + ], + "layer_name": "Div_315", + "layer_object_name": "compute_graph.flexml_layers[170]", + "layer_order": 192, + "mlir_dag_id_map": { + "dag_layer": 192, + "mlir_layer": 192 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[170].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[170].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[170].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[170].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[170].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[170].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[170].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[170].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[170].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[170].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[170].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[170].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[170].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[170].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[170].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[170].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_192" + ], + "l3": { + "l2l3_192_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_192_spill" + ] + }, + "super_iter": 1 + }, + "0193": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0", + "compute_graph.l2l3_189_spill", + "compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1", + "compute_graph.l2l3_192_spill" + ], + "out": [ + "compute_graph.l2_193", + "compute_graph.l2l3_193_spill" + ] + }, + "buffer_iter": 2, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[171].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[171].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[171].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[171].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[171].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[171].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[171].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[171].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[171].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[171].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[171].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[171].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[171].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[171].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[171].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[171].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 293888, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_189_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[171].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[171].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[171].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[171].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[171].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[171].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[171].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[171].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[171].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[171].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[171].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[171].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[171].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[171].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[171].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[171].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 2, + 0, + 63488, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_192_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_193.out[0]", + "to": "compute_graph.l2l3_193_spill.in[0]" + }, + { + "from": "compute_graph.l2_193.out[1]", + "to": "compute_graph.l2l3_193_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_189_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_189_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_189_for_layer_193_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_192_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_192_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_192_for_layer_193_port1.in[1]" + } + ], + "layer_name": "Mul_316", + "layer_object_name": "compute_graph.flexml_layers[171]", + "layer_order": 193, + "mlir_dag_id_map": { + "dag_layer": 193, + "mlir_layer": 193 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[171].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[171].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[171].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[171].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[171].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[171].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[171].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[171].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[171].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[171].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[171].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[171].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[171].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[171].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[171].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[171].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 153600, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_193" + ], + "l2_force_single_buffering": true, + "l3": { + "l2l3_193_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_193_spill" + ] + }, + "super_iter": 2 + }, + "0194": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_193_spill" + ], + "out": [ + "compute_graph.l2l3_194_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_193_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_193_spill" + ] + }, + "kernel_name": [ + "Transpose4dAdf" + ], + "layer_name": "Generated-#48", + "layer_object_name": "compute_graph.templated_graph_194", + "layer_order": 194, + "mlir_dag_id_map": { + "dag_layer": 194, + "mlir_layer": 194 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_194_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_194_spill" + ] + }, + "templated_graph": 1 + }, + "0195": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0", + "compute_graph.l2l3_194_spill" + ], + "out": [ + "compute_graph.l2_195" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[172].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[172].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[172].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[172].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[172].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[172].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[172].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[172].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[172].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[172].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[172].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[172].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[172].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[172].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[172].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[172].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_194_spill" + ] + }, + "kernel_name": [ + "superkernel_reduce_mean_c8" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_194_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_194_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_194_for_layer_195_port0.in[1]" + } + ], + "layer_name": "Generated-#50", + "layer_object_name": "compute_graph.flexml_layers[172]", + "layer_order": 195, + "mlir_dag_id_map": { + "dag_layer": 195, + "mlir_layer": 195 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[172].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[172].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[172].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[172].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[172].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[172].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[172].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[172].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[172].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[172].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[172].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[172].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[172].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[172].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[172].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[172].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb000", + 40 + ], + "l1_pong": [ + "0xcd80", + 40 + ], + "l2": [ + [ + 2, + 0, + 516608, + 3840, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_195" + ] + }, + "super_iter": 1 + }, + "0196": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_195" + ], + "out": [ + "compute_graph.l2_196" + ], + "wts": [ + "compute_graph.Layer_196_l2_wts", + "compute_graph.Layer_196_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[173].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[173].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[173].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[173].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[173].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[173].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[173].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[173].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[173].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[173].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[173].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[173].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[173].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[173].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[173].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[173].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3072 + ], + "l1_pong": [ + "0xcd80", + 3072 + ], + "l2": [ + [ + 2, + 0, + 516608, + 3840, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_195" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_195", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_318", + "layer_object_name": "compute_graph.flexml_layers[173]", + "layer_order": 196, + "mlir_dag_id_map": { + "dag_layer": 196, + "mlir_layer": 196 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[173].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[173].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[173].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[173].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[173].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[173].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[173].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[173].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[173].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[173].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[173].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[173].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[173].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[173].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[173].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[173].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 0, + 0, + 0, + 16384, + 240 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_196" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[173].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[173].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[173].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[173].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[173].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[173].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[173].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[173].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[173].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[173].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[173].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[173].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[173].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[173].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[173].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[173].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe580", + 3136 + ], + "l1_pong": [ + "0x9800", + 3136 + ], + "l2": [ + [ + 3, + 0, + 0, + 250880 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_196_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 5529600, + "size": 250880 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_196_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0197": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_196" + ], + "out": [ + "compute_graph.l2_197" + ], + "wts": [ + "compute_graph.Layer_197_l2_wts", + "compute_graph.Layer_197_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 3, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[174].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[174].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[174].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[174].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[174].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[174].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[174].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[174].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[174].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[174].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[174].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[174].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[174].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[174].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[174].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[174].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 640 + ], + "l1_pong": [ + "0xcd80", + 640 + ], + "l2": [ + [ + 0, + 0, + 0, + 16384, + 240 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_196" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_196", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_320", + "layer_object_name": "compute_graph.flexml_layers[174]", + "layer_order": 197, + "mlir_dag_id_map": { + "dag_layer": 197, + "mlir_layer": 197 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[174].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[174].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[174].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[174].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[174].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[174].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[174].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[174].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[174].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[174].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[174].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[174].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[174].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[174].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[174].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[174].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 384 + ], + "l1_pong": [ + "0x4000", + 384 + ], + "l2": [ + [ + 2, + 0, + 462848, + 30720, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_197" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[174].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[174].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[174].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[174].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[174].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[174].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[174].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[174].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[174].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[174].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[174].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[174].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[174].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[174].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[174].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[174].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xd280", + 4032 + ], + "l1_pong": [ + "0x8500", + 4032 + ], + "l2": [ + [ + 3, + 0, + 40448, + 241920 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_197_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 6031360, + "size": 241920 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_197_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0198": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_197" + ], + "out": [ + "compute_graph.l2_198" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[175].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[175].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[175].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[175].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[175].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[175].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[175].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[175].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[175].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[175].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[175].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[175].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[175].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[175].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[175].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[175].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 2, + 0, + 462848, + 30720, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_197" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_197", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_322", + "layer_object_name": "compute_graph.flexml_layers[175]", + "layer_order": 198, + "mlir_dag_id_map": { + "dag_layer": 198, + "mlir_layer": 198 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[175].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[175].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[175].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[175].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[175].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[175].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[175].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[175].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[175].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[175].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[175].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[175].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[175].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[175].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[175].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[175].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaf80", + 384 + ], + "l1_pong": [ + "0xcd80", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_198" + ] + }, + "super_iter": 1 + }, + "0199": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_198" + ], + "out": [ + "compute_graph.l2_199" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[176].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[176].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[176].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[176].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[176].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[176].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[176].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[176].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[176].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[176].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[176].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[176].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[176].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[176].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[176].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[176].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 6144, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_198" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_198", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_325", + "layer_object_name": "compute_graph.flexml_layers[176]", + "layer_order": 199, + "mlir_dag_id_map": { + "dag_layer": 199, + "mlir_layer": 199 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[176].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[176].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[176].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[176].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[176].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[176].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[176].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[176].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[176].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[176].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[176].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[176].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[176].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[176].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[176].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[176].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb080", + 256 + ], + "l1_pong": [ + "0xcd80", + 256 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_199" + ] + }, + "super_iter": 1 + }, + "0200": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_199" + ], + "out": [ + "compute_graph.l2_200", + "compute_graph.l2l3_200_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[177].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[177].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[177].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[177].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[177].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[177].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[177].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[177].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[177].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[177].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[177].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[177].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[177].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[177].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[177].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[177].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 2, + 0, + 516096, + 4096, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_199" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_199", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_200.out[0]", + "to": "compute_graph.l2l3_200_spill.in[0]" + }, + { + "from": "compute_graph.l2_200.out[1]", + "to": "compute_graph.l2l3_200_spill.in[1]" + } + ], + "layer_name": "Div_327", + "layer_object_name": "compute_graph.flexml_layers[177]", + "layer_order": 200, + "mlir_dag_id_map": { + "dag_layer": 200, + "mlir_layer": 200 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[177].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[177].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[177].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[177].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[177].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[177].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[177].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[177].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[177].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[177].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[177].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[177].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[177].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[177].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[177].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[177].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xae80", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 8192, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_200" + ], + "l3": { + "l2l3_200_spill": { + "size": 960 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_200_spill" + ] + }, + "super_iter": 1 + }, + "0201": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_200_spill", + "compute_graph.l2l3_scratch_0_201_spill", + "compute_graph.l2l3_scratch_1_201_spill" + ], + "out": [ + "compute_graph.l2l3_201_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_200_spill": { + "size": 960 + }, + "l2l3_scratch_0_201_spill": { + "size": 460800 + }, + "l2l3_scratch_1_201_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_200_spill", + "compute_graph.l2l3_scratch_0_201_spill", + "compute_graph.l2l3_scratch_1_201_spill" + ] + }, + "kernel_name": [ + "TileAdf" + ], + "layer_name": "Generated-#52", + "layer_object_name": "compute_graph.templated_graph_201", + "layer_order": 201, + "mlir_dag_id_map": { + "dag_layer": 201, + "mlir_layer": 201 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_201_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_201_spill" + ] + }, + "templated_graph": 1 + }, + "0202": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1", + "compute_graph.l2l3_193_spill", + "compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0", + "compute_graph.l2l3_201_spill" + ], + "out": [ + "compute_graph.l2_202" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[178].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[178].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[178].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[178].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[178].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[178].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[178].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[178].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[178].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[178].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[178].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[178].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[178].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[178].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[178].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[178].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 460800, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_201_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[178].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[178].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[178].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[178].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[178].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[178].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[178].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[178].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[178].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[178].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[178].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[178].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[178].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[178].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[178].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[178].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_193_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_193_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_193_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_193_for_layer_202_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_201_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_201_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_201_for_layer_202_port0.in[1]" + } + ], + "layer_name": "Mul_328", + "layer_object_name": "compute_graph.flexml_layers[178]", + "layer_order": 202, + "mlir_dag_id_map": { + "dag_layer": 202, + "mlir_layer": 202 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[178].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[178].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[178].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[178].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[178].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[178].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[178].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[178].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[178].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[178].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[178].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[178].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[178].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[178].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[178].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[178].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_202" + ] + }, + "super_iter": 1 + }, + "0203": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_202", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183" + ], + "out": [ + "compute_graph.l2_203" + ], + "wts": [ + "compute_graph.Layer_203_l2_wts", + "compute_graph.Layer_203_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 9, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[179].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[179].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[179].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[179].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[179].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[179].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[179].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[179].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[179].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[179].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[179].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[179].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[179].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[179].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[179].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[179].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3584 + ], + "l1_pong": [ + "0xcd80", + 3584 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_202" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[179].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[179].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[179].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[179].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[179].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[179].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[179].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[179].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[179].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[179].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[179].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[179].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[179].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[179].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[179].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[179].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 3072 + ], + "l1_pong": [ + "0x6000", + 3072 + ], + "l2": [ + [ + 0, + 0, + 0, + 73728, + 38400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_202", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_183183.in[1]" + } + ], + "layer_name": "Conv_329", + "layer_object_name": "compute_graph.flexml_layers[179]", + "layer_order": 203, + "mlir_dag_id_map": { + "dag_layer": 203, + "mlir_layer": 203 + }, + "num_iter": 54, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[179].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[179].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[179].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[179].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[179].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[179].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[179].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[179].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[179].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[179].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[179].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[179].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[179].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[179].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[179].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[179].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 768 + ], + "l1_pong": [ + "0x4000", + 768 + ], + "l2": [ + [ + 0, + 0, + 0, + 73728, + 38400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_203" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[179].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[179].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[179].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[179].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[179].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[179].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[179].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[179].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[179].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[179].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[179].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[179].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[179].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[179].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[179].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[179].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe980", + 2784 + ], + "l1_pong": [ + "0x9c00", + 2784 + ], + "l2": [ + [ + 3, + 0, + 0, + 200448 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_203_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 6515200, + "size": 200448 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_203_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0204": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_203" + ], + "out": [ + "compute_graph.l2_204", + "compute_graph.l2l3_204_spill" + ], + "wts": [ + "compute_graph.Layer_204_l2_wts", + "compute_graph.Layer_204_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[180].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[180].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[180].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[180].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[180].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[180].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[180].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[180].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[180].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[180].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[180].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[180].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[180].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[180].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[180].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[180].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2560 + ], + "l1_pong": [ + "0xcd80", + 2560 + ], + "l2": [ + [ + 0, + 0, + 0, + 73728, + 38400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_203" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_203", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_204.out[4]", + "to": "compute_graph.l2l3_204_spill.in[0]" + }, + { + "from": "compute_graph.l2_204.out[5]", + "to": "compute_graph.l2l3_204_spill.in[1]" + } + ], + "layer_name": "Conv_331", + "layer_object_name": "compute_graph.flexml_layers[180]", + "layer_order": 204, + "mlir_dag_id_map": { + "dag_layer": 204, + "mlir_layer": 204 + }, + "num_iter": 36, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[180].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[180].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[180].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[180].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[180].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[180].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[180].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[180].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[180].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[180].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[180].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[180].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[180].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[180].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[180].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[180].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1280 + ], + "l1_pong": [ + "0x4000", + 1280 + ], + "l2": [ + [ + 1, + 0, + 311296, + 368640, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_204" + ], + "l3": { + "l2l3_204_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_204_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[180].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[180].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[180].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[180].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[180].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[180].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[180].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[180].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[180].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[180].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[180].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[180].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[180].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[180].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[180].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[180].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe180", + 3360 + ], + "l1_pong": [ + "0x9400", + 3360 + ], + "l2": [ + [ + 3, + 0, + 201728, + 161280 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_204_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 6916096, + "size": 161280 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_204_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0205": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_204" + ], + "out": [ + "compute_graph.l2_205" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[181].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[181].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[181].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[181].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[181].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[181].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[181].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[181].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[181].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[181].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[181].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[181].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[181].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[181].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[181].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[181].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 1, + 0, + 311296, + 368640, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_204" + ] + }, + "kernel_name": [ + "superkernel_add1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_204", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Add_333", + "layer_object_name": "compute_graph.flexml_layers[181]", + "layer_order": 205, + "mlir_dag_id_map": { + "dag_layer": 205, + "mlir_layer": 205 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[181].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[181].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[181].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[181].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[181].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[181].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[181].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[181].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[181].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[181].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[181].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[181].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[181].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[181].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[181].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[181].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_205" + ] + }, + "super_iter": 1 + }, + "0206": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_205" + ], + "out": [ + "compute_graph.l2_206" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[182].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[182].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[182].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[182].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[182].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[182].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[182].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[182].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[182].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[182].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[182].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[182].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[182].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[182].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[182].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[182].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_205" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_205", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Clip_336", + "layer_object_name": "compute_graph.flexml_layers[182]", + "layer_order": 206, + "mlir_dag_id_map": { + "dag_layer": 206, + "mlir_layer": 206 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[182].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[182].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[182].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[182].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[182].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[182].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[182].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[182].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[182].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[182].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[182].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[182].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[182].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[182].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[182].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[182].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_206" + ] + }, + "super_iter": 1 + }, + "0207": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_206" + ], + "out": [ + "compute_graph.l2_207", + "compute_graph.l2l3_207_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[183].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[183].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[183].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[183].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[183].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[183].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[183].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[183].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[183].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[183].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[183].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[183].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[183].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[183].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[183].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[183].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 2, + 0, + 63488, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_206" + ] + }, + "kernel_name": [ + "superkernel_mul1d_attribute_broadcasting" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_206", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_207.out[0]", + "to": "compute_graph.l2l3_207_spill.in[0]" + }, + { + "from": "compute_graph.l2_207.out[1]", + "to": "compute_graph.l2l3_207_spill.in[1]" + } + ], + "layer_name": "Div_338", + "layer_object_name": "compute_graph.flexml_layers[183]", + "layer_order": 207, + "mlir_dag_id_map": { + "dag_layer": 207, + "mlir_layer": 207 + }, + "num_iter": 9, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[183].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[183].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[183].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[183].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[183].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[183].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[183].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[183].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[183].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[183].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[183].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[183].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[183].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[183].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[183].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[183].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_207" + ], + "l3": { + "l2l3_207_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_207_spill" + ] + }, + "super_iter": 1 + }, + "0208": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0", + "compute_graph.l2l3_204_spill", + "compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1", + "compute_graph.l2l3_207_spill" + ], + "out": [ + "compute_graph.l2_208", + "compute_graph.l2l3_208_spill" + ] + }, + "buffer_iter": 2, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[184].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[184].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[184].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[184].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[184].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[184].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[184].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[184].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[184].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[184].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[184].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[184].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[184].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[184].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[184].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[184].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 293888, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_204_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[184].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[184].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[184].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[184].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[184].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[184].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[184].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[184].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[184].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[184].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[184].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[184].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[184].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[184].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[184].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[184].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 2, + 0, + 63488, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1" + ], + "l2_force_single_buffering": true, + "l3_buffer_names": [ + "compute_graph.l2l3_207_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_208.out[0]", + "to": "compute_graph.l2l3_208_spill.in[0]" + }, + { + "from": "compute_graph.l2_208.out[1]", + "to": "compute_graph.l2l3_208_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_204_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_204_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_204_for_layer_208_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_207_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_207_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_207_for_layer_208_port1.in[1]" + } + ], + "layer_name": "Mul_339", + "layer_object_name": "compute_graph.flexml_layers[184]", + "layer_order": 208, + "mlir_dag_id_map": { + "dag_layer": 208, + "mlir_layer": 208 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[184].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[184].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[184].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[184].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[184].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[184].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[184].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[184].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[184].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[184].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[184].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[184].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[184].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[184].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[184].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[184].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 153600, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_208" + ], + "l2_force_single_buffering": true, + "l3": { + "l2l3_208_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_208_spill" + ] + }, + "super_iter": 2 + }, + "0209": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_208_spill" + ], + "out": [ + "compute_graph.l2l3_209_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_208_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_208_spill" + ] + }, + "kernel_name": [ + "Transpose4dAdf" + ], + "layer_name": "Generated-#54", + "layer_object_name": "compute_graph.templated_graph_209", + "layer_order": 209, + "mlir_dag_id_map": { + "dag_layer": 209, + "mlir_layer": 209 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_209_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_209_spill" + ] + }, + "templated_graph": 1 + }, + "0210": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0", + "compute_graph.l2l3_209_spill" + ], + "out": [ + "compute_graph.l2_210" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[185].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[185].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[185].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[185].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[185].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[185].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[185].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[185].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[185].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[185].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[185].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[185].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[185].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[185].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[185].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[185].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_209_spill" + ] + }, + "kernel_name": [ + "superkernel_reduce_mean_c8" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_209_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_209_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_209_for_layer_210_port0.in[1]" + } + ], + "layer_name": "Generated-#56", + "layer_object_name": "compute_graph.flexml_layers[185]", + "layer_order": 210, + "mlir_dag_id_map": { + "dag_layer": 210, + "mlir_layer": 210 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[185].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[185].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[185].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[185].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[185].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[185].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[185].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[185].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[185].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[185].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[185].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[185].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[185].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[185].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[185].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[185].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xb000", + 40 + ], + "l1_pong": [ + "0xcd80", + 40 + ], + "l2": [ + [ + 2, + 0, + 516608, + 3840, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_210" + ] + }, + "super_iter": 1 + }, + "0211": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_210" + ], + "out": [ + "compute_graph.l2_211" + ], + "wts": [ + "compute_graph.Layer_211_l2_wts", + "compute_graph.Layer_211_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[186].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[186].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[186].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[186].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[186].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[186].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[186].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[186].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[186].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[186].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[186].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[186].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[186].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[186].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[186].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[186].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3072 + ], + "l1_pong": [ + "0xcd80", + 3072 + ], + "l2": [ + [ + 2, + 0, + 516608, + 3840, + 960 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_210" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_210", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Conv_343", + "layer_object_name": "compute_graph.flexml_layers[186]", + "layer_order": 211, + "mlir_dag_id_map": { + "dag_layer": 211, + "mlir_layer": 211 + }, + "num_iter": 10, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[186].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[186].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[186].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[186].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[186].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[186].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[186].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[186].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[186].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[186].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[186].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[186].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[186].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[186].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[186].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[186].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 256 + ], + "l1_pong": [ + "0x4000", + 256 + ], + "l2": [ + [ + 0, + 0, + 0, + 8192, + 128 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_211" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[186].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[186].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[186].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[186].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[186].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[186].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[186].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[186].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[186].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[186].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[186].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[186].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[186].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[186].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[186].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[186].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe580", + 3136 + ], + "l1_pong": [ + "0x9800", + 3136 + ], + "l2": [ + [ + 3, + 0, + 273408, + 125440 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_211_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 7238656, + "size": 125440 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_211_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0212": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_211" + ], + "out": [ + "compute_graph.l2_212", + "compute_graph.l2l3_212_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[187].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[187].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[187].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[187].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[187].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[187].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[187].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[187].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[187].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[187].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[187].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[187].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[187].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[187].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[187].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[187].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 0, + 0, + 0, + 8192, + 128 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_211" + ] + }, + "kernel_name": [ + "superkernel_sigmoid1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_211", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_212.out[0]", + "to": "compute_graph.l2l3_212_spill.in[0]" + }, + { + "from": "compute_graph.l2_212.out[1]", + "to": "compute_graph.l2l3_212_spill.in[1]" + } + ], + "layer_name": "Sigmoid_344", + "layer_object_name": "compute_graph.flexml_layers[187]", + "layer_order": 212, + "mlir_dag_id_map": { + "dag_layer": 212, + "mlir_layer": 212 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[187].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[187].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[187].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[187].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[187].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[187].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[187].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[187].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[187].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[187].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[187].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[187].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[187].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[187].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[187].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[187].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xae80", + 512 + ], + "l1_pong": [ + "0xcd80", + 512 + ], + "l2": [ + [ + 0, + 0, + 16384, + 8192, + 128 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_212" + ], + "l3": { + "l2l3_212_spill": { + "size": 128 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_212_spill" + ] + }, + "super_iter": 1 + }, + "0213": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_212_spill", + "compute_graph.l2l3_scratch_0_213_spill", + "compute_graph.l2l3_scratch_1_213_spill" + ], + "out": [ + "compute_graph.l2l3_213_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_212_spill": { + "size": 128 + }, + "l2l3_scratch_0_213_spill": { + "size": 61440 + }, + "l2l3_scratch_1_213_spill": { + "size": 61440 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_212_spill", + "compute_graph.l2l3_scratch_0_213_spill", + "compute_graph.l2l3_scratch_1_213_spill" + ] + }, + "kernel_name": [ + "TileAdf" + ], + "layer_name": "Generated-#58", + "layer_object_name": "compute_graph.templated_graph_213", + "layer_order": 213, + "mlir_dag_id_map": { + "dag_layer": 213, + "mlir_layer": 213 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_213_spill": { + "size": 30720 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_213_spill" + ] + }, + "templated_graph": 1 + }, + "0214": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0", + "compute_graph.l2l3_208_spill", + "compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1", + "compute_graph.l2l3_213_spill" + ], + "out": [ + "compute_graph.l2_214", + "compute_graph.l2l3_214_spill" + ], + "wts": [ + "compute_graph.Layer_214_l2_wts", + "compute_graph.Layer_214_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 10, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[188].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[188].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[188].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[188].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[188].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[188].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[188].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[188].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[188].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[188].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[188].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[188].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[188].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[188].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[188].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[188].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3072 + ], + "l1_pong": [ + "0xcd80", + 3072 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_208_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[188].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[188].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[188].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[188].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[188].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[188].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[188].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[188].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[188].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[188].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[188].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[188].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[188].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[188].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[188].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[188].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x2000", + 4096 + ], + "l1_pong": [ + "0x6000", + 4096 + ], + "l2": [ + [ + 0, + 0, + 460800, + 49152, + 30720 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_213_spill" + ] + }, + "kernel_name": [ + "superkernel_conv_eltbinary" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_214.out[0]", + "to": "compute_graph.l2l3_214_spill.in[0]" + }, + { + "from": "compute_graph.l2_214.out[1]", + "to": "compute_graph.l2l3_214_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_208_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_208_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_208_for_layer_214_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_213_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_213_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_213_for_layer_214_port1.in[1]" + } + ], + "layer_name": "Conv_340", + "layer_object_name": "compute_graph.flexml_layers[188]", + "layer_order": 214, + "mlir_dag_id_map": { + "dag_layer": 214, + "mlir_layer": 214 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[188].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[188].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[188].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[188].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[188].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[188].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[188].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[188].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[188].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[188].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[188].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[188].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[188].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[188].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[188].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[188].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 0, + 0, + 460800, + 49152, + 30720 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_214" + ], + "l3": { + "l2l3_214_spill": { + "size": 30720 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_214_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[188].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[188].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[188].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[188].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[188].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[188].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[188].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[188].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[188].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[188].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[188].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[188].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[188].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[188].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[188].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[188].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe580", + 3200 + ], + "l1_pong": [ + "0x9800", + 3200 + ], + "l2": [ + [ + 3, + 0, + 268288, + 128000 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_214_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 7489536, + "size": 128000 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_214_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0215": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_214_spill" + ], + "out": [ + "compute_graph.l2l3_215_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_214_spill": { + "size": 30720 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_214_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_230.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_230.in[1]" + } + ], + "layer_name": "Split_349_Duplicated#0", + "layer_object_name": "compute_graph.templated_graph_215", + "layer_order": 215, + "mlir_dag_id_map": { + "dag_layer": 215, + "mlir_layer": 215 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_215_spill": { + "size": 15360 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_215_spill" + ] + }, + "templated_graph": 1 + }, + "0216": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_214_spill" + ], + "out": [ + "compute_graph.l2l3_216_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_214_spill": { + "size": 30720 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_214_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_217.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_217.in[1]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_225.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_225.in[1]" + } + ], + "layer_name": "Split_349_Duplicated#1", + "layer_object_name": "compute_graph.templated_graph_216", + "layer_order": 216, + "mlir_dag_id_map": { + "dag_layer": 216, + "mlir_layer": 216 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_216_spill": { + "size": 15360 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_216_spill" + ] + }, + "templated_graph": 1 + }, + "0218": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0", + "compute_graph.spill_L3_Concat_Buffer_layer_217" + ], + "out": [ + "compute_graph.l2_218" + ], + "wts": [ + "compute_graph.Layer_218_l2_wts", + "compute_graph.Layer_218_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 8, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[189].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[189].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[189].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[189].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[189].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[189].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[189].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[189].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[189].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[189].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[189].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[189].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[189].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[189].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[189].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[189].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2720 + ], + "l1_pong": [ + "0xcd80", + 2720 + ], + "l2": [ + [ + 0, + 0, + 0, + 30720, + 30720 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0" + ], + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_217" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_4.out[0]", + "to": "compute_graph.L2_IFM_Buffer_for_input4_0_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_4.out[1]", + "to": "compute_graph.L2_IFM_Buffer_for_input4_0_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_216_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_216_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_217_port0.in[1]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_217.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.in[0]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_217.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_217_for_layer_218_port0.in[1]" + } + ], + "layer_name": "Conv_351", + "layer_object_name": "compute_graph.flexml_layers[189]", + "layer_order": 218, + "mlir_dag_id_map": { + "dag_layer": 218, + "mlir_layer": 218 + }, + "num_iter": 16, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[189].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[189].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[189].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[189].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[189].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[189].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[189].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[189].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[189].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[189].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[189].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[189].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[189].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[189].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[189].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[189].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 2, + 0, + 391168, + 49152, + 30720 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_218" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[189].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[189].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[189].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[189].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[189].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[189].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[189].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[189].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[189].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[189].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[189].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[189].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[189].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[189].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[189].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[189].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe2c0", + 2368 + ], + "l1_pong": [ + "0x9540", + 2368 + ], + "l2": [ + [ + 2, + 0, + 489472, + 151552 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_218_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 7745536, + "size": 151552 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_218_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0219": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_218" + ], + "out": [ + "compute_graph.l2_219", + "compute_graph.l2l3_219_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[190].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[190].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[190].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[190].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[190].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[190].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[190].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[190].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[190].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[190].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[190].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[190].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[190].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[190].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[190].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[190].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 391168, + 49152, + 30720 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_218" + ] + }, + "kernel_name": [ + "superkernel_sigmoid1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_218", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_219.out[0]", + "to": "compute_graph.l2l3_219_spill.in[0]" + }, + { + "from": "compute_graph.l2_219.out[1]", + "to": "compute_graph.l2l3_219_spill.in[1]" + } + ], + "layer_name": "Sigmoid_352", + "layer_object_name": "compute_graph.flexml_layers[190]", + "layer_order": 219, + "mlir_dag_id_map": { + "dag_layer": 219, + "mlir_layer": 219 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[190].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[190].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[190].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[190].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[190].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[190].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[190].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[190].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[190].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[190].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[190].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[190].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[190].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[190].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[190].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[190].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 30720, + 30720 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_219" + ], + "l3": { + "l2l3_219_spill": { + "size": 30720 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_219_spill" + ] + }, + "super_iter": 1 + }, + "0220": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_219_spill" + ], + "out": [ + "compute_graph.l2l3_220_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_219_spill": { + "size": 30720 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_219_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_353_Duplicated#1", + "layer_object_name": "compute_graph.templated_graph_220", + "layer_order": 220, + "mlir_dag_id_map": { + "dag_layer": 220, + "mlir_layer": 220 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_220_spill": { + "size": 15360 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_220_spill" + ] + }, + "templated_graph": 1 + }, + "0221": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0", + "compute_graph.const_ifm_ddr_3", + "compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1", + "compute_graph.l2l3_220_spill" + ], + "out": [ + "compute_graph.l2_221" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[191].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[191].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[191].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[191].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[191].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[191].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[191].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[191].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[191].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[191].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[191].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[191].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[191].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[191].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[191].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[191].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0" + ], + "l3": { + "const_ifm_ddr_3": { + "size": 15360 + } + }, + "l3_buffer_names": [ + "compute_graph.const_ifm_ddr_3" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[191].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[191].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[191].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[191].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[191].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[191].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[191].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[191].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[191].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[191].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[191].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[191].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[191].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[191].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[191].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[191].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 30720, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_220_spill" + ] + }, + "kernel_name": [ + "superkernel_sub1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.const_ifm_ddr_3.out[0]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.in[0]" + }, + { + "from": "compute_graph.const_ifm_ddr_3.out[1]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input3_0_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_220_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_220_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_221_port1.in[1]" + } + ], + "layer_name": "Sub_359", + "layer_object_name": "compute_graph.flexml_layers[191]", + "layer_order": 221, + "mlir_dag_id_map": { + "dag_layer": 221, + "mlir_layer": 221 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[191].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[191].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[191].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[191].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[191].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[191].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[191].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[191].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[191].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[191].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[191].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[191].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[191].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[191].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[191].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[191].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 493568, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_221" + ] + }, + "super_iter": 1 + }, + "0222": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_221", + "compute_graph.L2_IFM_Buffer_for_input4_1_port1", + "compute_graph.ifm_ddr_4" + ], + "out": [ + "compute_graph.l2_222", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[192].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[192].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[192].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[192].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[192].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[192].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[192].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[192].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[192].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[192].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[192].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[192].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[192].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[192].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[192].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[192].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 493568, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_221" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[192].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[192].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[192].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[192].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[192].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[192].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[192].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[192].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[192].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[192].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[192].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[192].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[192].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[192].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[192].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[192].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_for_input4_1_port1" + ], + "l3": { + "ifm_ddr_4": { + "size": 15360 + } + }, + "l3_buffer_names": [ + "compute_graph.ifm_ddr_4" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_for_input4_1_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_221", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_222.out[0]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.in[0]" + }, + { + "from": "compute_graph.l2_222.out[1]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_4.out[2]", + "to": "compute_graph.L2_IFM_Buffer_for_input4_1_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_4.out[3]", + "to": "compute_graph.L2_IFM_Buffer_for_input4_1_port1.in[1]" + } + ], + "layer_name": "Mul_360", + "layer_object_name": "compute_graph.flexml_layers[192]", + "layer_order": 222, + "mlir_dag_id_map": { + "dag_layer": 222, + "mlir_layer": 222 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[192].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[192].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[192].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[192].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[192].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[192].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[192].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[192].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[192].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[192].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[192].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[192].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[192].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[192].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[192].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[192].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 30720, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_222" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_222222": { + "size": 15360 + } + }, + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222" + ] + }, + "super_iter": 1 + }, + "0223": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_219_spill" + ], + "out": [ + "compute_graph.l2l3_223_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_219_spill": { + "size": 30720 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_219_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_353_Duplicated#0", + "layer_object_name": "compute_graph.templated_graph_223", + "layer_order": 223, + "mlir_dag_id_map": { + "dag_layer": 223, + "mlir_layer": 223 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_223_spill": { + "size": 15360 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_223_spill" + ] + }, + "templated_graph": 1 + }, + "0224": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_for_input4_2_port1", + "compute_graph.ifm_ddr_4", + "compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0", + "compute_graph.l2l3_223_spill" + ], + "out": [ + "compute_graph.l2_224", + "compute_graph.spill_L3_Concat_Buffer_layer_225" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[193].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[193].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[193].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[193].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[193].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[193].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[193].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[193].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[193].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[193].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[193].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[193].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[193].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[193].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[193].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[193].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 30720, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_223_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[193].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[193].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[193].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[193].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[193].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[193].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[193].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[193].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[193].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[193].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[193].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[193].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[193].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[193].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[193].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[193].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_for_input4_2_port1" + ], + "l3": { + "ifm_ddr_4": { + "size": 15360 + } + }, + "l3_buffer_names": [ + "compute_graph.ifm_ddr_4" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_for_input4_2_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_224.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_225.in[2]" + }, + { + "from": "compute_graph.l2_224.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_225.in[3]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_4.out[4]", + "to": "compute_graph.L2_IFM_Buffer_for_input4_2_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_4.out[5]", + "to": "compute_graph.L2_IFM_Buffer_for_input4_2_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_223_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_223_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_223_for_layer_224_port0.in[1]" + } + ], + "layer_name": "Mul_354", + "layer_object_name": "compute_graph.flexml_layers[193]", + "layer_order": 224, + "mlir_dag_id_map": { + "dag_layer": 224, + "mlir_layer": 224 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[193].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[193].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[193].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[193].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[193].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[193].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[193].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[193].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[193].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[193].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[193].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[193].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[193].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[193].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[193].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[193].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 61440, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_224" + ], + "l3": { + "spill_L3_Concat_Buffer_layer_225": { + "buffer_bounds": [ + 1, + 64, + 12, + 20 + ], + "concat_offset": [ + 0, + 64, + 0, + 0 + ], + "size": 30720 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_225" + ] + }, + "super_iter": 1 + }, + "0226": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0", + "compute_graph.spill_L3_Concat_Buffer_layer_225" + ], + "out": [ + "compute_graph.l2_226" + ], + "wts": [ + "compute_graph.Layer_226_l2_wts", + "compute_graph.Layer_226_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 8, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[194].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[194].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[194].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[194].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[194].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[194].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[194].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[194].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[194].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[194].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[194].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[194].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[194].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[194].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[194].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[194].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 2720 + ], + "l1_pong": [ + "0xcd80", + 2720 + ], + "l2": [ + [ + 0, + 0, + 0, + 30720, + 30720 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0" + ], + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_225" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_216_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_216_spill.out[3]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_216_for_layer_225_port0.in[1]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_225.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.in[0]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_225.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_225_for_layer_226_port0.in[1]" + } + ], + "layer_name": "Conv_356", + "layer_object_name": "compute_graph.flexml_layers[194]", + "layer_order": 226, + "mlir_dag_id_map": { + "dag_layer": 226, + "mlir_layer": 226 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[194].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[194].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[194].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[194].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[194].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[194].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[194].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[194].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[194].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[194].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[194].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[194].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[194].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[194].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[194].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[194].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1536 + ], + "l1_pong": [ + "0x4000", + 1536 + ], + "l2": [ + [ + 2, + 0, + 475136, + 24576, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_226" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[194].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[194].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[194].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[194].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[194].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[194].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[194].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[194].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[194].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[194].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[194].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[194].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[194].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[194].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[194].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[194].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe2c0", + 2368 + ], + "l1_pong": [ + "0x9540", + 2368 + ], + "l2": [ + [ + 3, + 0, + 372736, + 75776 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_226_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 8048640, + "size": 75776 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_226_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0227": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_226" + ], + "out": [ + "compute_graph.l2_227" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[195].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[195].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[195].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[195].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[195].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[195].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[195].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[195].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[195].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[195].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[195].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[195].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[195].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[195].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[195].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[195].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 475136, + 24576, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_226" + ] + }, + "kernel_name": [ + "superkernel_tanh1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_226", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Tanh_357", + "layer_object_name": "compute_graph.flexml_layers[195]", + "layer_order": 227, + "mlir_dag_id_map": { + "dag_layer": 227, + "mlir_layer": 227 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[195].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[195].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[195].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[195].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[195].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[195].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[195].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[195].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[195].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[195].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[195].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[195].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[195].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[195].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[195].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[195].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 0, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_227" + ] + }, + "super_iter": 1 + }, + "0228": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_227", + "compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0", + "compute_graph.l2l3_220_spill" + ], + "out": [ + "compute_graph.l2_228" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[196].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[196].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[196].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[196].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[196].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[196].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[196].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[196].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[196].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[196].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[196].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[196].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[196].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[196].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[196].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[196].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 30720, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_220_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[196].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[196].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[196].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[196].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[196].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[196].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[196].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[196].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[196].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[196].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[196].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[196].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[196].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[196].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[196].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[196].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_227" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_227", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_220_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_220_spill.out[3]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_220_for_layer_228_port0.in[1]" + } + ], + "layer_name": "Mul_361", + "layer_object_name": "compute_graph.flexml_layers[196]", + "layer_order": 228, + "mlir_dag_id_map": { + "dag_layer": 228, + "mlir_layer": 228 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[196].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[196].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[196].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[196].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[196].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[196].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[196].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[196].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[196].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[196].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[196].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[196].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[196].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[196].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[196].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[196].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 493568, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_228" + ] + }, + "super_iter": 1 + }, + "0229": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_228", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222" + ], + "out": [ + "compute_graph.l2_229", + "compute_graph.ofm_ddr_0_l2l3_229_spill", + "compute_graph.spill_L3_Concat_Buffer_layer_230" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[197].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[197].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[197].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[197].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[197].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[197].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[197].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[197].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[197].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[197].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[197].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[197].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[197].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[197].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[197].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[197].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[197].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[197].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[197].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[197].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[197].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[197].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[197].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[197].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[197].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[197].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[197].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[197].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[197].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[197].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[197].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[197].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 2, + 0, + 493568, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_228" + ] + }, + "kernel_name": [ + "superkernel_add1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_228", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_229.out[0]", + "to": "compute_graph.ofm_ddr_0_l2l3_229_spill.in[0]" + }, + { + "from": "compute_graph.l2_229.out[1]", + "to": "compute_graph.ofm_ddr_0_l2l3_229_spill.in[1]" + }, + { + "from": "compute_graph.l2_229.out[2]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_230.in[2]" + }, + { + "from": "compute_graph.l2_229.out[3]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_230.in[3]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_222222.in[1]" + } + ], + "layer_name": "Add_362", + "layer_object_name": "compute_graph.flexml_layers[197]", + "layer_order": 229, + "mlir_dag_id_map": { + "dag_layer": 229, + "mlir_layer": 229 + }, + "num_iter": 1, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[197].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[197].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[197].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[197].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[197].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[197].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[197].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[197].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[197].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[197].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[197].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[197].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[197].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[197].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[197].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[197].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 30720, + 15360, + 15360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_229" + ], + "l3": { + "ofm_ddr_0_l2l3_229_spill": { + "size": 15360 + }, + "spill_L3_Concat_Buffer_layer_230": { + "buffer_bounds": [ + 1, + 64, + 12, + 20 + ], + "concat_offset": [ + 0, + 64, + 0, + 0 + ], + "size": 30720 + } + }, + "l3_buffer_names": [ + "compute_graph.ofm_ddr_0_l2l3_229_spill", + "compute_graph.spill_L3_Concat_Buffer_layer_230" + ] + }, + "super_iter": 1 + }, + "0231": { + "adf_layer_objects": { + "in": [ + "compute_graph.spill_L3_Concat_Buffer_layer_230" + ], + "out": [ + "compute_graph.l2l3_231_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "spill_L3_Concat_Buffer_layer_230": { + "size": 30720 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_230" + ] + }, + "kernel_name": [ + "ResizeAdf" + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_215_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_215_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_215_for_layer_230_port0.in[1]" + } + ], + "layer_name": "Resize_365", + "layer_object_name": "compute_graph.templated_graph_231", + "layer_order": 231, + "mlir_dag_id_map": { + "dag_layer": 231, + "mlir_layer": 231 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_231_spill": { + "size": 122880 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_231_spill" + ] + }, + "templated_graph": 1 + }, + "0232": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_231_spill" + ], + "out": [ + "compute_graph.l2l3_232_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_231_spill": { + "size": 122880 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_231_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_236.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_236.in[1]" + } + ], + "layer_name": "Slice_371", + "layer_object_name": "compute_graph.templated_graph_232", + "layer_order": 232, + "mlir_dag_id_map": { + "dag_layer": 232, + "mlir_layer": 232 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_232_spill": { + "size": 117760 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_232_spill" + ] + }, + "templated_graph": 1 + }, + "0233": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0", + "compute_graph.l2l3_5_spill" + ], + "out": [ + "compute_graph.l2_233", + "compute_graph.l2l3_233_spill", + "compute_graph.spill_L3_Concat_Buffer_layer_275" + ] + }, + "buffer_iter": 3, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[198].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[198].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[198].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[198].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[198].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[198].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[198].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[198].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[198].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[198].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[198].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[198].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[198].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[198].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[198].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[198].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 5120 + ], + "l1_pong": [ + "0xcd80", + 5120 + ], + "l2": [ + [ + 1, + 0, + 434176, + 153600, + 153600 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_5_spill" + ] + }, + "kernel_name": [ + "superkernel_avgpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_233.out[0]", + "to": "compute_graph.l2l3_233_spill.in[0]" + }, + { + "from": "compute_graph.l2_233.out[1]", + "to": "compute_graph.l2l3_233_spill.in[1]" + }, + { + "from": "compute_graph.l2_233.out[2]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_275.in[4]" + }, + { + "from": "compute_graph.l2_233.out[3]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_275.in[5]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_5_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_5_spill.out[3]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_233_port0.in[1]" + } + ], + "layer_name": "AveragePool_346", + "layer_object_name": "compute_graph.flexml_layers[198]", + "layer_order": 233, + "mlir_dag_id_map": { + "dag_layer": 233, + "mlir_layer": 233 + }, + "num_iter": 96, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[198].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[198].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[198].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[198].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[198].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[198].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[198].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[198].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[198].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[198].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[198].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[198].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[198].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[198].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[198].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[198].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 320 + ], + "l1_pong": [ + "0x4000", + 320 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 38400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_233" + ], + "l3": { + "l2l3_233_spill": { + "size": 115200 + }, + "spill_L3_Concat_Buffer_layer_275": { + "buffer_bounds": [ + 1, + 8, + 30, + 160 + ], + "concat_offset": [ + 0, + 56, + 0, + 0 + ], + "size": 921600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_233_spill", + "compute_graph.spill_L3_Concat_Buffer_layer_275" + ] + }, + "super_iter": 3 + }, + "0234": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0", + "compute_graph.l2l3_233_spill" + ], + "out": [ + "compute_graph.l2_234", + "compute_graph.spill_L3_Concat_Buffer_layer_256" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[199].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[199].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[199].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[199].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[199].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[199].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[199].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[199].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[199].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[199].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[199].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[199].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[199].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[199].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[199].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[199].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 6144 + ], + "l1_pong": [ + "0xcd80", + 6144 + ], + "l2": [ + [ + 0, + 0, + 0, + 115200, + 115200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_233_spill" + ] + }, + "kernel_name": [ + "superkernel_avgpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_234.out[4]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_256.in[4]" + }, + { + "from": "compute_graph.l2_234.out[5]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_256.in[5]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_233_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_233_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_233_for_layer_234_port0.in[1]" + } + ], + "layer_name": "AveragePool_347", + "layer_object_name": "compute_graph.flexml_layers[199]", + "layer_order": 234, + "mlir_dag_id_map": { + "dag_layer": 234, + "mlir_layer": 234 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[199].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[199].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[199].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[199].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[199].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[199].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[199].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[199].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[199].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[199].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[199].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[199].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[199].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[199].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[199].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[199].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 384 + ], + "l1_pong": [ + "0x4000", + 384 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 28800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_234" + ], + "l3": { + "spill_L3_Concat_Buffer_layer_256": { + "buffer_bounds": [ + 1, + 8, + 45, + 80 + ], + "concat_offset": [ + 0, + 104, + 0, + 0 + ], + "size": 403200 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_256" + ] + }, + "super_iter": 1 + }, + "0235": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_234" + ], + "out": [ + "compute_graph.l2_235", + "compute_graph.spill_L3_Concat_Buffer_layer_236" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[200].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[200].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[200].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[200].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[200].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[200].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[200].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[200].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[200].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[200].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[200].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[200].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[200].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[200].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[200].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[200].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 6144 + ], + "l1_pong": [ + "0xcd80", + 6144 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 28800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_234" + ] + }, + "kernel_name": [ + "superkernel_avgpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_234", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_235.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_236.in[4]" + }, + { + "from": "compute_graph.l2_235.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_236.in[5]" + } + ], + "layer_name": "AveragePool_348", + "layer_object_name": "compute_graph.flexml_layers[200]", + "layer_order": 235, + "mlir_dag_id_map": { + "dag_layer": 235, + "mlir_layer": 235 + }, + "num_iter": 5, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[200].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[200].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[200].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[200].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[200].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[200].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[200].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[200].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[200].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[200].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[200].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[200].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[200].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[200].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[200].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[200].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 384 + ], + "l1_pong": [ + "0x4000", + 384 + ], + "l2": [ + [ + 0, + 0, + 0, + 30720, + 7360 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_235" + ], + "l3": { + "spill_L3_Concat_Buffer_layer_236": { + "buffer_bounds": [ + 1, + 8, + 23, + 40 + ], + "concat_offset": [ + 0, + 168, + 0, + 0 + ], + "size": 161920 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_236" + ] + }, + "super_iter": 1 + }, + "0237": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0", + "compute_graph.spill_L3_Concat_Buffer_layer_236" + ], + "out": [ + "compute_graph.l2_237", + "compute_graph.l2l3_237_spill" + ], + "wts": [ + "compute_graph.Layer_237_l2_wts", + "compute_graph.Layer_237_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 11, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[201].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[201].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[201].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[201].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[201].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[201].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[201].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[201].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[201].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[201].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[201].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[201].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[201].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[201].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[201].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[201].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4000 + ], + "l1_pong": [ + "0xcd80", + 4000 + ], + "l2": [ + [ + 0, + 0, + 0, + 161920, + 161920 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0" + ], + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_236" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_237.out[0]", + "to": "compute_graph.l2l3_237_spill.in[0]" + }, + { + "from": "compute_graph.l2_237.out[1]", + "to": "compute_graph.l2l3_237_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_232_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_232_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_232_for_layer_236_port0.in[1]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_236.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.in[0]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_236.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_236_for_layer_237_port0.in[1]" + } + ], + "layer_name": "Conv_373", + "layer_object_name": "compute_graph.flexml_layers[201]", + "layer_order": 237, + "mlir_dag_id_map": { + "dag_layer": 237, + "mlir_layer": 237 + }, + "num_iter": 44, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[201].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[201].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[201].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[201].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[201].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[201].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[201].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[201].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[201].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[201].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[201].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[201].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[201].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[201].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[201].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[201].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 2304 + ], + "l1_pong": [ + "0x4000", + 2304 + ], + "l2": [ + [ + 0, + 0, + 323840, + 147456, + 73600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_237" + ], + "l3": { + "l2l3_237_spill": { + "size": 73600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_237_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[201].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[201].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[201].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[201].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[201].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[201].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[201].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[201].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[201].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[201].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[201].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[201].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[201].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[201].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[201].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[201].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xecc0", + 2368 + ], + "l1_pong": [ + "0x9f40", + 2368 + ], + "l2": [ + [ + 3, + 0, + 0, + 208384 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_237_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 8200192, + "size": 208384 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_237_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0238": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_237_spill" + ], + "out": [ + "compute_graph.l2l3_238_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_237_spill": { + "size": 73600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_237_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_253.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_253.in[1]" + } + ], + "layer_name": "Split_375_Duplicated#0", + "layer_object_name": "compute_graph.templated_graph_238", + "layer_order": 238, + "mlir_dag_id_map": { + "dag_layer": 238, + "mlir_layer": 238 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_238_spill": { + "size": 36800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_238_spill" + ] + }, + "templated_graph": 1 + }, + "0239": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_237_spill" + ], + "out": [ + "compute_graph.l2l3_239_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_237_spill": { + "size": 73600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_237_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_240.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_240.in[1]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_248.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_248.in[1]" + } + ], + "layer_name": "Split_375_Duplicated#1", + "layer_object_name": "compute_graph.templated_graph_239", + "layer_order": 239, + "mlir_dag_id_map": { + "dag_layer": 239, + "mlir_layer": 239 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_239_spill": { + "size": 36800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_239_spill" + ] + }, + "templated_graph": 1 + }, + "0241": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0", + "compute_graph.spill_L3_Concat_Buffer_layer_240" + ], + "out": [ + "compute_graph.l2_241" + ], + "wts": [ + "compute_graph.Layer_241_l2_wts", + "compute_graph.Layer_241_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[202].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[202].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[202].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[202].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[202].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[202].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[202].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[202].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[202].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[202].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[202].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[202].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[202].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[202].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[202].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[202].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4000 + ], + "l1_pong": [ + "0xcd80", + 4000 + ], + "l2": [ + [ + 0, + 0, + 0, + 73600, + 73600 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0" + ], + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_240" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_3.out[0]", + "to": "compute_graph.L2_IFM_Buffer_for_input3_0_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_3.out[1]", + "to": "compute_graph.L2_IFM_Buffer_for_input3_0_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_239_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_239_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_240_port0.in[1]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_240.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.in[0]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_240.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_240_for_layer_241_port0.in[1]" + } + ], + "layer_name": "Conv_377", + "layer_object_name": "compute_graph.flexml_layers[202]", + "layer_order": 241, + "mlir_dag_id_map": { + "dag_layer": 241, + "mlir_layer": 241 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[202].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[202].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[202].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[202].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[202].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[202].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[202].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[202].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[202].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[202].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[202].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[202].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[202].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[202].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[202].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[202].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 2304 + ], + "l1_pong": [ + "0x4000", + 2304 + ], + "l2": [ + [ + 2, + 0, + 229376, + 147456, + 73600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_241" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[202].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[202].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[202].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[202].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[202].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[202].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[202].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[202].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[202].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[202].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[202].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[202].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[202].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[202].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[202].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[202].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xecc0", + 2368 + ], + "l1_pong": [ + "0x9f40", + 2368 + ], + "l2": [ + [ + 3, + 0, + 334848, + 94720 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_241_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 8616960, + "size": 94720 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_241_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0242": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_241" + ], + "out": [ + "compute_graph.l2_242", + "compute_graph.l2l3_242_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[203].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[203].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[203].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[203].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[203].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[203].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[203].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[203].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[203].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[203].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[203].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[203].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[203].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[203].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[203].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[203].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 229376, + 147456, + 73600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_241" + ] + }, + "kernel_name": [ + "superkernel_sigmoid1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_241", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_242.out[0]", + "to": "compute_graph.l2l3_242_spill.in[0]" + }, + { + "from": "compute_graph.l2_242.out[1]", + "to": "compute_graph.l2l3_242_spill.in[1]" + } + ], + "layer_name": "Sigmoid_378", + "layer_object_name": "compute_graph.flexml_layers[203]", + "layer_order": 242, + "mlir_dag_id_map": { + "dag_layer": 242, + "mlir_layer": 242 + }, + "num_iter": 3, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[203].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[203].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[203].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[203].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[203].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[203].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[203].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[203].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[203].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[203].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[203].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[203].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[203].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[203].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[203].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[203].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 92160, + 73600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_242" + ], + "l3": { + "l2l3_242_spill": { + "size": 73600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_242_spill" + ] + }, + "super_iter": 1 + }, + "0243": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_242_spill" + ], + "out": [ + "compute_graph.l2l3_243_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_242_spill": { + "size": 73600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_242_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_379_Duplicated#1", + "layer_object_name": "compute_graph.templated_graph_243", + "layer_order": 243, + "mlir_dag_id_map": { + "dag_layer": 243, + "mlir_layer": 243 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_243_spill": { + "size": 36800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_243_spill" + ] + }, + "templated_graph": 1 + }, + "0244": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0", + "compute_graph.const_ifm_ddr_2", + "compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1", + "compute_graph.l2l3_243_spill" + ], + "out": [ + "compute_graph.l2_244" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[204].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[204].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[204].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[204].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[204].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[204].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[204].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[204].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[204].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[204].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[204].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[204].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[204].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[204].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[204].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[204].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 36800, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0" + ], + "l3": { + "const_ifm_ddr_2": { + "size": 36800 + } + }, + "l3_buffer_names": [ + "compute_graph.const_ifm_ddr_2" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[204].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[204].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[204].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[204].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[204].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[204].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[204].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[204].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[204].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[204].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[204].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[204].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[204].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[204].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[204].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[204].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73600, + 36800, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_243_spill" + ] + }, + "kernel_name": [ + "superkernel_sub1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.const_ifm_ddr_2.out[0]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.in[0]" + }, + { + "from": "compute_graph.const_ifm_ddr_2.out[1]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input2_0_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_243_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_243_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_244_port1.in[1]" + } + ], + "layer_name": "Sub_385", + "layer_object_name": "compute_graph.flexml_layers[204]", + "layer_order": 244, + "mlir_dag_id_map": { + "dag_layer": 244, + "mlir_layer": 244 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[204].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[204].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[204].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[204].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[204].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[204].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[204].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[204].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[204].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[204].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[204].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[204].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[204].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[204].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[204].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[204].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_244" + ] + }, + "super_iter": 1 + }, + "0245": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_244", + "compute_graph.L2_IFM_Buffer_for_input3_1_port1", + "compute_graph.ifm_ddr_3" + ], + "out": [ + "compute_graph.l2_245", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[205].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[205].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[205].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[205].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[205].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[205].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[205].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[205].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[205].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[205].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[205].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[205].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[205].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[205].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[205].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[205].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_244" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[205].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[205].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[205].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[205].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[205].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[205].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[205].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[205].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[205].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[205].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[205].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[205].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[205].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[205].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[205].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[205].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 36800, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_for_input3_1_port1" + ], + "l3": { + "ifm_ddr_3": { + "size": 36800 + } + }, + "l3_buffer_names": [ + "compute_graph.ifm_ddr_3" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_for_input3_1_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_244", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_245.out[0]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.in[0]" + }, + { + "from": "compute_graph.l2_245.out[1]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_3.out[2]", + "to": "compute_graph.L2_IFM_Buffer_for_input3_1_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_3.out[3]", + "to": "compute_graph.L2_IFM_Buffer_for_input3_1_port1.in[1]" + } + ], + "layer_name": "Mul_386", + "layer_object_name": "compute_graph.flexml_layers[205]", + "layer_order": 245, + "mlir_dag_id_map": { + "dag_layer": 245, + "mlir_layer": 245 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[205].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[205].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[205].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[205].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[205].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[205].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[205].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[205].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[205].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[205].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[205].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[205].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[205].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[205].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[205].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[205].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 73600, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_245" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_245245": { + "size": 36800 + } + }, + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245" + ] + }, + "super_iter": 1 + }, + "0246": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_242_spill" + ], + "out": [ + "compute_graph.l2l3_246_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_242_spill": { + "size": 73600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_242_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_379_Duplicated#0", + "layer_object_name": "compute_graph.templated_graph_246", + "layer_order": 246, + "mlir_dag_id_map": { + "dag_layer": 246, + "mlir_layer": 246 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_246_spill": { + "size": 36800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_246_spill" + ] + }, + "templated_graph": 1 + }, + "0247": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_for_input3_2_port1", + "compute_graph.ifm_ddr_3", + "compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0", + "compute_graph.l2l3_246_spill" + ], + "out": [ + "compute_graph.l2_247", + "compute_graph.spill_L3_Concat_Buffer_layer_248" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[206].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[206].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[206].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[206].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[206].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[206].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[206].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[206].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[206].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[206].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[206].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[206].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[206].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[206].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[206].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[206].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 73600, + 36800, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_246_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[206].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[206].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[206].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[206].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[206].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[206].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[206].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[206].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[206].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[206].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[206].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[206].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[206].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[206].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[206].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[206].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 36800, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_for_input3_2_port1" + ], + "l3": { + "ifm_ddr_3": { + "size": 36800 + } + }, + "l3_buffer_names": [ + "compute_graph.ifm_ddr_3" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_for_input3_2_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_247.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_248.in[2]" + }, + { + "from": "compute_graph.l2_247.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_248.in[3]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_3.out[4]", + "to": "compute_graph.L2_IFM_Buffer_for_input3_2_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_3.out[5]", + "to": "compute_graph.L2_IFM_Buffer_for_input3_2_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_246_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_246_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_246_for_layer_247_port0.in[1]" + } + ], + "layer_name": "Mul_380", + "layer_object_name": "compute_graph.flexml_layers[206]", + "layer_order": 247, + "mlir_dag_id_map": { + "dag_layer": 247, + "mlir_layer": 247 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[206].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[206].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[206].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[206].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[206].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[206].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[206].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[206].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[206].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[206].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[206].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[206].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[206].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[206].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[206].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[206].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 147200, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_247" + ], + "l3": { + "spill_L3_Concat_Buffer_layer_248": { + "buffer_bounds": [ + 1, + 40, + 23, + 40 + ], + "concat_offset": [ + 0, + 40, + 0, + 0 + ], + "size": 73600 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_248" + ] + }, + "super_iter": 1 + }, + "0249": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0", + "compute_graph.spill_L3_Concat_Buffer_layer_248" + ], + "out": [ + "compute_graph.l2_249" + ], + "wts": [ + "compute_graph.Layer_249_l2_wts", + "compute_graph.Layer_249_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[207].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[207].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[207].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[207].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[207].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[207].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[207].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[207].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[207].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[207].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[207].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[207].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[207].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[207].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[207].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[207].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4000 + ], + "l1_pong": [ + "0xcd80", + 4000 + ], + "l2": [ + [ + 0, + 0, + 0, + 73600, + 73600 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0" + ], + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_248" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_239_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_239_spill.out[3]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_239_for_layer_248_port0.in[1]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_248.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.in[0]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_248.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_248_for_layer_249_port0.in[1]" + } + ], + "layer_name": "Conv_382", + "layer_object_name": "compute_graph.flexml_layers[207]", + "layer_order": 249, + "mlir_dag_id_map": { + "dag_layer": 249, + "mlir_layer": 249 + }, + "num_iter": 10, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[207].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[207].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[207].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[207].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[207].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[207].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[207].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[207].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[207].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[207].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[207].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[207].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[207].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[207].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[207].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[207].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 2304 + ], + "l1_pong": [ + "0x4000", + 2304 + ], + "l2": [ + [ + 2, + 0, + 376832, + 73728, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_249" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[207].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[207].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[207].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[207].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[207].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[207].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[207].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[207].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[207].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[207].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[207].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[207].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[207].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[207].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[207].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[207].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xecc0", + 2368 + ], + "l1_pong": [ + "0x9f40", + 2368 + ], + "l2": [ + [ + 3, + 0, + 0, + 47360 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_249_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 8806400, + "size": 47360 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_249_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0250": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_249" + ], + "out": [ + "compute_graph.l2_250" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[208].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[208].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[208].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[208].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[208].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[208].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[208].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[208].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[208].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[208].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[208].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[208].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[208].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[208].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[208].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[208].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 376832, + 73728, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_249" + ] + }, + "kernel_name": [ + "superkernel_tanh1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_249", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Tanh_383", + "layer_object_name": "compute_graph.flexml_layers[208]", + "layer_order": 250, + "mlir_dag_id_map": { + "dag_layer": 250, + "mlir_layer": 250 + }, + "num_iter": 2, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[208].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[208].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[208].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[208].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[208].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[208].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[208].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[208].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[208].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[208].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[208].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[208].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[208].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[208].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[208].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[208].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_250" + ] + }, + "super_iter": 1 + }, + "0251": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_250", + "compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0", + "compute_graph.l2l3_243_spill" + ], + "out": [ + "compute_graph.l2_251" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[209].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[209].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[209].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[209].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[209].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[209].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[209].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[209].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[209].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[209].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[209].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[209].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[209].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[209].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[209].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[209].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 122880, + 36800, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_243_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[209].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[209].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[209].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[209].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[209].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[209].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[209].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[209].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[209].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[209].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[209].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[209].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[209].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[209].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[209].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[209].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_250" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_250", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_243_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_243_spill.out[3]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_243_for_layer_251_port0.in[1]" + } + ], + "layer_name": "Mul_387", + "layer_object_name": "compute_graph.flexml_layers[209]", + "layer_order": 251, + "mlir_dag_id_map": { + "dag_layer": 251, + "mlir_layer": 251 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[209].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[209].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[209].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[209].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[209].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[209].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[209].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[209].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[209].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[209].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[209].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[209].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[209].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[209].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[209].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[209].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_251" + ] + }, + "super_iter": 1 + }, + "0252": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_251", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245" + ], + "out": [ + "compute_graph.l2_252", + "compute_graph.ofm_ddr_1_l2l3_252_spill", + "compute_graph.spill_L3_Concat_Buffer_layer_253" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[210].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[210].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[210].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[210].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[210].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[210].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[210].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[210].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[210].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[210].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[210].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[210].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[210].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[210].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[210].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[210].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[210].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[210].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[210].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[210].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[210].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[210].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[210].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[210].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[210].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[210].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[210].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[210].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[210].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[210].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[210].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[210].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 2, + 0, + 401408, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_251" + ] + }, + "kernel_name": [ + "superkernel_add1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_251", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_252.out[0]", + "to": "compute_graph.ofm_ddr_1_l2l3_252_spill.in[0]" + }, + { + "from": "compute_graph.l2_252.out[1]", + "to": "compute_graph.ofm_ddr_1_l2l3_252_spill.in[1]" + }, + { + "from": "compute_graph.l2_252.out[2]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_253.in[2]" + }, + { + "from": "compute_graph.l2_252.out[3]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_253.in[3]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_245245.in[1]" + } + ], + "layer_name": "Add_388", + "layer_object_name": "compute_graph.flexml_layers[210]", + "layer_order": 252, + "mlir_dag_id_map": { + "dag_layer": 252, + "mlir_layer": 252 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[210].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[210].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[210].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[210].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[210].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[210].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[210].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[210].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[210].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[210].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[210].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[210].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[210].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[210].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[210].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[210].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 122880, + 61440, + 36800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_252" + ], + "l3": { + "ofm_ddr_1_l2l3_252_spill": { + "size": 36800 + }, + "spill_L3_Concat_Buffer_layer_253": { + "buffer_bounds": [ + 1, + 40, + 23, + 40 + ], + "concat_offset": [ + 0, + 40, + 0, + 0 + ], + "size": 73600 + } + }, + "l3_buffer_names": [ + "compute_graph.ofm_ddr_1_l2l3_252_spill", + "compute_graph.spill_L3_Concat_Buffer_layer_253" + ] + }, + "super_iter": 1 + }, + "0254": { + "adf_layer_objects": { + "in": [ + "compute_graph.spill_L3_Concat_Buffer_layer_253" + ], + "out": [ + "compute_graph.l2l3_254_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "spill_L3_Concat_Buffer_layer_253": { + "size": 73600 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_253" + ] + }, + "kernel_name": [ + "ResizeAdf" + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_238_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_238_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_238_for_layer_253_port0.in[1]" + } + ], + "layer_name": "Resize_391", + "layer_object_name": "compute_graph.templated_graph_254", + "layer_order": 254, + "mlir_dag_id_map": { + "dag_layer": 254, + "mlir_layer": 254 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_254_spill": { + "size": 294400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_254_spill" + ] + }, + "templated_graph": 1 + }, + "0255": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_254_spill" + ], + "out": [ + "compute_graph.l2l3_255_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_254_spill": { + "size": 294400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_254_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_256.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_256.in[1]" + } + ], + "layer_name": "Slice_397", + "layer_object_name": "compute_graph.templated_graph_255", + "layer_order": 255, + "mlir_dag_id_map": { + "dag_layer": 255, + "mlir_layer": 255 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_255_spill": { + "size": 288000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_255_spill" + ] + }, + "templated_graph": 1 + }, + "0257": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0", + "compute_graph.spill_L3_Concat_Buffer_layer_256" + ], + "out": [ + "compute_graph.l2_257", + "compute_graph.l2l3_257_spill" + ], + "wts": [ + "compute_graph.Layer_257_l2_wts", + "compute_graph.Layer_257_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 7, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[211].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[211].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[211].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[211].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[211].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[211].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[211].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[211].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[211].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[211].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[211].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[211].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[211].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[211].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[211].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[211].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4032 + ], + "l1_pong": [ + "0xcd80", + 4032 + ], + "l2": [ + [ + 0, + 0, + 0, + 403200, + 403200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0" + ], + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_256" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_257.out[0]", + "to": "compute_graph.l2l3_257_spill.in[0]" + }, + { + "from": "compute_graph.l2_257.out[1]", + "to": "compute_graph.l2l3_257_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_255_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_255_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_255_for_layer_256_port0.in[1]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_256.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.in[0]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_256.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_256_for_layer_257_port0.in[1]" + } + ], + "layer_name": "Conv_399", + "layer_object_name": "compute_graph.flexml_layers[211]", + "layer_order": 257, + "mlir_dag_id_map": { + "dag_layer": 257, + "mlir_layer": 257 + }, + "num_iter": 35, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[211].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[211].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[211].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[211].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[211].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[211].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[211].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[211].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[211].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[211].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[211].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[211].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[211].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[211].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[211].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[211].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 3072 + ], + "l1_pong": [ + "0x4000", + 3072 + ], + "l2": [ + [ + 1, + 0, + 282112, + 245760, + 144000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_257" + ], + "l3": { + "l2l3_257_spill": { + "size": 144000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_257_spill" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[211].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[211].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[211].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[211].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[211].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[211].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[211].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[211].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[211].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[211].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[211].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[211].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[211].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[211].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[211].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[211].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xed00", + 2368 + ], + "l1_pong": [ + "0x9f80", + 2368 + ], + "l2": [ + [ + 3, + 0, + 0, + 66304 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_257_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 8901120, + "size": 66304 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_257_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0258": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_257_spill" + ], + "out": [ + "compute_graph.l2l3_258_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_257_spill": { + "size": 144000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_257_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_401_Duplicated#0", + "layer_object_name": "compute_graph.templated_graph_258", + "layer_order": 258, + "mlir_dag_id_map": { + "dag_layer": 258, + "mlir_layer": 258 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_258_spill": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_258_spill" + ] + }, + "templated_graph": 1 + }, + "0259": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_257_spill" + ], + "out": [ + "compute_graph.l2l3_259_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_257_spill": { + "size": 144000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_257_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_401_Duplicated#1", + "layer_object_name": "compute_graph.templated_graph_259", + "layer_order": 259, + "mlir_dag_id_map": { + "dag_layer": 259, + "mlir_layer": 259 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_259_spill": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_259_spill" + ] + }, + "templated_graph": 1 + }, + "0260": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_259_spill", + "compute_graph.ifm_ddr_2" + ], + "out": [ + "compute_graph.l2l3_260_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "ifm_ddr_2": { + "size": 86400 + }, + "l2l3_259_spill": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_259_spill", + "compute_graph.ifm_ddr_2" + ] + }, + "kernel_name": [ + "ConcatC8Adf" + ], + "layer_name": "Concat_402", + "layer_object_name": "compute_graph.templated_graph_260", + "layer_order": 260, + "mlir_dag_id_map": { + "dag_layer": 260, + "mlir_layer": 260 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_260_spill": { + "size": 144000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_260_spill" + ] + }, + "templated_graph": 1 + }, + "0261": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0", + "compute_graph.l2l3_260_spill" + ], + "out": [ + "compute_graph.l2_261" + ], + "wts": [ + "compute_graph.Layer_261_l2_wts", + "compute_graph.Layer_261_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 3, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[212].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[212].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[212].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[212].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[212].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[212].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[212].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[212].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[212].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[212].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[212].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[212].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[212].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[212].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[212].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[212].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4032 + ], + "l1_pong": [ + "0xcd80", + 4032 + ], + "l2": [ + [ + 0, + 0, + 0, + 144000, + 144000 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_260_spill" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_260_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_260_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_260_for_layer_261_port0.in[1]" + } + ], + "layer_name": "Conv_403", + "layer_object_name": "compute_graph.flexml_layers[212]", + "layer_order": 261, + "mlir_dag_id_map": { + "dag_layer": 261, + "mlir_layer": 261 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[212].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[212].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[212].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[212].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[212].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[212].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[212].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[212].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[212].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[212].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[212].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[212].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[212].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[212].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[212].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[212].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 3072 + ], + "l1_pong": [ + "0x4000", + 3072 + ], + "l2": [ + [ + 2, + 0, + 32768, + 245760, + 144000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_261" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[212].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[212].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[212].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[212].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[212].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[212].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[212].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[212].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[212].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[212].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[212].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[212].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[212].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[212].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[212].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[212].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xed00", + 2368 + ], + "l1_pong": [ + "0x9f80", + 2368 + ], + "l2": [ + [ + 3, + 0, + 467456, + 28416 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_261_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 9033728, + "size": 28416 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_261_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0262": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_261" + ], + "out": [ + "compute_graph.l2_262", + "compute_graph.l2l3_262_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[213].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[213].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[213].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[213].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[213].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[213].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[213].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[213].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[213].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[213].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[213].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[213].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[213].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[213].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[213].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[213].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 32768, + 245760, + 144000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_261" + ] + }, + "kernel_name": [ + "superkernel_sigmoid1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_261", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_262.out[0]", + "to": "compute_graph.l2l3_262_spill.in[0]" + }, + { + "from": "compute_graph.l2_262.out[1]", + "to": "compute_graph.l2l3_262_spill.in[1]" + } + ], + "layer_name": "Sigmoid_404", + "layer_object_name": "compute_graph.flexml_layers[213]", + "layer_order": 262, + "mlir_dag_id_map": { + "dag_layer": 262, + "mlir_layer": 262 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[213].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[213].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[213].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[213].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[213].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[213].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[213].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[213].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[213].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[213].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[213].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[213].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[213].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[213].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[213].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[213].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 245760, + 144000 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_262" + ], + "l3": { + "l2l3_262_spill": { + "size": 144000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_262_spill" + ] + }, + "super_iter": 1 + }, + "0263": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_262_spill" + ], + "out": [ + "compute_graph.l2l3_263_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_262_spill": { + "size": 144000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_262_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_405_Duplicated#1", + "layer_object_name": "compute_graph.templated_graph_263", + "layer_order": 263, + "mlir_dag_id_map": { + "dag_layer": 263, + "mlir_layer": 263 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_263_spill": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_263_spill" + ] + }, + "templated_graph": 1 + }, + "0264": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0", + "compute_graph.const_ifm_ddr_1", + "compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1", + "compute_graph.l2l3_263_spill" + ], + "out": [ + "compute_graph.l2_264" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[214].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[214].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[214].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[214].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[214].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[214].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[214].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[214].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[214].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[214].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[214].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[214].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[214].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[214].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[214].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[214].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 86400, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0" + ], + "l3": { + "const_ifm_ddr_1": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.const_ifm_ddr_1" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[214].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[214].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[214].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[214].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[214].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[214].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[214].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[214].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[214].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[214].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[214].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[214].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[214].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[214].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[214].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[214].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 172800, + 86400, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_263_spill" + ] + }, + "kernel_name": [ + "superkernel_sub1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.const_ifm_ddr_1.out[0]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.in[0]" + }, + { + "from": "compute_graph.const_ifm_ddr_1.out[1]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input1_0_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_263_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_263_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_264_port1.in[1]" + } + ], + "layer_name": "Sub_411", + "layer_object_name": "compute_graph.flexml_layers[214]", + "layer_order": 264, + "mlir_dag_id_map": { + "dag_layer": 264, + "mlir_layer": 264 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[214].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[214].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[214].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[214].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[214].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[214].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[214].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[214].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[214].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[214].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[214].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[214].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[214].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[214].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[214].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[214].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_264" + ] + }, + "super_iter": 1 + }, + "0265": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_264", + "compute_graph.L2_IFM_Buffer_for_input2_1_port1", + "compute_graph.ifm_ddr_2" + ], + "out": [ + "compute_graph.l2_265", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[215].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[215].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[215].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[215].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[215].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[215].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[215].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[215].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[215].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[215].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[215].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[215].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[215].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[215].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[215].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[215].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_264" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[215].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[215].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[215].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[215].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[215].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[215].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[215].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[215].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[215].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[215].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[215].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[215].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[215].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[215].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[215].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[215].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 86400, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_for_input2_1_port1" + ], + "l3": { + "ifm_ddr_2": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.ifm_ddr_2" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_for_input2_1_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_264", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_265.out[0]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.in[0]" + }, + { + "from": "compute_graph.l2_265.out[1]", + "to": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_2.out[1]", + "to": "compute_graph.L2_IFM_Buffer_for_input2_1_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_2.out[2]", + "to": "compute_graph.L2_IFM_Buffer_for_input2_1_port1.in[1]" + } + ], + "layer_name": "Mul_412", + "layer_object_name": "compute_graph.flexml_layers[215]", + "layer_order": 265, + "mlir_dag_id_map": { + "dag_layer": 265, + "mlir_layer": 265 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[215].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[215].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[215].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[215].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[215].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[215].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[215].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[215].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[215].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[215].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[215].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[215].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[215].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[215].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[215].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[215].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 172800, + 122880, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_265" + ], + "l3": { + "L3_OFM_Buffer_layer_TGSpilling_265265": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265" + ] + }, + "super_iter": 1 + }, + "0266": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_262_spill" + ], + "out": [ + "compute_graph.l2l3_266_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_262_spill": { + "size": 144000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_262_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_405_Duplicated#0", + "layer_object_name": "compute_graph.templated_graph_266", + "layer_order": 266, + "mlir_dag_id_map": { + "dag_layer": 266, + "mlir_layer": 266 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_266_spill": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_266_spill" + ] + }, + "templated_graph": 1 + }, + "0267": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_for_input2_2_port1", + "compute_graph.ifm_ddr_2", + "compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0", + "compute_graph.l2l3_266_spill" + ], + "out": [ + "compute_graph.l2_267", + "compute_graph.l2l3_267_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[216].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[216].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[216].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[216].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[216].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[216].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[216].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[216].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[216].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[216].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[216].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[216].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[216].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[216].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[216].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[216].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 172800, + 86400, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_266_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[216].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[216].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[216].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[216].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[216].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[216].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[216].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[216].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[216].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[216].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[216].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[216].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[216].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[216].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[216].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[216].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 86400, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_for_input2_2_port1" + ], + "l3": { + "ifm_ddr_2": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.ifm_ddr_2" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_for_input2_2_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_267.out[0]", + "to": "compute_graph.l2l3_267_spill.in[0]" + }, + { + "from": "compute_graph.l2_267.out[1]", + "to": "compute_graph.l2l3_267_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_2.out[3]", + "to": "compute_graph.L2_IFM_Buffer_for_input2_2_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_2.out[4]", + "to": "compute_graph.L2_IFM_Buffer_for_input2_2_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_266_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_266_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_266_for_layer_267_port0.in[1]" + } + ], + "layer_name": "Mul_406", + "layer_object_name": "compute_graph.flexml_layers[216]", + "layer_order": 267, + "mlir_dag_id_map": { + "dag_layer": 267, + "mlir_layer": 267 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[216].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[216].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[216].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[216].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[216].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[216].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[216].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[216].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[216].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[216].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[216].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[216].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[216].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[216].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[216].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[216].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 345600, + 122880, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_267" + ], + "l3": { + "l2l3_267_spill": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_267_spill" + ] + }, + "super_iter": 1 + }, + "0268": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_259_spill", + "compute_graph.l2l3_267_spill" + ], + "out": [ + "compute_graph.l2l3_268_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_259_spill": { + "size": 86400 + }, + "l2l3_267_spill": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_259_spill", + "compute_graph.l2l3_267_spill" + ] + }, + "kernel_name": [ + "ConcatC8Adf" + ], + "layer_name": "Concat_407", + "layer_object_name": "compute_graph.templated_graph_268", + "layer_order": 268, + "mlir_dag_id_map": { + "dag_layer": 268, + "mlir_layer": 268 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_268_spill": { + "size": 144000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_268_spill" + ] + }, + "templated_graph": 1 + }, + "0269": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0", + "compute_graph.l2l3_268_spill" + ], + "out": [ + "compute_graph.l2_269" + ], + "wts": [ + "compute_graph.Layer_269_l2_wts", + "compute_graph.Layer_269_wts_ddr" + ] + }, + "buffer_iter": 1, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[217].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[217].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[217].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[217].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[217].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[217].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[217].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[217].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[217].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[217].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[217].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[217].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[217].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[217].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[217].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[217].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3808 + ], + "l1_pong": [ + "0xcd80", + 3808 + ], + "l2": [ + [ + 0, + 0, + 0, + 144000, + 144000 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_268_spill" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_268_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_268_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_268_for_layer_269_port0.in[1]" + } + ], + "layer_name": "Conv_408", + "layer_object_name": "compute_graph.flexml_layers[217]", + "layer_order": 269, + "mlir_dag_id_map": { + "dag_layer": 269, + "mlir_layer": 269 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[217].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[217].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[217].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[217].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[217].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[217].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[217].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[217].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[217].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[217].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[217].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[217].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[217].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[217].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[217].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[217].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 3072 + ], + "l1_pong": [ + "0x4000", + 3072 + ], + "l2": [ + [ + 2, + 0, + 229376, + 147456, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_269" + ] + }, + "super_iter": 1, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[217].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[217].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[217].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[217].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[217].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[217].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[217].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[217].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[217].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[217].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[217].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[217].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[217].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[217].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[217].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[217].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xeb40", + 608 + ], + "l1_pong": [ + "0x9dc0", + 608 + ], + "l2": [ + [ + 3, + 0, + 0, + 12160 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_269_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 9090560, + "size": 12160 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_269_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0270": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_269" + ], + "out": [ + "compute_graph.l2_270" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[218].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[218].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[218].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[218].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[218].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[218].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[218].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[218].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[218].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[218].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[218].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[218].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[218].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[218].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[218].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[218].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 7680 + ], + "l1_pong": [ + "0x4000", + 7680 + ], + "l2": [ + [ + 2, + 0, + 229376, + 147456, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_269" + ] + }, + "kernel_name": [ + "superkernel_tanh1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.l2_269", + "dest_port": 0, + "src_offset": 0 + } + ], + "layer_name": "Tanh_409", + "layer_object_name": "compute_graph.flexml_layers[218]", + "layer_order": 270, + "mlir_dag_id_map": { + "dag_layer": 270, + "mlir_layer": 270 + }, + "num_iter": 4, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[218].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[218].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[218].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[218].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[218].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[218].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[218].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[218].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[218].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[218].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[218].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[218].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[218].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[218].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[218].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[218].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa380", + 1920 + ], + "l1_pong": [ + "0xcd80", + 1920 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_270" + ] + }, + "super_iter": 1 + }, + "0271": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_270", + "compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0", + "compute_graph.l2l3_263_spill" + ], + "out": [ + "compute_graph.l2_271" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[219].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[219].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[219].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[219].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[219].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[219].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[219].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[219].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[219].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[219].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[219].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[219].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[219].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[219].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[219].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[219].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 245760, + 86400, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_263_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[219].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[219].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[219].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[219].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[219].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[219].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[219].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[219].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[219].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[219].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[219].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[219].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[219].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[219].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[219].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[219].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_270" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_270", + "dest_port": 2, + "src_offset": 0 + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_263_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_263_spill.out[3]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_263_for_layer_271_port0.in[1]" + } + ], + "layer_name": "Mul_413", + "layer_object_name": "compute_graph.flexml_layers[219]", + "layer_order": 271, + "mlir_dag_id_map": { + "dag_layer": 271, + "mlir_layer": 271 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[219].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[219].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[219].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[219].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[219].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[219].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[219].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[219].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[219].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[219].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[219].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[219].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[219].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[219].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[219].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[219].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_271" + ] + }, + "super_iter": 1 + }, + "0272": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2_271", + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265", + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265" + ], + "out": [ + "compute_graph.l2_272", + "compute_graph.ofm_ddr_2_l2l3_272_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[220].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[220].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[220].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[220].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[220].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[220].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[220].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[220].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[220].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[220].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[220].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[220].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[220].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[220].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[220].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[220].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 3840 + ], + "l1_pong": [ + "0x4000", + 3840 + ], + "l2": [ + [ + 0, + 0, + 0, + 122880, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265" + ], + "l3_buffer_names": [ + "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[220].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[220].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[220].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[220].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[220].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[220].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[220].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[220].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[220].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[220].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[220].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[220].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[220].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[220].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[220].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[220].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5e00", + 3840 + ], + "l1_pong": [ + "0x1e00", + 3840 + ], + "l2": [ + [ + 2, + 0, + 278528, + 122880, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_271" + ] + }, + "kernel_name": [ + "superkernel_add1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.l2_271", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_272.out[0]", + "to": "compute_graph.ofm_ddr_2_l2l3_272_spill.in[0]" + }, + { + "from": "compute_graph.l2_272.out[1]", + "to": "compute_graph.ofm_ddr_2_l2l3_272_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.out[0]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.in[0]" + }, + { + "from": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265.out[1]", + "to": "compute_graph.L2_OFM_Buffer_layer_TGSpilling_265265.in[1]" + } + ], + "layer_name": "Add_414", + "layer_object_name": "compute_graph.flexml_layers[220]", + "layer_order": 272, + "mlir_dag_id_map": { + "dag_layer": 272, + "mlir_layer": 272 + }, + "num_iter": 8, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[220].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[220].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[220].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[220].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[220].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[220].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[220].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[220].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[220].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[220].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[220].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[220].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[220].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[220].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[220].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[220].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xab00", + 960 + ], + "l1_pong": [ + "0xcd80", + 960 + ], + "l2": [ + [ + 0, + 0, + 245760, + 122880, + 86400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_272" + ], + "l3": { + "ofm_ddr_2_l2l3_272_spill": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.ofm_ddr_2_l2l3_272_spill" + ] + }, + "super_iter": 1 + }, + "0273": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_258_spill", + "compute_graph.ofm_ddr_2_l2l3_272_spill" + ], + "out": [ + "compute_graph.l2l3_273_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_258_spill": { + "size": 86400 + }, + "ofm_ddr_2_l2l3_272_spill": { + "size": 86400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_258_spill", + "compute_graph.ofm_ddr_2_l2l3_272_spill" + ] + }, + "kernel_name": [ + "ConcatC8Adf" + ], + "layer_name": "Concat_415", + "layer_object_name": "compute_graph.templated_graph_273", + "layer_order": 273, + "mlir_dag_id_map": { + "dag_layer": 273, + "mlir_layer": 273 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_273_spill": { + "size": 144000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_273_spill" + ] + }, + "templated_graph": 1 + }, + "0274": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_273_spill" + ], + "out": [ + "compute_graph.spill_L3_Concat_Buffer_layer_275" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_273_spill": { + "size": 144000 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_273_spill" + ] + }, + "kernel_name": [ + "ResizeAdf" + ], + "layer_name": "Resize_417", + "layer_object_name": "compute_graph.templated_graph_274", + "layer_order": 274, + "mlir_dag_id_map": { + "dag_layer": 274, + "mlir_layer": 274 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "spill_L3_Concat_Buffer_layer_275": { + "size": 921600 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_275" + ] + }, + "templated_graph": 1 + }, + "0276": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0", + "compute_graph.spill_L3_Concat_Buffer_layer_275" + ], + "out": [ + "compute_graph.l2_276", + "compute_graph.l2l3_276_spill" + ], + "wts": [ + "compute_graph.Layer_276_l2_wts", + "compute_graph.Layer_276_wts_ddr" + ] + }, + "buffer_iter": 5, + "depth_iter": 4, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[221].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[221].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[221].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[221].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[221].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[221].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[221].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[221].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[221].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[221].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[221].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[221].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[221].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[221].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[221].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[221].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3808 + ], + "l1_pong": [ + "0xcd80", + 3808 + ], + "l2": [ + [ + 0, + 0, + 0, + 204800, + 204800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0" + ], + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_275" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_276.out[0]", + "to": "compute_graph.l2l3_276_spill.in[0]" + }, + { + "from": "compute_graph.l2_276.out[1]", + "to": "compute_graph.l2l3_276_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_275.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[0]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_275.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_275_for_layer_276_port0.in[1]" + } + ], + "layer_name": "Conv_419", + "layer_object_name": "compute_graph.flexml_layers[221]", + "layer_order": 276, + "mlir_dag_id_map": { + "dag_layer": 276, + "mlir_layer": 276 + }, + "num_iter": 100, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[221].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[221].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[221].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[221].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[221].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[221].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[221].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[221].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[221].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[221].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[221].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[221].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[221].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[221].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[221].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[221].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1280 + ], + "l1_pong": [ + "0x4000", + 1280 + ], + "l2": [ + [ + 2, + 0, + 114688, + 102400, + 92160 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_276" + ], + "l3": { + "l2l3_276_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_276_spill" + ] + }, + "super_iter": 5, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[221].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[221].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[221].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[221].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[221].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[221].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[221].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[221].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[221].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[221].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[221].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[221].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[221].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[221].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[221].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[221].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xeb40", + 1184 + ], + "l1_pong": [ + "0x9dc0", + 1184 + ], + "l2": [ + [ + 3, + 0, + 0, + 18944 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_276_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 9114880, + "size": 18944 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_276_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0277": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_276_spill" + ], + "out": [ + "compute_graph.l2l3_277_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_276_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_276_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_292.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_292.in[1]" + } + ], + "layer_name": "Split_421_Duplicated#0", + "layer_object_name": "compute_graph.templated_graph_277", + "layer_order": 277, + "mlir_dag_id_map": { + "dag_layer": 277, + "mlir_layer": 277 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_277_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_277_spill" + ] + }, + "templated_graph": 1 + }, + "0278": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_276_spill" + ], + "out": [ + "compute_graph.l2l3_278_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_276_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_276_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_279.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_279.in[1]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_287.in[0]" + }, + { + "from": "compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_287.in[1]" + } + ], + "layer_name": "Split_421_Duplicated#1", + "layer_object_name": "compute_graph.templated_graph_278", + "layer_order": 278, + "mlir_dag_id_map": { + "dag_layer": 278, + "mlir_layer": 278 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_278_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_278_spill" + ] + }, + "templated_graph": 1 + }, + "0280": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0", + "compute_graph.spill_L3_Concat_Buffer_layer_279" + ], + "out": [ + "compute_graph.l2_280", + "compute_graph.l2l3_280_spill" + ], + "wts": [ + "compute_graph.Layer_280_l2_wts", + "compute_graph.Layer_280_wts_ddr" + ] + }, + "buffer_iter": 3, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[222].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[222].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[222].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[222].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[222].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[222].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[222].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[222].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[222].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[222].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[222].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[222].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[222].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[222].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[222].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[222].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3264 + ], + "l1_pong": [ + "0xcd80", + 3264 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 163840 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0" + ], + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_279" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_280.out[0]", + "to": "compute_graph.l2l3_280_spill.in[0]" + }, + { + "from": "compute_graph.l2_280.out[1]", + "to": "compute_graph.l2l3_280_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_1.out[0]", + "to": "compute_graph.L2_IFM_Buffer_for_input1_0_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_1.out[1]", + "to": "compute_graph.L2_IFM_Buffer_for_input1_0_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_278_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_278_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_279_port0.in[1]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_279.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.in[0]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_279.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_279_for_layer_280_port0.in[1]" + } + ], + "layer_name": "Conv_423", + "layer_object_name": "compute_graph.flexml_layers[222]", + "layer_order": 280, + "mlir_dag_id_map": { + "dag_layer": 280, + "mlir_layer": 280 + }, + "num_iter": 60, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[222].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[222].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[222].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[222].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[222].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[222].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[222].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[222].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[222].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[222].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[222].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[222].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[222].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[222].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[222].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[222].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 1, + 0, + 393216, + 163840, + 153600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_280" + ], + "l3": { + "l2l3_280_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_280_spill" + ] + }, + "super_iter": 3, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[222].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[222].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[222].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[222].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[222].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[222].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[222].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[222].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[222].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[222].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[222].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[222].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[222].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[222].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[222].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[222].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe700", + 1184 + ], + "l1_pong": [ + "0x9980", + 1184 + ], + "l2": [ + [ + 3, + 0, + 505344, + 9472 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_280_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 9152768, + "size": 9472 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_280_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0281": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0", + "compute_graph.l2l3_280_spill" + ], + "out": [ + "compute_graph.l2_281", + "compute_graph.l2l3_281_spill" + ] + }, + "buffer_iter": 3, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[223].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[223].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[223].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[223].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[223].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[223].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[223].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[223].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[223].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[223].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[223].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[223].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[223].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[223].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[223].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[223].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 8192 + ], + "l1_pong": [ + "0x4000", + 8192 + ], + "l2": [ + [ + 1, + 0, + 434176, + 153600, + 153600 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_280_spill" + ] + }, + "kernel_name": [ + "superkernel_sigmoid1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_281.out[0]", + "to": "compute_graph.l2l3_281_spill.in[0]" + }, + { + "from": "compute_graph.l2_281.out[1]", + "to": "compute_graph.l2l3_281_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_280_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_280_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_280_for_layer_281_port0.in[1]" + } + ], + "layer_name": "Sigmoid_424", + "layer_object_name": "compute_graph.flexml_layers[223]", + "layer_order": 281, + "mlir_dag_id_map": { + "dag_layer": 281, + "mlir_layer": 281 + }, + "num_iter": 15, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[223].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[223].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[223].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[223].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[223].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[223].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[223].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[223].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[223].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[223].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[223].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[223].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[223].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[223].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[223].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[223].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa280", + 2048 + ], + "l1_pong": [ + "0xcd80", + 2048 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 153600 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_281" + ], + "l3": { + "l2l3_281_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_281_spill" + ] + }, + "super_iter": 3 + }, + "0282": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_281_spill" + ], + "out": [ + "compute_graph.l2l3_282_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_281_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_281_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_425_Duplicated#1", + "layer_object_name": "compute_graph.templated_graph_282", + "layer_order": 282, + "mlir_dag_id_map": { + "dag_layer": 282, + "mlir_layer": 282 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_282_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_282_spill" + ] + }, + "templated_graph": 1 + }, + "0283": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0", + "compute_graph.const_ifm_ddr", + "compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1", + "compute_graph.l2l3_282_spill" + ], + "out": [ + "compute_graph.l2_283", + "compute_graph.l2l3_283_spill" + ] + }, + "buffer_iter": 3, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[224].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[224].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[224].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[224].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[224].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[224].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[224].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[224].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[224].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[224].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[224].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[224].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[224].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[224].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[224].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[224].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 4096 + ], + "l1_pong": [ + "0x4000", + 4096 + ], + "l2": [ + [ + 2, + 0, + 217088, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0" + ], + "l3": { + "const_ifm_ddr": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.const_ifm_ddr" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[224].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[224].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[224].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[224].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[224].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[224].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[224].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[224].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[224].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[224].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[224].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[224].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[224].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[224].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[224].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[224].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x6000", + 4096 + ], + "l1_pong": [ + "0x2000", + 4096 + ], + "l2": [ + [ + 1, + 0, + 434176, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_282_spill" + ] + }, + "kernel_name": [ + "superkernel_sub1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_283.out[0]", + "to": "compute_graph.l2l3_283_spill.in[0]" + }, + { + "from": "compute_graph.l2_283.out[1]", + "to": "compute_graph.l2l3_283_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.const_ifm_ddr.out[0]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.in[0]" + }, + { + "from": "compute_graph.const_ifm_ddr.out[1]", + "to": "compute_graph.L2_CONST_IFM_Buffer_for_input0_0_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_282_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_282_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_283_port1.in[1]" + } + ], + "layer_name": "Sub_431", + "layer_object_name": "compute_graph.flexml_layers[224]", + "layer_order": 283, + "mlir_dag_id_map": { + "dag_layer": 283, + "mlir_layer": 283 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[224].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[224].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[224].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[224].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[224].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[224].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[224].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[224].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[224].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[224].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[224].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[224].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[224].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[224].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[224].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[224].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaa80", + 1024 + ], + "l1_pong": [ + "0xcd80", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_283" + ], + "l3": { + "l2l3_283_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_283_spill" + ] + }, + "super_iter": 3 + }, + "0284": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_for_input1_1_port1", + "compute_graph.ifm_ddr_1", + "compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0", + "compute_graph.l2l3_283_spill" + ], + "out": [ + "compute_graph.l2_284", + "compute_graph.l2l3_284_spill" + ] + }, + "buffer_iter": 3, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[225].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[225].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[225].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[225].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[225].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[225].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[225].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[225].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[225].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[225].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[225].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[225].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[225].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[225].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[225].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[225].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 4096 + ], + "l1_pong": [ + "0x4000", + 4096 + ], + "l2": [ + [ + 1, + 0, + 434176, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_283_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[225].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[225].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[225].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[225].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[225].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[225].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[225].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[225].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[225].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[225].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[225].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[225].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[225].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[225].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[225].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[225].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x6000", + 4096 + ], + "l1_pong": [ + "0x2000", + 4096 + ], + "l2": [ + [ + 2, + 0, + 217088, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_for_input1_1_port1" + ], + "l3": { + "ifm_ddr_1": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.ifm_ddr_1" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_for_input1_1_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_284.out[0]", + "to": "compute_graph.l2l3_284_spill.in[0]" + }, + { + "from": "compute_graph.l2_284.out[1]", + "to": "compute_graph.l2l3_284_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_1.out[2]", + "to": "compute_graph.L2_IFM_Buffer_for_input1_1_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_1.out[3]", + "to": "compute_graph.L2_IFM_Buffer_for_input1_1_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_283_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_283_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_283_for_layer_284_port0.in[1]" + } + ], + "layer_name": "Mul_432", + "layer_object_name": "compute_graph.flexml_layers[225]", + "layer_order": 284, + "mlir_dag_id_map": { + "dag_layer": 284, + "mlir_layer": 284 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[225].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[225].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[225].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[225].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[225].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[225].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[225].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[225].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[225].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[225].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[225].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[225].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[225].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[225].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[225].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[225].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaa80", + 1024 + ], + "l1_pong": [ + "0xcd80", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_284" + ], + "l3": { + "l2l3_284_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_284_spill" + ] + }, + "super_iter": 3 + }, + "0285": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_281_spill" + ], + "out": [ + "compute_graph.l2l3_285_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_281_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_281_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_425_Duplicated#0", + "layer_object_name": "compute_graph.templated_graph_285", + "layer_order": 285, + "mlir_dag_id_map": { + "dag_layer": 285, + "mlir_layer": 285 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_285_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_285_spill" + ] + }, + "templated_graph": 1 + }, + "0286": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_for_input1_2_port1", + "compute_graph.ifm_ddr_1", + "compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0", + "compute_graph.l2l3_285_spill" + ], + "out": [ + "compute_graph.l2_286", + "compute_graph.spill_L3_Concat_Buffer_layer_287" + ] + }, + "buffer_iter": 3, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[226].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[226].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[226].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[226].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[226].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[226].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[226].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[226].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[226].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[226].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[226].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[226].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[226].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[226].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[226].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[226].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 4096 + ], + "l1_pong": [ + "0x4000", + 4096 + ], + "l2": [ + [ + 1, + 0, + 434176, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_285_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[226].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[226].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[226].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[226].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[226].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[226].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[226].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[226].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[226].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[226].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[226].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[226].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[226].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[226].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[226].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[226].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x6000", + 4096 + ], + "l1_pong": [ + "0x2000", + 4096 + ], + "l2": [ + [ + 2, + 0, + 217088, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_for_input1_2_port1" + ], + "l3": { + "ifm_ddr_1": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.ifm_ddr_1" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_for_input1_2_port1", + "dest_port": 2, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_286.out[0]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_287.in[2]" + }, + { + "from": "compute_graph.l2_286.out[1]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_287.in[3]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.ifm_ddr_1.out[4]", + "to": "compute_graph.L2_IFM_Buffer_for_input1_2_port1.in[0]" + }, + { + "from": "compute_graph.ifm_ddr_1.out[5]", + "to": "compute_graph.L2_IFM_Buffer_for_input1_2_port1.in[1]" + }, + { + "from": "compute_graph.l2l3_285_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_285_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_285_for_layer_286_port0.in[1]" + } + ], + "layer_name": "Mul_426", + "layer_object_name": "compute_graph.flexml_layers[226]", + "layer_order": 286, + "mlir_dag_id_map": { + "dag_layer": 286, + "mlir_layer": 286 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[226].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[226].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[226].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[226].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[226].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[226].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[226].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[226].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[226].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[226].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[226].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[226].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[226].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[226].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[226].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[226].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaa80", + 1024 + ], + "l1_pong": [ + "0xcd80", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_286" + ], + "l3": { + "spill_L3_Concat_Buffer_layer_287": { + "buffer_bounds": [ + 1, + 16, + 30, + 160 + ], + "concat_offset": [ + 0, + 16, + 0, + 0 + ], + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_287" + ] + }, + "super_iter": 3 + }, + "0288": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0", + "compute_graph.spill_L3_Concat_Buffer_layer_287" + ], + "out": [ + "compute_graph.l2_288", + "compute_graph.l2l3_288_spill" + ], + "wts": [ + "compute_graph.Layer_288_l2_wts", + "compute_graph.Layer_288_wts_ddr" + ] + }, + "buffer_iter": 3, + "depth_iter": 2, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[227].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[227].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[227].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[227].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[227].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[227].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[227].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[227].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[227].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[227].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[227].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[227].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[227].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[227].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[227].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[227].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3264 + ], + "l1_pong": [ + "0xcd80", + 3264 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 163840 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0" + ], + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_287" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_288.out[0]", + "to": "compute_graph.l2l3_288_spill.in[0]" + }, + { + "from": "compute_graph.l2_288.out[1]", + "to": "compute_graph.l2l3_288_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_278_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_278_spill.out[3]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_278_for_layer_287_port0.in[1]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_287.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.in[0]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_287.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_287_for_layer_288_port0.in[1]" + } + ], + "layer_name": "Conv_428", + "layer_object_name": "compute_graph.flexml_layers[227]", + "layer_order": 288, + "mlir_dag_id_map": { + "dag_layer": 288, + "mlir_layer": 288 + }, + "num_iter": 60, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[227].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[227].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[227].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[227].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[227].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[227].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[227].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[227].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[227].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[227].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[227].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[227].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[227].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[227].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[227].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[227].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1024 + ], + "l1_pong": [ + "0x4000", + 1024 + ], + "l2": [ + [ + 1, + 0, + 393216, + 163840, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_288" + ], + "l3": { + "l2l3_288_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_288_spill" + ] + }, + "super_iter": 3, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[227].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[227].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[227].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[227].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[227].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[227].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[227].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[227].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[227].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[227].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[227].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[227].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[227].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[227].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[227].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[227].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe700", + 1184 + ], + "l1_pong": [ + "0x9980", + 1184 + ], + "l2": [ + [ + 3, + 0, + 0, + 9472 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_288_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 9171712, + "size": 9472 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_288_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0289": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0", + "compute_graph.l2l3_288_spill" + ], + "out": [ + "compute_graph.l2_289", + "compute_graph.l2l3_289_spill" + ] + }, + "buffer_iter": 1, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[228].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[228].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[228].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[228].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[228].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[228].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[228].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[228].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[228].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[228].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[228].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[228].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[228].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[228].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[228].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[228].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 5888 + ], + "l1_pong": [ + "0x4000", + 5888 + ], + "l2": [ + [ + 0, + 0, + 0, + 230400, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_288_spill" + ] + }, + "kernel_name": [ + "superkernel_tanh1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_289.out[0]", + "to": "compute_graph.l2l3_289_spill.in[0]" + }, + { + "from": "compute_graph.l2_289.out[1]", + "to": "compute_graph.l2l3_289_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_288_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_288_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_288_for_layer_289_port0.in[1]" + } + ], + "layer_name": "Tanh_429", + "layer_object_name": "compute_graph.flexml_layers[228]", + "layer_order": 289, + "mlir_dag_id_map": { + "dag_layer": 289, + "mlir_layer": 289 + }, + "num_iter": 20, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[228].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[228].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[228].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[228].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[228].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[228].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[228].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[228].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[228].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[228].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[228].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[228].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[228].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[228].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[228].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[228].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa700", + 1472 + ], + "l1_pong": [ + "0xcd80", + 1472 + ], + "l2": [ + [ + 0, + 0, + 460800, + 471040, + 230400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_289" + ], + "l3": { + "l2l3_289_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_289_spill" + ] + }, + "super_iter": 1 + }, + "0290": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0", + "compute_graph.l2l3_282_spill", + "compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1", + "compute_graph.l2l3_289_spill" + ], + "out": [ + "compute_graph.l2_290", + "compute_graph.l2l3_290_spill" + ] + }, + "buffer_iter": 3, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[229].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[229].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[229].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[229].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[229].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[229].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[229].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[229].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[229].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[229].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[229].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[229].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[229].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[229].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[229].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[229].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 4096 + ], + "l1_pong": [ + "0x4000", + 4096 + ], + "l2": [ + [ + 2, + 0, + 217088, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_282_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[229].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[229].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[229].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[229].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[229].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[229].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[229].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[229].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[229].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[229].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[229].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[229].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[229].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[229].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[229].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[229].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x6000", + 4096 + ], + "l1_pong": [ + "0x2000", + 4096 + ], + "l2": [ + [ + 1, + 0, + 434176, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_289_spill" + ] + }, + "kernel_name": [ + "superkernel_mul1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_290.out[0]", + "to": "compute_graph.l2l3_290_spill.in[0]" + }, + { + "from": "compute_graph.l2_290.out[1]", + "to": "compute_graph.l2l3_290_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_282_spill.out[2]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_282_spill.out[3]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_282_for_layer_290_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_289_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_289_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_289_for_layer_290_port1.in[1]" + } + ], + "layer_name": "Mul_433", + "layer_object_name": "compute_graph.flexml_layers[229]", + "layer_order": 290, + "mlir_dag_id_map": { + "dag_layer": 290, + "mlir_layer": 290 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[229].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[229].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[229].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[229].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[229].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[229].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[229].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[229].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[229].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[229].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[229].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[229].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[229].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[229].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[229].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[229].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaa80", + 1024 + ], + "l1_pong": [ + "0xcd80", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_290" + ], + "l3": { + "l2l3_290_spill": { + "size": 230400 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_290_spill" + ] + }, + "super_iter": 3 + }, + "0291": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0", + "compute_graph.l2l3_284_spill", + "compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1", + "compute_graph.l2l3_290_spill" + ], + "out": [ + "compute_graph.l2_291", + "compute_graph.ofm_ddr_3_l2l3_291_spill", + "compute_graph.spill_L3_Concat_Buffer_layer_292" + ] + }, + "buffer_iter": 3, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[230].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[230].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[230].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[230].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[230].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[230].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[230].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[230].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[230].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[230].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[230].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[230].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[230].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[230].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[230].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[230].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 4096 + ], + "l1_pong": [ + "0x4000", + 4096 + ], + "l2": [ + [ + 2, + 0, + 217088, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_284_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[230].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[230].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[230].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[230].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[230].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[230].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[230].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[230].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[230].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[230].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[230].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[230].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[230].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[230].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[230].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[230].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x6000", + 4096 + ], + "l1_pong": [ + "0x2000", + 4096 + ], + "l2": [ + [ + 1, + 0, + 434176, + 76800, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_290_spill" + ] + }, + "kernel_name": [ + "superkernel_add1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_291.out[0]", + "to": "compute_graph.ofm_ddr_3_l2l3_291_spill.in[0]" + }, + { + "from": "compute_graph.l2_291.out[1]", + "to": "compute_graph.ofm_ddr_3_l2l3_291_spill.in[1]" + }, + { + "from": "compute_graph.l2_291.out[2]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_292.in[2]" + }, + { + "from": "compute_graph.l2_291.out[3]", + "to": "compute_graph.spill_L3_Concat_Buffer_layer_292.in[3]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_284_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_284_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_284_for_layer_291_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_290_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_290_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_290_for_layer_291_port1.in[1]" + } + ], + "layer_name": "Add_434", + "layer_object_name": "compute_graph.flexml_layers[230]", + "layer_order": 291, + "mlir_dag_id_map": { + "dag_layer": 291, + "mlir_layer": 291 + }, + "num_iter": 30, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[230].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[230].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[230].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[230].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[230].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[230].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[230].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[230].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[230].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[230].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[230].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[230].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[230].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[230].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[230].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[230].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xaa80", + 1024 + ], + "l1_pong": [ + "0xcd80", + 1024 + ], + "l2": [ + [ + 0, + 0, + 0, + 163840, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_291" + ], + "l3": { + "ofm_ddr_3_l2l3_291_spill": { + "size": 230400 + }, + "spill_L3_Concat_Buffer_layer_292": { + "buffer_bounds": [ + 1, + 16, + 30, + 160 + ], + "concat_offset": [ + 0, + 16, + 0, + 0 + ], + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.ofm_ddr_3_l2l3_291_spill", + "compute_graph.spill_L3_Concat_Buffer_layer_292" + ] + }, + "super_iter": 3 + }, + "0293": { + "adf_layer_objects": { + "in": [ + "compute_graph.spill_L3_Concat_Buffer_layer_292" + ], + "out": [ + "compute_graph.spill_L3_Concat_Buffer_layer_294" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "spill_L3_Concat_Buffer_layer_292": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_292" + ] + }, + "kernel_name": [ + "ResizeAdf" + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_277_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_277_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_277_for_layer_292_port0.in[1]" + } + ], + "layer_name": "Resize_437", + "layer_object_name": "compute_graph.templated_graph_293", + "layer_order": 293, + "mlir_dag_id_map": { + "dag_layer": 293, + "mlir_layer": 293 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "spill_L3_Concat_Buffer_layer_294": { + "size": 2304000 + } + }, + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_294" + ] + }, + "templated_graph": 1 + }, + "0295": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0", + "compute_graph.spill_L3_Concat_Buffer_layer_294" + ], + "out": [ + "compute_graph.l2_295", + "compute_graph.l2l3_295_spill" + ], + "wts": [ + "compute_graph.Layer_295_l2_wts", + "compute_graph.Layer_295_wts_ddr" + ] + }, + "buffer_iter": 12, + "depth_iter": 5, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[231].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[231].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[231].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[231].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[231].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[231].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[231].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[231].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[231].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[231].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[231].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[231].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[231].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[231].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[231].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[231].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3168 + ], + "l1_pong": [ + "0xcd80", + 3168 + ], + "l2": [ + [ + 0, + 0, + 0, + 217600, + 217600 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0" + ], + "l3_buffer_names": [ + "compute_graph.spill_L3_Concat_Buffer_layer_294" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_295.out[0]", + "to": "compute_graph.l2l3_295_spill.in[0]" + }, + { + "from": "compute_graph.l2_295.out[1]", + "to": "compute_graph.l2l3_295_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_5_spill.out[4]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_5_spill.out[5]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_294_port1.in[1]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_294.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[0]" + }, + { + "from": "compute_graph.spill_L3_Concat_Buffer_layer_294.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_294_for_layer_295_port0.in[1]" + } + ], + "layer_name": "Conv_439", + "layer_object_name": "compute_graph.flexml_layers[231]", + "layer_order": 295, + "mlir_dag_id_map": { + "dag_layer": 295, + "mlir_layer": 295 + }, + "num_iter": 300, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[231].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[231].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[231].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[231].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[231].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[231].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[231].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[231].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[231].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[231].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[231].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[231].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[231].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[231].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[231].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[231].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 2048 + ], + "l1_pong": [ + "0x4000", + 2048 + ], + "l2": [ + [ + 1, + 0, + 393216, + 163840, + 76800 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_295" + ], + "l3": { + "l2l3_295_spill": { + "size": 921600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_295_spill" + ] + }, + "super_iter": 12, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[231].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[231].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[231].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[231].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[231].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[231].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[231].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[231].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[231].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[231].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[231].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[231].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[231].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[231].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[231].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[231].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xe640", + 608 + ], + "l1_pong": [ + "0x98c0", + 608 + ], + "l2": [ + [ + 3, + 0, + 0, + 12160 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_295_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 9190656, + "size": 12160 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_295_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0296": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0", + "compute_graph.l2l3_295_spill" + ], + "out": [ + "compute_graph.l2_296", + "compute_graph.l2l3_296_spill" + ], + "wts": [ + "compute_graph.Layer_296_l2_wts", + "compute_graph.Layer_296_wts_ddr" + ] + }, + "buffer_iter": 9, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[232].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[232].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[232].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[232].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[232].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[232].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[232].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[232].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[232].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[232].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[232].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[232].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[232].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[232].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[232].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[232].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 3808 + ], + "l1_pong": [ + "0xcd80", + 3808 + ], + "l2": [ + [ + 2, + 0, + 73728, + 112640, + 112640 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_295_spill" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_296.out[0]", + "to": "compute_graph.l2l3_296_spill.in[0]" + }, + { + "from": "compute_graph.l2_296.out[1]", + "to": "compute_graph.l2l3_296_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_295_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_295_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_295_for_layer_296_port0.in[1]" + } + ], + "layer_name": "Conv_441", + "layer_object_name": "compute_graph.flexml_layers[232]", + "layer_order": 296, + "mlir_dag_id_map": { + "dag_layer": 296, + "mlir_layer": 296 + }, + "num_iter": 90, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[232].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[232].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[232].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[232].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[232].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[232].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[232].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[232].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[232].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[232].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[232].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[232].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[232].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[232].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[232].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[232].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 1280 + ], + "l1_pong": [ + "0x4000", + 1280 + ], + "l2": [ + [ + 0, + 0, + 0, + 204800, + 102400 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_296" + ], + "l3": { + "l2l3_296_spill": { + "size": 921600 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_296_spill" + ] + }, + "super_iter": 9, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[232].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[232].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[232].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[232].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[232].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[232].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[232].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[232].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[232].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[232].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[232].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[232].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[232].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[232].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[232].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[232].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xeb40", + 1184 + ], + "l1_pong": [ + "0x9dc0", + 1184 + ], + "l2": [ + [ + 3, + 0, + 514816, + 4736 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_296_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 9214976, + "size": 4736 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_296_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0297": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0", + "compute_graph.l2l3_296_spill" + ], + "out": [ + "compute_graph.l2_297", + "compute_graph.l2l3_297_spill" + ], + "wts": [ + "compute_graph.Layer_297_l2_wts", + "compute_graph.Layer_297_wts_ddr" + ] + }, + "buffer_iter": 9, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[233].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[233].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[233].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[233].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[233].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[233].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[233].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[233].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[233].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[233].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[233].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[233].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[233].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[233].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[233].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[233].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x8000", + 4096 + ], + "l1_pong": [ + "0xcd80", + 4096 + ], + "l2": [ + [ + 2, + 0, + 114688, + 102400, + 102400 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_296_spill" + ] + }, + "kernel_name": [ + "conv2d_maxpool" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_297.out[0]", + "to": "compute_graph.l2l3_297_spill.in[0]" + }, + { + "from": "compute_graph.l2_297.out[1]", + "to": "compute_graph.l2l3_297_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_296_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_296_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_296_for_layer_297_port0.in[1]" + } + ], + "layer_name": "Conv_443", + "layer_object_name": "compute_graph.flexml_layers[233]", + "layer_order": 297, + "mlir_dag_id_map": { + "dag_layer": 297, + "mlir_layer": 297 + }, + "num_iter": 225, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[233].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[233].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[233].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[233].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[233].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[233].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[233].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[233].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[233].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[233].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[233].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[233].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[233].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[233].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[233].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[233].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0x0", + 512 + ], + "l1_pong": [ + "0x4000", + 512 + ], + "l2": [ + [ + 0, + 0, + 0, + 204800, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_297" + ], + "l3": { + "l2l3_297_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_297_spill" + ] + }, + "super_iter": 9, + "wts": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[233].compute_node[0][0].in[1]", + "compute_graph.flexml_layers[233].compute_node[0][1].in[1]", + "compute_graph.flexml_layers[233].compute_node[0][2].in[1]", + "compute_graph.flexml_layers[233].compute_node[0][3].in[1]", + "compute_graph.flexml_layers[233].compute_node[1][0].in[1]", + "compute_graph.flexml_layers[233].compute_node[1][1].in[1]", + "compute_graph.flexml_layers[233].compute_node[1][2].in[1]", + "compute_graph.flexml_layers[233].compute_node[1][3].in[1]", + "compute_graph.flexml_layers[233].compute_node[2][0].in[1]", + "compute_graph.flexml_layers[233].compute_node[2][1].in[1]", + "compute_graph.flexml_layers[233].compute_node[2][2].in[1]", + "compute_graph.flexml_layers[233].compute_node[2][3].in[1]", + "compute_graph.flexml_layers[233].compute_node[3][0].in[1]", + "compute_graph.flexml_layers[233].compute_node[3][1].in[1]", + "compute_graph.flexml_layers[233].compute_node[3][2].in[1]", + "compute_graph.flexml_layers[233].compute_node[3][3].in[1]" + ], + "l1_ping": [ + "0xed80", + 544 + ], + "l1_pong": [ + "0xa000", + 544 + ], + "l2": [ + [ + 3, + 0, + 0, + 2176 + ] + ], + "l2_buffer_names": [ + "compute_graph.Layer_297_l2_wts" + ], + "l3": { + "wts_ddr": { + "offset": 9224448, + "size": 2176 + } + }, + "l3_buffer_names": [ + "compute_graph.Layer_297_wts_ddr" + ] + }, + "wts_iter": 1 + }, + "0298": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_297_spill" + ], + "out": [ + "compute_graph.l2l3_298_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_297_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_297_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_444_Duplicated#1", + "layer_object_name": "compute_graph.templated_graph_298", + "layer_order": 298, + "mlir_dag_id_map": { + "dag_layer": 298, + "mlir_layer": 298 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_298_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_298_spill" + ] + }, + "templated_graph": 1 + }, + "0299": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0", + "compute_graph.l2l3_298_spill" + ], + "out": [ + "compute_graph.L2_OFM_Buffer_layer_299", + "compute_graph.ofm_ddr_4" + ] + }, + "buffer_iter": 9, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[234].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[234].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[234].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[234].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[234].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[234].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[234].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[234].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[234].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[234].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[234].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[234].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[234].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[234].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[234].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[234].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 2, + 0, + 319488, + 51200, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_298_spill" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_OFM_Buffer_layer_299.out[0]", + "to": "compute_graph.ofm_ddr_4.in[0]" + }, + { + "from": "compute_graph.L2_OFM_Buffer_layer_299.out[1]", + "to": "compute_graph.ofm_ddr_4.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_298_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_298_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_298_for_layer_299_port0.in[1]" + } + ], + "layer_name": "Clip_447", + "layer_object_name": "compute_graph.flexml_layers[234]", + "layer_order": 299, + "mlir_dag_id_map": { + "dag_layer": 299, + "mlir_layer": 299 + }, + "num_iter": 72, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[234].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[234].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[234].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[234].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[234].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[234].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[234].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[234].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[234].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[234].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[234].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[234].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[234].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[234].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[234].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[234].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 204800, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_299" + ], + "l3": { + "ofm_ddr_4": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.ofm_ddr_4" + ] + }, + "super_iter": 9 + }, + "0300": { + "adf_layer_objects": { + "in": [ + "compute_graph.l2l3_297_spill" + ], + "out": [ + "compute_graph.l2l3_300_spill" + ] + }, + "ifm": { + "dtype": "bfloat16", + "l3": { + "l2l3_297_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_297_spill" + ] + }, + "kernel_name": [ + "SliceHCWC8Adf" + ], + "layer_name": "Split_444_Duplicated#0", + "layer_object_name": "compute_graph.templated_graph_300", + "layer_order": 300, + "mlir_dag_id_map": { + "dag_layer": 300, + "mlir_layer": 300 + }, + "num_iter": 0, + "ofm": { + "dtype": "bfloat16", + "l3": { + "l2l3_300_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_300_spill" + ] + }, + "templated_graph": 1 + }, + "0301": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1", + "compute_graph.l2l3_5_spill", + "compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0", + "compute_graph.l2l3_300_spill" + ], + "out": [ + "compute_graph.l2_301", + "compute_graph.l2l3_301_spill" + ] + }, + "buffer_iter": 9, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[235].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[235].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[235].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[235].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[235].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[235].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[235].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[235].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[235].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[235].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[235].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[235].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[235].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[235].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[235].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[235].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 2560 + ], + "l1_pong": [ + "0x4000", + 2560 + ], + "l2": [ + [ + 2, + 0, + 114688, + 51200, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_300_spill" + ] + }, + "ifm2": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[235].compute_node[0][0].in[2]", + "compute_graph.flexml_layers[235].compute_node[0][1].in[2]", + "compute_graph.flexml_layers[235].compute_node[0][2].in[2]", + "compute_graph.flexml_layers[235].compute_node[0][3].in[2]", + "compute_graph.flexml_layers[235].compute_node[1][0].in[2]", + "compute_graph.flexml_layers[235].compute_node[1][1].in[2]", + "compute_graph.flexml_layers[235].compute_node[1][2].in[2]", + "compute_graph.flexml_layers[235].compute_node[1][3].in[2]", + "compute_graph.flexml_layers[235].compute_node[2][0].in[2]", + "compute_graph.flexml_layers[235].compute_node[2][1].in[2]", + "compute_graph.flexml_layers[235].compute_node[2][2].in[2]", + "compute_graph.flexml_layers[235].compute_node[2][3].in[2]", + "compute_graph.flexml_layers[235].compute_node[3][0].in[2]", + "compute_graph.flexml_layers[235].compute_node[3][1].in[2]", + "compute_graph.flexml_layers[235].compute_node[3][2].in[2]", + "compute_graph.flexml_layers[235].compute_node[3][3].in[2]" + ], + "l1_ping": [ + "0x5400", + 2560 + ], + "l1_pong": [ + "0x1400", + 2560 + ], + "l2": [ + [ + 2, + 0, + 319488, + 51200, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_5_spill" + ] + }, + "kernel_name": [ + "superkernel_add1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0", + "dest_port": 0, + "src_offset": 0 + }, + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1", + "dest_port": 2, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.l2_301.out[0]", + "to": "compute_graph.l2l3_301_spill.in[0]" + }, + { + "from": "compute_graph.l2_301.out[1]", + "to": "compute_graph.l2l3_301_spill.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_300_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_300_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_300_for_layer_301_port0.in[1]" + }, + { + "from": "compute_graph.l2l3_5_spill.out[6]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.in[0]" + }, + { + "from": "compute_graph.l2l3_5_spill.out[7]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_5_for_layer_301_port1.in[1]" + } + ], + "layer_name": "Add_445", + "layer_object_name": "compute_graph.flexml_layers[235]", + "layer_order": 301, + "mlir_dag_id_map": { + "dag_layer": 301, + "mlir_layer": 301 + }, + "num_iter": 180, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[235].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[235].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[235].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[235].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[235].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[235].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[235].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[235].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[235].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[235].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[235].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[235].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[235].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[235].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[235].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[235].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xad80", + 640 + ], + "l1_pong": [ + "0xcd80", + 640 + ], + "l2": [ + [ + 0, + 0, + 0, + 204800, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.l2_301" + ], + "l3": { + "l2l3_301_spill": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.l2l3_301_spill" + ] + }, + "super_iter": 9 + }, + "0302": { + "adf_layer_objects": { + "in": [ + "compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0", + "compute_graph.l2l3_301_spill" + ], + "out": [ + "compute_graph.L2_OFM_Buffer_layer_302", + "compute_graph.ofm_ddr_5" + ] + }, + "buffer_iter": 9, + "depth_iter": 1, + "ifm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[236].compute_node[0][0].in[0]", + "compute_graph.flexml_layers[236].compute_node[0][1].in[0]", + "compute_graph.flexml_layers[236].compute_node[0][2].in[0]", + "compute_graph.flexml_layers[236].compute_node[0][3].in[0]", + "compute_graph.flexml_layers[236].compute_node[1][0].in[0]", + "compute_graph.flexml_layers[236].compute_node[1][1].in[0]", + "compute_graph.flexml_layers[236].compute_node[1][2].in[0]", + "compute_graph.flexml_layers[236].compute_node[1][3].in[0]", + "compute_graph.flexml_layers[236].compute_node[2][0].in[0]", + "compute_graph.flexml_layers[236].compute_node[2][1].in[0]", + "compute_graph.flexml_layers[236].compute_node[2][2].in[0]", + "compute_graph.flexml_layers[236].compute_node[2][3].in[0]", + "compute_graph.flexml_layers[236].compute_node[3][0].in[0]", + "compute_graph.flexml_layers[236].compute_node[3][1].in[0]", + "compute_graph.flexml_layers[236].compute_node[3][2].in[0]", + "compute_graph.flexml_layers[236].compute_node[3][3].in[0]" + ], + "l1_ping": [ + "0x0", + 6400 + ], + "l1_pong": [ + "0x4000", + 6400 + ], + "l2": [ + [ + 2, + 0, + 319488, + 51200, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0" + ], + "l3_buffer_names": [ + "compute_graph.l2l3_301_spill" + ] + }, + "kernel_name": [ + "superkernel_clip1d" + ], + "l2_ifm_buffer_ports": [ + { + "buff_obj": "compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0", + "dest_port": 0, + "src_offset": 0 + } + ], + "l2tol3_connections": [ + { + "from": "compute_graph.L2_OFM_Buffer_layer_302.out[0]", + "to": "compute_graph.ofm_ddr_5.in[0]" + }, + { + "from": "compute_graph.L2_OFM_Buffer_layer_302.out[1]", + "to": "compute_graph.ofm_ddr_5.in[1]" + } + ], + "l3tol2_connections": [ + { + "from": "compute_graph.l2l3_301_spill.out[0]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.in[0]" + }, + { + "from": "compute_graph.l2l3_301_spill.out[1]", + "to": "compute_graph.L2_IFM_Buffer_from_layer_301_for_layer_302_port0.in[1]" + } + ], + "layer_name": "Clip_446", + "layer_object_name": "compute_graph.flexml_layers[236]", + "layer_order": 302, + "mlir_dag_id_map": { + "dag_layer": 302, + "mlir_layer": 302 + }, + "num_iter": 72, + "ofm": { + "dtype": "bfloat16", + "l1_buffer_names": [ + "compute_graph.flexml_layers[236].compute_node[0][0].out[0]", + "compute_graph.flexml_layers[236].compute_node[0][1].out[0]", + "compute_graph.flexml_layers[236].compute_node[0][2].out[0]", + "compute_graph.flexml_layers[236].compute_node[0][3].out[0]", + "compute_graph.flexml_layers[236].compute_node[1][0].out[0]", + "compute_graph.flexml_layers[236].compute_node[1][1].out[0]", + "compute_graph.flexml_layers[236].compute_node[1][2].out[0]", + "compute_graph.flexml_layers[236].compute_node[1][3].out[0]", + "compute_graph.flexml_layers[236].compute_node[2][0].out[0]", + "compute_graph.flexml_layers[236].compute_node[2][1].out[0]", + "compute_graph.flexml_layers[236].compute_node[2][2].out[0]", + "compute_graph.flexml_layers[236].compute_node[2][3].out[0]", + "compute_graph.flexml_layers[236].compute_node[3][0].out[0]", + "compute_graph.flexml_layers[236].compute_node[3][1].out[0]", + "compute_graph.flexml_layers[236].compute_node[3][2].out[0]", + "compute_graph.flexml_layers[236].compute_node[3][3].out[0]" + ], + "l1_ping": [ + "0xa600", + 1600 + ], + "l1_pong": [ + "0xcd80", + 1600 + ], + "l2": [ + [ + 0, + 0, + 0, + 204800, + 51200 + ] + ], + "l2_buffer_names": [ + "compute_graph.L2_OFM_Buffer_layer_302" + ], + "l3": { + "ofm_ddr_5": { + "size": 460800 + } + }, + "l3_buffer_names": [ + "compute_graph.ofm_ddr_5" + ] + }, + "super_iter": 9 + } +} diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/ctrlPktPatch.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/ctrlPktPatch.json new file mode 100644 index 0000000000000000000000000000000000000000..4dba14de6caaf4d3aec1d0a92fc763ab8eae9dbd --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/ctrlPktPatch.json @@ -0,0 +1,8256 @@ +{ + "offset_unit": "byte", + "size_unit": "byte", + "control_packet_patch": { + "patch0": { + "offset": 405436, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr" + }, + "patch1": { + "offset": 405496, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr" + }, + "patch2": { + "offset": 415036, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_1_spill" + }, + "patch3": { + "offset": 415096, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_1_spill" + }, + "patch4": { + "offset": 415396, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_1_spill" + }, + "patch5": { + "offset": 415456, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_1_spill" + }, + "patch6": { + "offset": 415516, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_2_spill" + }, + "patch7": { + "offset": 415576, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_2_spill" + }, + "patch8": { + "offset": 416212, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_2_spill" + }, + "patch9": { + "offset": 416272, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_3_spill" + }, + "patch10": { + "offset": 416884, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_3_spill" + }, + "patch11": { + "offset": 416944, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_3_spill" + }, + "patch12": { + "offset": 417364, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_3_spill" + }, + "patch13": { + "offset": 417424, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_4_spill" + }, + "patch14": { + "offset": 428236, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_4_spill" + }, + "patch15": { + "offset": 428296, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_5_spill" + }, + "patch16": { + "offset": 428884, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_5_spill" + }, + "patch17": { + "offset": 428944, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_5_spill" + }, + "patch18": { + "offset": 429412, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_5_spill" + }, + "patch19": { + "offset": 429460, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr_5" + }, + "patch20": { + "offset": 429520, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_5_spill" + }, + "patch21": { + "offset": 429568, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr_5" + }, + "patch22": { + "offset": 442012, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_6_spill" + }, + "patch23": { + "offset": 442072, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_6_spill" + }, + "patch24": { + "offset": 452812, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_6_spill" + }, + "patch25": { + "offset": 452860, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr_4" + }, + "patch26": { + "offset": 452920, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_6_spill" + }, + "patch27": { + "offset": 452968, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr_4" + }, + "patch28": { + "offset": 453028, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_8_wts_ddr" + }, + "patch29": { + "offset": 453088, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_8_wts_ddr" + }, + "patch30": { + "offset": 465652, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_7_spill" + }, + "patch31": { + "offset": 465712, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_7_spill" + }, + "patch32": { + "offset": 476284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_7_spill" + }, + "patch33": { + "offset": 476344, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_7_spill" + }, + "patch34": { + "offset": 476404, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_7_spill" + }, + "patch35": { + "offset": 476464, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_7_spill" + }, + "patch36": { + "offset": 476524, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_7_spill" + }, + "patch37": { + "offset": 476584, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_7_spill" + }, + "patch38": { + "offset": 477052, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_8_spill" + }, + "patch39": { + "offset": 477112, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_8_spill" + }, + "patch40": { + "offset": 490324, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_8_spill" + }, + "patch41": { + "offset": 490384, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_8_spill" + }, + "patch42": { + "offset": 506236, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_9_spill" + }, + "patch43": { + "offset": 506296, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_9_spill" + }, + "patch44": { + "offset": 506596, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_10_spill" + }, + "patch45": { + "offset": 506656, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_10_spill" + }, + "patch46": { + "offset": 506716, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_9_spill" + }, + "patch47": { + "offset": 506776, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_9_spill" + }, + "patch48": { + "offset": 521572, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_10_spill" + }, + "patch49": { + "offset": 521632, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_10_spill" + }, + "patch50": { + "offset": 539404, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_11_spill" + }, + "patch51": { + "offset": 539464, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_11_spill" + }, + "patch52": { + "offset": 539932, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_8_spill" + }, + "patch53": { + "offset": 539980, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_11_spill" + }, + "patch54": { + "offset": 540040, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_8_spill" + }, + "patch55": { + "offset": 540088, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_11_spill" + }, + "patch56": { + "offset": 540148, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_13_wts_ddr" + }, + "patch57": { + "offset": 540208, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_13_wts_ddr" + }, + "patch58": { + "offset": 552292, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_12_spill" + }, + "patch59": { + "offset": 552352, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_12_spill" + }, + "patch60": { + "offset": 807300, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_12_spill" + }, + "patch61": { + "offset": 807360, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_12_spill" + }, + "patch62": { + "offset": 807612, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_14_wts_ddr" + }, + "patch63": { + "offset": 807672, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_14_wts_ddr" + }, + "patch64": { + "offset": 835836, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_13_spill" + }, + "patch65": { + "offset": 835896, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_13_spill" + }, + "patch66": { + "offset": 836268, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_13_spill" + }, + "patch67": { + "offset": 836316, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_12_spill" + }, + "patch68": { + "offset": 836376, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_13_spill" + }, + "patch69": { + "offset": 836424, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_12_spill" + }, + "patch70": { + "offset": 836484, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_15_wts_ddr" + }, + "patch71": { + "offset": 836544, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_15_wts_ddr" + }, + "patch72": { + "offset": 848628, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_14_spill" + }, + "patch73": { + "offset": 848676, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch74": { + "offset": 848736, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_14_spill" + }, + "patch75": { + "offset": 848784, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch76": { + "offset": 855276, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_14_spill" + }, + "patch77": { + "offset": 855336, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_14_spill" + }, + "patch78": { + "offset": 855948, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch79": { + "offset": 856008, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch80": { + "offset": 856068, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_16_wts_ddr" + }, + "patch81": { + "offset": 856128, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_16_wts_ddr" + }, + "patch82": { + "offset": 872604, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch83": { + "offset": 872664, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch84": { + "offset": 872724, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch85": { + "offset": 872784, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch86": { + "offset": 872844, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch87": { + "offset": 872904, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch88": { + "offset": 884700, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_16_spill" + }, + "patch89": { + "offset": 884760, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_16_spill" + }, + "patch90": { + "offset": 887460, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch91": { + "offset": 887520, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch92": { + "offset": 887580, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch93": { + "offset": 887640, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch94": { + "offset": 887700, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch95": { + "offset": 887760, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_15_spill" + }, + "patch96": { + "offset": 891492, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_16_spill" + }, + "patch97": { + "offset": 891552, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_17_spill" + }, + "patch98": { + "offset": 892164, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_17_spill" + }, + "patch99": { + "offset": 892224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_17_spill" + }, + "patch100": { + "offset": 892620, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_17_spill" + }, + "patch101": { + "offset": 892680, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_17_spill" + }, + "patch102": { + "offset": 892740, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_18_wts_ddr" + }, + "patch103": { + "offset": 892800, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_18_wts_ddr" + }, + "patch104": { + "offset": 893172, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_19_wts_ddr" + }, + "patch105": { + "offset": 893232, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_19_wts_ddr" + }, + "patch106": { + "offset": 904188, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818" + }, + "patch107": { + "offset": 904248, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818" + }, + "patch108": { + "offset": 904308, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_20_wts_ddr" + }, + "patch109": { + "offset": 904368, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_20_wts_ddr" + }, + "patch110": { + "offset": 918300, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_19_spill" + }, + "patch111": { + "offset": 918360, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_19_spill" + }, + "patch112": { + "offset": 918660, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_20_spill" + }, + "patch113": { + "offset": 918720, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_20_spill" + }, + "patch114": { + "offset": 918780, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_19_spill" + }, + "patch115": { + "offset": 918840, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_19_spill" + }, + "patch116": { + "offset": 918900, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_19_spill" + }, + "patch117": { + "offset": 936552, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_20_spill" + }, + "patch118": { + "offset": 936612, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_21_spill" + }, + "patch119": { + "offset": 937224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_21_spill" + }, + "patch120": { + "offset": 937284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_21_spill" + }, + "patch121": { + "offset": 937680, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818" + }, + "patch122": { + "offset": 937740, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818" + }, + "patch123": { + "offset": 937800, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_22_wts_ddr" + }, + "patch124": { + "offset": 937860, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_22_wts_ddr" + }, + "patch125": { + "offset": 938232, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_21_spill" + }, + "patch126": { + "offset": 938292, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_21_spill" + }, + "patch127": { + "offset": 938544, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_23_wts_ddr" + }, + "patch128": { + "offset": 938604, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_23_wts_ddr" + }, + "patch129": { + "offset": 951720, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_256" + }, + "patch130": { + "offset": 951780, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_256" + }, + "patch131": { + "offset": 951840, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_24_wts_ddr" + }, + "patch132": { + "offset": 951900, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_24_wts_ddr" + }, + "patch133": { + "offset": 979608, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424" + }, + "patch134": { + "offset": 979668, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424" + }, + "patch135": { + "offset": 979872, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_24_spill" + }, + "patch136": { + "offset": 979932, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_24_spill" + }, + "patch137": { + "offset": 980232, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_24_spill" + }, + "patch138": { + "offset": 980292, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_25_spill" + }, + "patch139": { + "offset": 980352, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_25_spill" + }, + "patch140": { + "offset": 1229172, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_25_spill" + }, + "patch141": { + "offset": 1229364, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_27_wts_ddr" + }, + "patch142": { + "offset": 1229424, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_27_wts_ddr" + }, + "patch143": { + "offset": 1551756, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_28_wts_ddr" + }, + "patch144": { + "offset": 1551816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_28_wts_ddr" + }, + "patch145": { + "offset": 1596516, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_31_spill" + }, + "patch146": { + "offset": 1596756, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_31_spill" + }, + "patch147": { + "offset": 1596816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_32_spill" + }, + "patch148": { + "offset": 1597380, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_32_spill" + }, + "patch149": { + "offset": 1597548, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_32_spill" + }, + "patch150": { + "offset": 1597812, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_32_spill" + }, + "patch151": { + "offset": 1597872, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_32_spill" + }, + "patch152": { + "offset": 1598412, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424" + }, + "patch153": { + "offset": 1598472, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424" + }, + "patch154": { + "offset": 1598724, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_32_spill" + }, + "patch155": { + "offset": 1598784, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_32_spill" + }, + "patch156": { + "offset": 1599036, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_34_wts_ddr" + }, + "patch157": { + "offset": 1599096, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_34_wts_ddr" + }, + "patch158": { + "offset": 1611492, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_35_wts_ddr" + }, + "patch159": { + "offset": 1611552, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_35_wts_ddr" + }, + "patch160": { + "offset": 1621956, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434" + }, + "patch161": { + "offset": 1622016, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434" + }, + "patch162": { + "offset": 1622076, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_36_wts_ddr" + }, + "patch163": { + "offset": 1622136, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_36_wts_ddr" + }, + "patch164": { + "offset": 1643748, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_36_spill" + }, + "patch165": { + "offset": 1643808, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_36_spill" + }, + "patch166": { + "offset": 1644012, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636" + }, + "patch167": { + "offset": 1644072, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636" + }, + "patch168": { + "offset": 1644372, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_36_spill" + }, + "patch169": { + "offset": 1644432, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_37_spill" + }, + "patch170": { + "offset": 1644480, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_37_spill" + }, + "patch171": { + "offset": 1893300, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_37_spill" + }, + "patch172": { + "offset": 1893492, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_39_wts_ddr" + }, + "patch173": { + "offset": 1893552, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_39_wts_ddr" + }, + "patch174": { + "offset": 2250492, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_40_wts_ddr" + }, + "patch175": { + "offset": 2250552, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_40_wts_ddr" + }, + "patch176": { + "offset": 2295156, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_43_spill" + }, + "patch177": { + "offset": 2295396, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_43_spill" + }, + "patch178": { + "offset": 2295456, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_44_spill" + }, + "patch179": { + "offset": 2296020, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_44_spill" + }, + "patch180": { + "offset": 2296188, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_44_spill" + }, + "patch181": { + "offset": 2296452, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_44_spill" + }, + "patch182": { + "offset": 2296512, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_44_spill" + }, + "patch183": { + "offset": 2297052, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636" + }, + "patch184": { + "offset": 2297112, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636" + }, + "patch185": { + "offset": 2297364, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_44_spill" + }, + "patch186": { + "offset": 2297424, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_44_spill" + }, + "patch187": { + "offset": 2297676, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_46_wts_ddr" + }, + "patch188": { + "offset": 2297736, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_46_wts_ddr" + }, + "patch189": { + "offset": 2312268, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434" + }, + "patch190": { + "offset": 2312328, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434" + }, + "patch191": { + "offset": 2312580, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_47_wts_ddr" + }, + "patch192": { + "offset": 2312640, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_47_wts_ddr" + }, + "patch193": { + "offset": 2323920, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646" + }, + "patch194": { + "offset": 2323980, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646" + }, + "patch195": { + "offset": 2324040, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_48_wts_ddr" + }, + "patch196": { + "offset": 2324100, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_48_wts_ddr" + }, + "patch197": { + "offset": 2345712, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848" + }, + "patch198": { + "offset": 2345772, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848" + }, + "patch199": { + "offset": 2345976, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_48_spill" + }, + "patch200": { + "offset": 2346036, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_48_spill" + }, + "patch201": { + "offset": 2346336, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_48_spill" + }, + "patch202": { + "offset": 2346396, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_49_spill" + }, + "patch203": { + "offset": 2346444, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_49_spill" + }, + "patch204": { + "offset": 2595264, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_49_spill" + }, + "patch205": { + "offset": 2595456, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_51_wts_ddr" + }, + "patch206": { + "offset": 2595516, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_51_wts_ddr" + }, + "patch207": { + "offset": 2952456, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_52_wts_ddr" + }, + "patch208": { + "offset": 2952516, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_52_wts_ddr" + }, + "patch209": { + "offset": 2997120, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_55_spill" + }, + "patch210": { + "offset": 2997360, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_55_spill" + }, + "patch211": { + "offset": 2997420, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_56_spill" + }, + "patch212": { + "offset": 2997984, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_56_spill" + }, + "patch213": { + "offset": 2998152, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_56_spill" + }, + "patch214": { + "offset": 2998416, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_56_spill" + }, + "patch215": { + "offset": 2998476, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_56_spill" + }, + "patch216": { + "offset": 2999016, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848" + }, + "patch217": { + "offset": 2999076, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848" + }, + "patch218": { + "offset": 2999328, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_56_spill" + }, + "patch219": { + "offset": 2999388, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_56_spill" + }, + "patch220": { + "offset": 2999640, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_58_wts_ddr" + }, + "patch221": { + "offset": 2999700, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_58_wts_ddr" + }, + "patch222": { + "offset": 3014232, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646" + }, + "patch223": { + "offset": 3014292, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646" + }, + "patch224": { + "offset": 3014544, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_59_wts_ddr" + }, + "patch225": { + "offset": 3014604, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_59_wts_ddr" + }, + "patch226": { + "offset": 3025836, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_236" + }, + "patch227": { + "offset": 3025896, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_236" + }, + "patch228": { + "offset": 3037452, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_59_spill" + }, + "patch229": { + "offset": 3037512, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_59_spill" + }, + "patch230": { + "offset": 3065340, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_62_spill" + }, + "patch231": { + "offset": 3065400, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_62_spill" + }, + "patch232": { + "offset": 3065724, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_59_spill" + }, + "patch233": { + "offset": 3065772, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_62_spill" + }, + "patch234": { + "offset": 3065832, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_59_spill" + }, + "patch235": { + "offset": 3065880, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_62_spill" + }, + "patch236": { + "offset": 3065940, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_59_spill" + }, + "patch237": { + "offset": 3065988, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_62_spill" + }, + "patch238": { + "offset": 3077376, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_63_spill" + }, + "patch239": { + "offset": 3077436, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_63_spill" + }, + "patch240": { + "offset": 3082296, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_63_spill" + }, + "patch241": { + "offset": 3082356, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_64_spill" + }, + "patch242": { + "offset": 3082968, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_64_spill" + }, + "patch243": { + "offset": 3083028, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_64_spill" + }, + "patch244": { + "offset": 3083424, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_64_spill" + }, + "patch245": { + "offset": 3083484, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_64_spill" + }, + "patch246": { + "offset": 3083544, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_65_wts_ddr" + }, + "patch247": { + "offset": 3083604, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_65_wts_ddr" + }, + "patch248": { + "offset": 3119976, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_70_wts_ddr" + }, + "patch249": { + "offset": 3120036, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_70_wts_ddr" + }, + "patch250": { + "offset": 3131184, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_71_wts_ddr" + }, + "patch251": { + "offset": 3131244, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_71_wts_ddr" + }, + "patch252": { + "offset": 3177144, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_76_wts_ddr" + }, + "patch253": { + "offset": 3177204, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_76_wts_ddr" + }, + "patch254": { + "offset": 3223824, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_81_wts_ddr" + }, + "patch255": { + "offset": 3223884, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_81_wts_ddr" + }, + "patch256": { + "offset": 3238488, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_82_wts_ddr" + }, + "patch257": { + "offset": 3238548, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_82_wts_ddr" + }, + "patch258": { + "offset": 3284400, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_87_wts_ddr" + }, + "patch259": { + "offset": 3284460, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_87_wts_ddr" + }, + "patch260": { + "offset": 3330504, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_92_wts_ddr" + }, + "patch261": { + "offset": 3330564, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_92_wts_ddr" + }, + "patch262": { + "offset": 3345168, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_93_wts_ddr" + }, + "patch263": { + "offset": 3345228, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_93_wts_ddr" + }, + "patch264": { + "offset": 3391080, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_98_wts_ddr" + }, + "patch265": { + "offset": 3391140, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_98_wts_ddr" + }, + "patch266": { + "offset": 3437184, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_103_wts_ddr" + }, + "patch267": { + "offset": 3437244, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_103_wts_ddr" + }, + "patch268": { + "offset": 3451848, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_104_wts_ddr" + }, + "patch269": { + "offset": 3451908, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_104_wts_ddr" + }, + "patch270": { + "offset": 3498048, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_109_wts_ddr" + }, + "patch271": { + "offset": 3498108, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_109_wts_ddr" + }, + "patch272": { + "offset": 3552912, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113" + }, + "patch273": { + "offset": 3552972, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113" + }, + "patch274": { + "offset": 3553176, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_113_spill" + }, + "patch275": { + "offset": 3553236, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_113_spill" + }, + "patch276": { + "offset": 3553536, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_113_spill" + }, + "patch277": { + "offset": 3553596, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_114_spill" + }, + "patch278": { + "offset": 3802416, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_114_spill" + }, + "patch279": { + "offset": 3802608, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_116_wts_ddr" + }, + "patch280": { + "offset": 3802668, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_116_wts_ddr" + }, + "patch281": { + "offset": 4122504, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_117_wts_ddr" + }, + "patch282": { + "offset": 4122564, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_117_wts_ddr" + }, + "patch283": { + "offset": 4167168, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_120_spill" + }, + "patch284": { + "offset": 4167408, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_120_spill" + }, + "patch285": { + "offset": 4167468, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_121_spill" + }, + "patch286": { + "offset": 4168032, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_121_spill" + }, + "patch287": { + "offset": 4168200, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_121_spill" + }, + "patch288": { + "offset": 4168464, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_121_spill" + }, + "patch289": { + "offset": 4168524, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_121_spill" + }, + "patch290": { + "offset": 4169040, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113" + }, + "patch291": { + "offset": 4169100, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113" + }, + "patch292": { + "offset": 4169352, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_121_spill" + }, + "patch293": { + "offset": 4169412, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_121_spill" + }, + "patch294": { + "offset": 4169664, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_123_wts_ddr" + }, + "patch295": { + "offset": 4169724, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_123_wts_ddr" + }, + "patch296": { + "offset": 4180872, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_124_wts_ddr" + }, + "patch297": { + "offset": 4180932, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_124_wts_ddr" + }, + "patch298": { + "offset": 4191288, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123" + }, + "patch299": { + "offset": 4191348, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123" + }, + "patch300": { + "offset": 4227720, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_129_wts_ddr" + }, + "patch301": { + "offset": 4227780, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_129_wts_ddr" + }, + "patch302": { + "offset": 4282584, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_133_spill" + }, + "patch303": { + "offset": 4282644, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_133_spill" + }, + "patch304": { + "offset": 4282848, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133" + }, + "patch305": { + "offset": 4282908, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133" + }, + "patch306": { + "offset": 4283208, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_133_spill" + }, + "patch307": { + "offset": 4283268, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_134_spill" + }, + "patch308": { + "offset": 4532136, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_134_spill" + }, + "patch309": { + "offset": 4532328, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_136_wts_ddr" + }, + "patch310": { + "offset": 4532388, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_136_wts_ddr" + }, + "patch311": { + "offset": 4887984, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_137_wts_ddr" + }, + "patch312": { + "offset": 4888044, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_137_wts_ddr" + }, + "patch313": { + "offset": 4932744, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_140_spill" + }, + "patch314": { + "offset": 4932984, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_140_spill" + }, + "patch315": { + "offset": 4933044, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_141_spill" + }, + "patch316": { + "offset": 4933608, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_141_spill" + }, + "patch317": { + "offset": 4933776, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_141_spill" + }, + "patch318": { + "offset": 4934040, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_141_spill" + }, + "patch319": { + "offset": 4934100, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_141_spill" + }, + "patch320": { + "offset": 4934616, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133" + }, + "patch321": { + "offset": 4934676, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133" + }, + "patch322": { + "offset": 4934928, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_141_spill" + }, + "patch323": { + "offset": 4934988, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_141_spill" + }, + "patch324": { + "offset": 4935240, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_143_wts_ddr" + }, + "patch325": { + "offset": 4935300, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_143_wts_ddr" + }, + "patch326": { + "offset": 4955640, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123" + }, + "patch327": { + "offset": 4955700, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123" + }, + "patch328": { + "offset": 4955952, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_144_wts_ddr" + }, + "patch329": { + "offset": 4956012, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_144_wts_ddr" + }, + "patch330": { + "offset": 5002920, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_149_wts_ddr" + }, + "patch331": { + "offset": 5002980, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_149_wts_ddr" + }, + "patch332": { + "offset": 5058744, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153" + }, + "patch333": { + "offset": 5058804, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153" + }, + "patch334": { + "offset": 5059008, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_153_spill" + }, + "patch335": { + "offset": 5059068, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_153_spill" + }, + "patch336": { + "offset": 5059368, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_153_spill" + }, + "patch337": { + "offset": 5059428, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_154_spill" + }, + "patch338": { + "offset": 5308296, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_154_spill" + }, + "patch339": { + "offset": 5308488, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_156_wts_ddr" + }, + "patch340": { + "offset": 5308548, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_156_wts_ddr" + }, + "patch341": { + "offset": 5628624, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_157_wts_ddr" + }, + "patch342": { + "offset": 5628684, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_157_wts_ddr" + }, + "patch343": { + "offset": 5673384, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_160_spill" + }, + "patch344": { + "offset": 5673624, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_160_spill" + }, + "patch345": { + "offset": 5673684, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_161_spill" + }, + "patch346": { + "offset": 5674248, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_161_spill" + }, + "patch347": { + "offset": 5674416, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_161_spill" + }, + "patch348": { + "offset": 5674680, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_161_spill" + }, + "patch349": { + "offset": 5674740, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_161_spill" + }, + "patch350": { + "offset": 5675256, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153" + }, + "patch351": { + "offset": 5675316, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153" + }, + "patch352": { + "offset": 5675568, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_161_spill" + }, + "patch353": { + "offset": 5675628, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_161_spill" + }, + "patch354": { + "offset": 5675880, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_163_wts_ddr" + }, + "patch355": { + "offset": 5675940, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_163_wts_ddr" + }, + "patch356": { + "offset": 5698176, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_164_wts_ddr" + }, + "patch357": { + "offset": 5698236, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_164_wts_ddr" + }, + "patch358": { + "offset": 5698464, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163" + }, + "patch359": { + "offset": 5698524, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163" + }, + "patch360": { + "offset": 5708160, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_164_spill" + }, + "patch361": { + "offset": 5708220, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_164_spill" + }, + "patch362": { + "offset": 5734320, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_167_spill" + }, + "patch363": { + "offset": 5734380, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_167_spill" + }, + "patch364": { + "offset": 5734752, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_164_spill" + }, + "patch365": { + "offset": 5734800, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_167_spill" + }, + "patch366": { + "offset": 5734860, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_164_spill" + }, + "patch367": { + "offset": 5734908, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_167_spill" + }, + "patch368": { + "offset": 5734968, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_169_wts_ddr" + }, + "patch369": { + "offset": 5735028, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_169_wts_ddr" + }, + "patch370": { + "offset": 5746344, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_168_spill" + }, + "patch371": { + "offset": 5746404, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_168_spill" + }, + "patch372": { + "offset": 5747592, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_168_spill" + }, + "patch373": { + "offset": 5747652, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_168_spill" + }, + "patch374": { + "offset": 5757072, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_169_spill" + }, + "patch375": { + "offset": 5757132, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_169_spill" + }, + "patch376": { + "offset": 5783232, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_172_spill" + }, + "patch377": { + "offset": 5783292, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_172_spill" + }, + "patch378": { + "offset": 5783616, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_169_spill" + }, + "patch379": { + "offset": 5783664, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_172_spill" + }, + "patch380": { + "offset": 5783724, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_169_spill" + }, + "patch381": { + "offset": 5783772, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_172_spill" + }, + "patch382": { + "offset": 5793048, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_173_spill" + }, + "patch383": { + "offset": 5793108, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_173_spill" + }, + "patch384": { + "offset": 5794368, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_173_spill" + }, + "patch385": { + "offset": 5794428, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_174_spill" + }, + "patch386": { + "offset": 6043248, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_174_spill" + }, + "patch387": { + "offset": 6043440, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_176_wts_ddr" + }, + "patch388": { + "offset": 6043500, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_176_wts_ddr" + }, + "patch389": { + "offset": 6408984, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_177_wts_ddr" + }, + "patch390": { + "offset": 6409044, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_177_wts_ddr" + }, + "patch391": { + "offset": 6443568, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_180_spill" + }, + "patch392": { + "offset": 6443808, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_180_spill" + }, + "patch393": { + "offset": 6443868, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_181_spill" + }, + "patch394": { + "offset": 6444432, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_181_spill" + }, + "patch395": { + "offset": 6444600, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_181_spill" + }, + "patch396": { + "offset": 6444864, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_181_spill" + }, + "patch397": { + "offset": 6444924, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_181_spill" + }, + "patch398": { + "offset": 6445440, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_173_spill" + }, + "patch399": { + "offset": 6445500, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_173_spill" + }, + "patch400": { + "offset": 6445752, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_181_spill" + }, + "patch401": { + "offset": 6445812, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_181_spill" + }, + "patch402": { + "offset": 6446064, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_183_wts_ddr" + }, + "patch403": { + "offset": 6446124, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_183_wts_ddr" + }, + "patch404": { + "offset": 6467952, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163" + }, + "patch405": { + "offset": 6468012, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163" + }, + "patch406": { + "offset": 6480024, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_184_wts_ddr" + }, + "patch407": { + "offset": 6480084, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_184_wts_ddr" + }, + "patch408": { + "offset": 6480312, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183" + }, + "patch409": { + "offset": 6480372, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183" + }, + "patch410": { + "offset": 6490008, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_184_spill" + }, + "patch411": { + "offset": 6490068, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_184_spill" + }, + "patch412": { + "offset": 6516168, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_187_spill" + }, + "patch413": { + "offset": 6516228, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_187_spill" + }, + "patch414": { + "offset": 6516600, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_184_spill" + }, + "patch415": { + "offset": 6516648, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_187_spill" + }, + "patch416": { + "offset": 6516708, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_184_spill" + }, + "patch417": { + "offset": 6516756, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_187_spill" + }, + "patch418": { + "offset": 6516816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_189_wts_ddr" + }, + "patch419": { + "offset": 6516876, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_189_wts_ddr" + }, + "patch420": { + "offset": 6528192, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_188_spill" + }, + "patch421": { + "offset": 6528252, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_188_spill" + }, + "patch422": { + "offset": 6529440, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_188_spill" + }, + "patch423": { + "offset": 6529500, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_188_spill" + }, + "patch424": { + "offset": 6538920, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_189_spill" + }, + "patch425": { + "offset": 6538980, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_189_spill" + }, + "patch426": { + "offset": 6565080, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_192_spill" + }, + "patch427": { + "offset": 6565140, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_192_spill" + }, + "patch428": { + "offset": 6565464, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_189_spill" + }, + "patch429": { + "offset": 6565512, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_192_spill" + }, + "patch430": { + "offset": 6565572, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_189_spill" + }, + "patch431": { + "offset": 6565620, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_192_spill" + }, + "patch432": { + "offset": 6574896, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_193_spill" + }, + "patch433": { + "offset": 6574956, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_193_spill" + }, + "patch434": { + "offset": 6576216, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_193_spill" + }, + "patch435": { + "offset": 6576276, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_194_spill" + }, + "patch436": { + "offset": 6825096, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_194_spill" + }, + "patch437": { + "offset": 6825288, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_196_wts_ddr" + }, + "patch438": { + "offset": 6825348, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_196_wts_ddr" + }, + "patch439": { + "offset": 7135024, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_197_wts_ddr" + }, + "patch440": { + "offset": 7135084, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_197_wts_ddr" + }, + "patch441": { + "offset": 7169608, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_200_spill" + }, + "patch442": { + "offset": 7169848, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_200_spill" + }, + "patch443": { + "offset": 7169908, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_201_spill" + }, + "patch444": { + "offset": 7170472, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_201_spill" + }, + "patch445": { + "offset": 7170640, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_201_spill" + }, + "patch446": { + "offset": 7170904, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_201_spill" + }, + "patch447": { + "offset": 7170964, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_201_spill" + }, + "patch448": { + "offset": 7171480, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_193_spill" + }, + "patch449": { + "offset": 7171540, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_193_spill" + }, + "patch450": { + "offset": 7171792, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_201_spill" + }, + "patch451": { + "offset": 7171852, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_201_spill" + }, + "patch452": { + "offset": 7172104, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_203_wts_ddr" + }, + "patch453": { + "offset": 7172164, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_203_wts_ddr" + }, + "patch454": { + "offset": 7193992, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183" + }, + "patch455": { + "offset": 7194052, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183" + }, + "patch456": { + "offset": 7206064, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_204_wts_ddr" + }, + "patch457": { + "offset": 7206124, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_204_wts_ddr" + }, + "patch458": { + "offset": 7215784, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_204_spill" + }, + "patch459": { + "offset": 7215844, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_204_spill" + }, + "patch460": { + "offset": 7241944, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_207_spill" + }, + "patch461": { + "offset": 7242004, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_207_spill" + }, + "patch462": { + "offset": 7242328, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_204_spill" + }, + "patch463": { + "offset": 7242376, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_207_spill" + }, + "patch464": { + "offset": 7242436, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_204_spill" + }, + "patch465": { + "offset": 7242484, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_207_spill" + }, + "patch466": { + "offset": 7251760, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_208_spill" + }, + "patch467": { + "offset": 7251820, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_208_spill" + }, + "patch468": { + "offset": 7253080, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_208_spill" + }, + "patch469": { + "offset": 7253140, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_209_spill" + }, + "patch470": { + "offset": 7501960, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_209_spill" + }, + "patch471": { + "offset": 7502152, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_211_wts_ddr" + }, + "patch472": { + "offset": 7502212, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_211_wts_ddr" + }, + "patch473": { + "offset": 7856592, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_212_spill" + }, + "patch474": { + "offset": 7856832, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_212_spill" + }, + "patch475": { + "offset": 7856892, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_213_spill" + }, + "patch476": { + "offset": 7857456, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_0_213_spill" + }, + "patch477": { + "offset": 7857624, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_213_spill" + }, + "patch478": { + "offset": 7857888, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_scratch_1_213_spill" + }, + "patch479": { + "offset": 7857948, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_213_spill" + }, + "patch480": { + "offset": 7858536, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_213_spill" + }, + "patch481": { + "offset": 7858596, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_213_spill" + }, + "patch482": { + "offset": 7858656, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_214_wts_ddr" + }, + "patch483": { + "offset": 7858716, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_214_wts_ddr" + }, + "patch484": { + "offset": 7859064, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_208_spill" + }, + "patch485": { + "offset": 7859124, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_208_spill" + }, + "patch486": { + "offset": 7869672, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_214_spill" + }, + "patch487": { + "offset": 7869732, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_214_spill" + }, + "patch488": { + "offset": 7869984, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_214_spill" + }, + "patch489": { + "offset": 7870044, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_214_spill" + }, + "patch490": { + "offset": 7870272, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_215_spill" + }, + "patch491": { + "offset": 7870332, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_215_spill" + }, + "patch492": { + "offset": 7870584, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_214_spill" + }, + "patch493": { + "offset": 7870644, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_214_spill" + }, + "patch494": { + "offset": 7870872, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_216_spill" + }, + "patch495": { + "offset": 7870932, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_216_spill" + }, + "patch496": { + "offset": 7871160, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_5_spill" + }, + "patch497": { + "offset": 7871220, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_5_spill" + }, + "patch498": { + "offset": 7871448, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch499": { + "offset": 7871508, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch500": { + "offset": 7871760, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_1" + }, + "patch501": { + "offset": 7871820, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_1" + }, + "patch502": { + "offset": 7872024, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279" + }, + "patch503": { + "offset": 7872084, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279" + }, + "patch504": { + "offset": 7872312, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_3" + }, + "patch505": { + "offset": 7872372, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_3" + }, + "patch506": { + "offset": 7872576, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_240" + }, + "patch507": { + "offset": 7872636, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_240" + }, + "patch508": { + "offset": 7872888, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_215_spill" + }, + "patch509": { + "offset": 7872948, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_215_spill" + }, + "patch510": { + "offset": 7873152, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_230" + }, + "patch511": { + "offset": 7873212, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_230" + }, + "patch512": { + "offset": 7873464, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_216_spill" + }, + "patch513": { + "offset": 7873524, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_216_spill" + }, + "patch514": { + "offset": 7873728, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_225" + }, + "patch515": { + "offset": 7873788, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_225" + }, + "patch516": { + "offset": 7874016, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_4" + }, + "patch517": { + "offset": 7874076, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_4" + }, + "patch518": { + "offset": 7874280, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_217" + }, + "patch519": { + "offset": 7874340, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_217" + }, + "patch520": { + "offset": 7874592, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_216_spill" + }, + "patch521": { + "offset": 7874652, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_216_spill" + }, + "patch522": { + "offset": 7874856, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_217" + }, + "patch523": { + "offset": 7874916, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_217" + }, + "patch524": { + "offset": 7875192, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_217" + }, + "patch525": { + "offset": 7875252, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_217" + }, + "patch526": { + "offset": 7875312, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_218_wts_ddr" + }, + "patch527": { + "offset": 7875372, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_218_wts_ddr" + }, + "patch528": { + "offset": 7893840, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_219_spill" + }, + "patch529": { + "offset": 7893900, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_219_spill" + }, + "patch530": { + "offset": 7894152, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_219_spill" + }, + "patch531": { + "offset": 7894212, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_219_spill" + }, + "patch532": { + "offset": 7894440, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_220_spill" + }, + "patch533": { + "offset": 7894500, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_220_spill" + }, + "patch534": { + "offset": 7894752, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_220_spill" + }, + "patch535": { + "offset": 7894812, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_220_spill" + }, + "patch536": { + "offset": 7895040, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr_3" + }, + "patch537": { + "offset": 7895100, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr_3" + }, + "patch538": { + "offset": 7905432, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_4" + }, + "patch539": { + "offset": 7905492, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_4" + }, + "patch540": { + "offset": 7915800, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222" + }, + "patch541": { + "offset": 7915860, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222" + }, + "patch542": { + "offset": 7916112, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_219_spill" + }, + "patch543": { + "offset": 7916172, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_219_spill" + }, + "patch544": { + "offset": 7916400, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_223_spill" + }, + "patch545": { + "offset": 7916460, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_223_spill" + }, + "patch546": { + "offset": 7916712, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_223_spill" + }, + "patch547": { + "offset": 7916772, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_223_spill" + }, + "patch548": { + "offset": 7917000, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_4" + }, + "patch549": { + "offset": 7917060, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_4" + }, + "patch550": { + "offset": 7917312, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_226_wts_ddr" + }, + "patch551": { + "offset": 7917372, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_226_wts_ddr" + }, + "patch552": { + "offset": 7927656, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_225" + }, + "patch553": { + "offset": 7927716, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_225" + }, + "patch554": { + "offset": 7927944, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_225" + }, + "patch555": { + "offset": 7928004, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_225" + }, + "patch556": { + "offset": 7946400, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_220_spill" + }, + "patch557": { + "offset": 7946460, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_220_spill" + }, + "patch558": { + "offset": 7956792, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222" + }, + "patch559": { + "offset": 7956852, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222" + }, + "patch560": { + "offset": 7965240, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_230" + }, + "patch561": { + "offset": 7965300, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_230" + }, + "patch562": { + "offset": 7965504, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_0_l2l3_229_spill" + }, + "patch563": { + "offset": 7965564, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_0_l2l3_229_spill" + }, + "patch564": { + "offset": 8304544, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_230" + }, + "patch565": { + "offset": 8304604, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_230" + }, + "patch566": { + "offset": 8312824, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_231_spill" + }, + "patch567": { + "offset": 8312884, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_231_spill" + }, + "patch568": { + "offset": 8313160, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_231_spill" + }, + "patch569": { + "offset": 8313220, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_231_spill" + }, + "patch570": { + "offset": 8313448, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_232_spill" + }, + "patch571": { + "offset": 8313508, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_232_spill" + }, + "patch572": { + "offset": 8313808, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_5_spill" + }, + "patch573": { + "offset": 8313868, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_5_spill" + }, + "patch574": { + "offset": 8324008, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_233_spill" + }, + "patch575": { + "offset": 8324056, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch576": { + "offset": 8324116, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_233_spill" + }, + "patch577": { + "offset": 8324164, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch578": { + "offset": 8359288, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_233_spill" + }, + "patch579": { + "offset": 8359348, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_233_spill" + }, + "patch580": { + "offset": 8375224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_256" + }, + "patch581": { + "offset": 8375284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_256" + }, + "patch582": { + "offset": 8387104, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_236" + }, + "patch583": { + "offset": 8387164, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_236" + }, + "patch584": { + "offset": 8387416, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_232_spill" + }, + "patch585": { + "offset": 8387476, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_232_spill" + }, + "patch586": { + "offset": 8387680, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_236" + }, + "patch587": { + "offset": 8387740, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_236" + }, + "patch588": { + "offset": 8388016, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_236" + }, + "patch589": { + "offset": 8388076, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_236" + }, + "patch590": { + "offset": 8388136, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_237_wts_ddr" + }, + "patch591": { + "offset": 8388196, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_237_wts_ddr" + }, + "patch592": { + "offset": 8399008, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_237_spill" + }, + "patch593": { + "offset": 8399068, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_237_spill" + }, + "patch594": { + "offset": 8399320, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_237_spill" + }, + "patch595": { + "offset": 8399380, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_237_spill" + }, + "patch596": { + "offset": 8399608, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_238_spill" + }, + "patch597": { + "offset": 8399668, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_238_spill" + }, + "patch598": { + "offset": 8399920, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_237_spill" + }, + "patch599": { + "offset": 8399980, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_237_spill" + }, + "patch600": { + "offset": 8400208, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_239_spill" + }, + "patch601": { + "offset": 8400268, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_239_spill" + }, + "patch602": { + "offset": 8400520, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_239_spill" + }, + "patch603": { + "offset": 8400580, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_239_spill" + }, + "patch604": { + "offset": 8400784, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_248" + }, + "patch605": { + "offset": 8400844, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_248" + }, + "patch606": { + "offset": 8401096, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_239_spill" + }, + "patch607": { + "offset": 8401156, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_239_spill" + }, + "patch608": { + "offset": 8401360, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_240" + }, + "patch609": { + "offset": 8401420, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_240" + }, + "patch610": { + "offset": 8401696, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_240" + }, + "patch611": { + "offset": 8401756, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_240" + }, + "patch612": { + "offset": 8401816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_241_wts_ddr" + }, + "patch613": { + "offset": 8401876, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_241_wts_ddr" + }, + "patch614": { + "offset": 8706888, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_242_spill" + }, + "patch615": { + "offset": 8706948, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_242_spill" + }, + "patch616": { + "offset": 8707200, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_242_spill" + }, + "patch617": { + "offset": 8707260, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_242_spill" + }, + "patch618": { + "offset": 8707488, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_243_spill" + }, + "patch619": { + "offset": 8707548, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_243_spill" + }, + "patch620": { + "offset": 8707800, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_243_spill" + }, + "patch621": { + "offset": 8707860, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_243_spill" + }, + "patch622": { + "offset": 8708088, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr_2" + }, + "patch623": { + "offset": 8708148, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr_2" + }, + "patch624": { + "offset": 8719248, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_3" + }, + "patch625": { + "offset": 8719308, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_3" + }, + "patch626": { + "offset": 8730384, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245" + }, + "patch627": { + "offset": 8730444, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245" + }, + "patch628": { + "offset": 8730696, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_242_spill" + }, + "patch629": { + "offset": 8730756, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_242_spill" + }, + "patch630": { + "offset": 8730984, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_246_spill" + }, + "patch631": { + "offset": 8731044, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_246_spill" + }, + "patch632": { + "offset": 8731296, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_246_spill" + }, + "patch633": { + "offset": 8731356, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_246_spill" + }, + "patch634": { + "offset": 8731584, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_3" + }, + "patch635": { + "offset": 8731644, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_3" + }, + "patch636": { + "offset": 8731896, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_249_wts_ddr" + }, + "patch637": { + "offset": 8731956, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_249_wts_ddr" + }, + "patch638": { + "offset": 8743008, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_248" + }, + "patch639": { + "offset": 8743068, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_248" + }, + "patch640": { + "offset": 8743296, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_248" + }, + "patch641": { + "offset": 8743356, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_248" + }, + "patch642": { + "offset": 8762808, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_243_spill" + }, + "patch643": { + "offset": 8762868, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_243_spill" + }, + "patch644": { + "offset": 8773968, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245" + }, + "patch645": { + "offset": 8774028, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245" + }, + "patch646": { + "offset": 8783184, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_1_l2l3_252_spill" + }, + "patch647": { + "offset": 8783244, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_1_l2l3_252_spill" + }, + "patch648": { + "offset": 8783448, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_253" + }, + "patch649": { + "offset": 8783508, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_253" + }, + "patch650": { + "offset": 8783760, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_238_spill" + }, + "patch651": { + "offset": 8783820, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_238_spill" + }, + "patch652": { + "offset": 8784024, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_253" + }, + "patch653": { + "offset": 8784084, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_253" + }, + "patch654": { + "offset": 9124480, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_253" + }, + "patch655": { + "offset": 9124540, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_253" + }, + "patch656": { + "offset": 9124600, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_254_spill" + }, + "patch657": { + "offset": 9124660, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_254_spill" + }, + "patch658": { + "offset": 9137224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_254_spill" + }, + "patch659": { + "offset": 9137284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_254_spill" + }, + "patch660": { + "offset": 9137344, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_255_spill" + }, + "patch661": { + "offset": 9137404, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_255_spill" + }, + "patch662": { + "offset": 9138136, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_255_spill" + }, + "patch663": { + "offset": 9138196, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_255_spill" + }, + "patch664": { + "offset": 9138424, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_256" + }, + "patch665": { + "offset": 9138484, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_256" + }, + "patch666": { + "offset": 9138784, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_256" + }, + "patch667": { + "offset": 9138844, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_256" + }, + "patch668": { + "offset": 9138904, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_257_wts_ddr" + }, + "patch669": { + "offset": 9138964, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_257_wts_ddr" + }, + "patch670": { + "offset": 9147256, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_257_spill" + }, + "patch671": { + "offset": 9147316, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_257_spill" + }, + "patch672": { + "offset": 9147688, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_257_spill" + }, + "patch673": { + "offset": 9147748, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_257_spill" + }, + "patch674": { + "offset": 9147808, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_258_spill" + }, + "patch675": { + "offset": 9147868, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_258_spill" + }, + "patch676": { + "offset": 9156808, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_257_spill" + }, + "patch677": { + "offset": 9156868, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_257_spill" + }, + "patch678": { + "offset": 9156928, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_259_spill" + }, + "patch679": { + "offset": 9156988, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_259_spill" + }, + "patch680": { + "offset": 9157048, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_2" + }, + "patch681": { + "offset": 9168016, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_259_spill" + }, + "patch682": { + "offset": 9168076, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_260_spill" + }, + "patch683": { + "offset": 9168136, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_261_wts_ddr" + }, + "patch684": { + "offset": 9168196, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_261_wts_ddr" + }, + "patch685": { + "offset": 9180040, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_260_spill" + }, + "patch686": { + "offset": 9180100, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_260_spill" + }, + "patch687": { + "offset": 9548112, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_262_spill" + }, + "patch688": { + "offset": 9548172, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_262_spill" + }, + "patch689": { + "offset": 9548520, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_262_spill" + }, + "patch690": { + "offset": 9548580, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_262_spill" + }, + "patch691": { + "offset": 9548640, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_263_spill" + }, + "patch692": { + "offset": 9548700, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_263_spill" + }, + "patch693": { + "offset": 9560112, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_263_spill" + }, + "patch694": { + "offset": 9560172, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_263_spill" + }, + "patch695": { + "offset": 9560400, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr_1" + }, + "patch696": { + "offset": 9560460, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr_1" + }, + "patch697": { + "offset": 9571200, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_2" + }, + "patch698": { + "offset": 9571260, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_2" + }, + "patch699": { + "offset": 9579264, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265" + }, + "patch700": { + "offset": 9579324, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265" + }, + "patch701": { + "offset": 9579672, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_262_spill" + }, + "patch702": { + "offset": 9579732, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_262_spill" + }, + "patch703": { + "offset": 9579792, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_266_spill" + }, + "patch704": { + "offset": 9579852, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_266_spill" + }, + "patch705": { + "offset": 9591264, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_266_spill" + }, + "patch706": { + "offset": 9591324, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_266_spill" + }, + "patch707": { + "offset": 9591576, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_2" + }, + "patch708": { + "offset": 9591636, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_2" + }, + "patch709": { + "offset": 9601584, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_267_spill" + }, + "patch710": { + "offset": 9601644, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_267_spill" + }, + "patch711": { + "offset": 9602112, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_259_spill" + }, + "patch712": { + "offset": 9602172, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_268_spill" + }, + "patch713": { + "offset": 9602232, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_267_spill" + }, + "patch714": { + "offset": 9614160, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_268_spill" + }, + "patch715": { + "offset": 9614220, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_268_spill" + }, + "patch716": { + "offset": 9614280, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_269_wts_ddr" + }, + "patch717": { + "offset": 9614340, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_269_wts_ddr" + }, + "patch718": { + "offset": 9634584, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_263_spill" + }, + "patch719": { + "offset": 9634644, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_263_spill" + }, + "patch720": { + "offset": 9645360, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265" + }, + "patch721": { + "offset": 9645420, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265" + }, + "patch722": { + "offset": 9655368, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_2_l2l3_272_spill" + }, + "patch723": { + "offset": 9655428, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_2_l2l3_272_spill" + }, + "patch724": { + "offset": 9655896, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_258_spill" + }, + "patch725": { + "offset": 9655956, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch726": { + "offset": 9656016, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_2_l2l3_272_spill" + }, + "patch727": { + "offset": 10010032, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch728": { + "offset": 10010092, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch729": { + "offset": 10010152, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch730": { + "offset": 10010212, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch731": { + "offset": 10010272, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch732": { + "offset": 10010332, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch733": { + "offset": 10010392, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch734": { + "offset": 10010452, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch735": { + "offset": 10010512, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_276_wts_ddr" + }, + "patch736": { + "offset": 10010572, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_276_wts_ddr" + }, + "patch737": { + "offset": 10022248, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch738": { + "offset": 10022308, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch739": { + "offset": 10022368, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch740": { + "offset": 10022428, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch741": { + "offset": 10022488, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch742": { + "offset": 10022548, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch743": { + "offset": 10022608, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch744": { + "offset": 10022668, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch745": { + "offset": 10023136, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch746": { + "offset": 10023196, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch747": { + "offset": 10023256, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch748": { + "offset": 10023316, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch749": { + "offset": 10023376, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch750": { + "offset": 10023436, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch751": { + "offset": 10023496, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch752": { + "offset": 10023556, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch753": { + "offset": 10023616, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch754": { + "offset": 10023676, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch755": { + "offset": 10023736, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch756": { + "offset": 10023796, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch757": { + "offset": 10023856, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch758": { + "offset": 10023916, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch759": { + "offset": 10023976, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch760": { + "offset": 10024036, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch761": { + "offset": 10024504, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch762": { + "offset": 10024564, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch763": { + "offset": 10024624, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch764": { + "offset": 10024684, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch765": { + "offset": 10024744, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch766": { + "offset": 10024804, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch767": { + "offset": 10024864, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch768": { + "offset": 10024924, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch769": { + "offset": 10024984, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch770": { + "offset": 10025044, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch771": { + "offset": 10025104, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch772": { + "offset": 10025164, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch773": { + "offset": 10025224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch774": { + "offset": 10025284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch775": { + "offset": 10025344, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch776": { + "offset": 10025404, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch777": { + "offset": 10025872, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch778": { + "offset": 10025932, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch779": { + "offset": 10025992, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch780": { + "offset": 10026052, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch781": { + "offset": 10026112, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch782": { + "offset": 10026172, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch783": { + "offset": 10026232, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch784": { + "offset": 10026292, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch785": { + "offset": 10026352, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch786": { + "offset": 10026412, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch787": { + "offset": 10026472, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch788": { + "offset": 10026532, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch789": { + "offset": 10026592, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch790": { + "offset": 10026652, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch791": { + "offset": 10026712, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch792": { + "offset": 10026772, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch793": { + "offset": 10027240, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch794": { + "offset": 10027300, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch795": { + "offset": 10027360, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch796": { + "offset": 10027420, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch797": { + "offset": 10027480, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch798": { + "offset": 10027540, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch799": { + "offset": 10027600, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch800": { + "offset": 10027660, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch801": { + "offset": 10027720, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch802": { + "offset": 10027780, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch803": { + "offset": 10027840, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch804": { + "offset": 10027900, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch805": { + "offset": 10027960, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch806": { + "offset": 10028020, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch807": { + "offset": 10028080, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch808": { + "offset": 10028140, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch809": { + "offset": 10028608, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch810": { + "offset": 10028668, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch811": { + "offset": 10028728, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch812": { + "offset": 10028788, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch813": { + "offset": 10028848, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch814": { + "offset": 10028908, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch815": { + "offset": 10028968, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch816": { + "offset": 10029028, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch817": { + "offset": 10029088, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch818": { + "offset": 10029148, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch819": { + "offset": 10029208, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch820": { + "offset": 10029268, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch821": { + "offset": 10029328, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch822": { + "offset": 10029388, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch823": { + "offset": 10029448, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch824": { + "offset": 10029508, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch825": { + "offset": 10029976, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch826": { + "offset": 10030036, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch827": { + "offset": 10030096, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch828": { + "offset": 10030156, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch829": { + "offset": 10030216, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch830": { + "offset": 10030276, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch831": { + "offset": 10030336, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch832": { + "offset": 10030396, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch833": { + "offset": 10030456, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch834": { + "offset": 10030516, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch835": { + "offset": 10030576, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch836": { + "offset": 10030636, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch837": { + "offset": 10030696, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch838": { + "offset": 10030756, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch839": { + "offset": 10030816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch840": { + "offset": 10030876, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch841": { + "offset": 10031344, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch842": { + "offset": 10031404, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch843": { + "offset": 10031464, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch844": { + "offset": 10031524, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch845": { + "offset": 10031584, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch846": { + "offset": 10031644, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch847": { + "offset": 10031704, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch848": { + "offset": 10031764, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch849": { + "offset": 10031824, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch850": { + "offset": 10031884, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch851": { + "offset": 10031944, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch852": { + "offset": 10032004, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch853": { + "offset": 10032064, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch854": { + "offset": 10032124, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch855": { + "offset": 10032184, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch856": { + "offset": 10032244, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch857": { + "offset": 10032712, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch858": { + "offset": 10032772, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch859": { + "offset": 10032832, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch860": { + "offset": 10032892, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch861": { + "offset": 10032952, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch862": { + "offset": 10033012, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch863": { + "offset": 10033072, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch864": { + "offset": 10033132, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch865": { + "offset": 10033192, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch866": { + "offset": 10033252, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch867": { + "offset": 10033312, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch868": { + "offset": 10033372, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch869": { + "offset": 10033432, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch870": { + "offset": 10033492, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch871": { + "offset": 10033552, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch872": { + "offset": 10033612, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch873": { + "offset": 10034080, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch874": { + "offset": 10034140, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch875": { + "offset": 10034200, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch876": { + "offset": 10034260, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch877": { + "offset": 10034320, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch878": { + "offset": 10034380, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch879": { + "offset": 10034440, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch880": { + "offset": 10034500, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch881": { + "offset": 10034560, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch882": { + "offset": 10034620, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch883": { + "offset": 10034680, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch884": { + "offset": 10034740, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch885": { + "offset": 10034800, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch886": { + "offset": 10034860, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch887": { + "offset": 10034920, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch888": { + "offset": 10034980, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch889": { + "offset": 10035448, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch890": { + "offset": 10035508, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch891": { + "offset": 10035568, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch892": { + "offset": 10035628, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch893": { + "offset": 10035688, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch894": { + "offset": 10035748, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch895": { + "offset": 10035808, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch896": { + "offset": 10035868, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch897": { + "offset": 10035928, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch898": { + "offset": 10035988, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch899": { + "offset": 10036048, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch900": { + "offset": 10036108, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch901": { + "offset": 10036168, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch902": { + "offset": 10036228, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch903": { + "offset": 10036288, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch904": { + "offset": 10036348, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch905": { + "offset": 10036816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch906": { + "offset": 10036876, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch907": { + "offset": 10036936, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch908": { + "offset": 10036996, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch909": { + "offset": 10037056, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch910": { + "offset": 10037116, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch911": { + "offset": 10037176, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch912": { + "offset": 10037236, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_273_spill" + }, + "patch913": { + "offset": 10037296, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch914": { + "offset": 10037356, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch915": { + "offset": 10037416, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch916": { + "offset": 10037476, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch917": { + "offset": 10037536, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch918": { + "offset": 10037596, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch919": { + "offset": 10037656, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch920": { + "offset": 10037716, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch921": { + "offset": 10038328, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch922": { + "offset": 10038388, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch923": { + "offset": 10050088, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_276_spill" + }, + "patch924": { + "offset": 10050148, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_276_spill" + }, + "patch925": { + "offset": 10050640, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch926": { + "offset": 10050700, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch927": { + "offset": 10051600, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch928": { + "offset": 10051660, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch929": { + "offset": 10052560, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch930": { + "offset": 10052620, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch931": { + "offset": 10053520, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch932": { + "offset": 10053580, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275" + }, + "patch933": { + "offset": 10054624, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_276_spill" + }, + "patch934": { + "offset": 10054684, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_276_spill" + }, + "patch935": { + "offset": 10054744, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_277_spill" + }, + "patch936": { + "offset": 10054804, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_277_spill" + }, + "patch937": { + "offset": 10055824, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_276_spill" + }, + "patch938": { + "offset": 10055884, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_276_spill" + }, + "patch939": { + "offset": 10055944, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_278_spill" + }, + "patch940": { + "offset": 10056004, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_278_spill" + }, + "patch941": { + "offset": 10056952, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_278_spill" + }, + "patch942": { + "offset": 10057012, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_278_spill" + }, + "patch943": { + "offset": 10057216, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287" + }, + "patch944": { + "offset": 10057276, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287" + }, + "patch945": { + "offset": 10057528, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_278_spill" + }, + "patch946": { + "offset": 10057588, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_278_spill" + }, + "patch947": { + "offset": 10057792, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279" + }, + "patch948": { + "offset": 10057852, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279" + }, + "patch949": { + "offset": 10058152, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279" + }, + "patch950": { + "offset": 10058212, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279" + }, + "patch951": { + "offset": 10058272, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279" + }, + "patch952": { + "offset": 10058332, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279" + }, + "patch953": { + "offset": 10058392, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279" + }, + "patch954": { + "offset": 10058452, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279" + }, + "patch955": { + "offset": 10058512, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_280_wts_ddr" + }, + "patch956": { + "offset": 10058572, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_280_wts_ddr" + }, + "patch957": { + "offset": 10059136, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_280_spill" + }, + "patch958": { + "offset": 10059196, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_280_spill" + }, + "patch959": { + "offset": 10073200, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_281_spill" + }, + "patch960": { + "offset": 10073260, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_281_spill" + }, + "patch961": { + "offset": 10073320, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_280_spill" + }, + "patch962": { + "offset": 10073380, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_280_spill" + }, + "patch963": { + "offset": 10084816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_281_spill" + }, + "patch964": { + "offset": 10084876, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_281_spill" + }, + "patch965": { + "offset": 10084936, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_282_spill" + }, + "patch966": { + "offset": 10084996, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_282_spill" + }, + "patch967": { + "offset": 10086064, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr" + }, + "patch968": { + "offset": 10086112, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_282_spill" + }, + "patch969": { + "offset": 10086172, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.const_ifm_ddr" + }, + "patch970": { + "offset": 10086220, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_282_spill" + }, + "patch971": { + "offset": 10098208, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_283_spill" + }, + "patch972": { + "offset": 10098268, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_283_spill" + }, + "patch973": { + "offset": 10101376, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_283_spill" + }, + "patch974": { + "offset": 10101424, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_1" + }, + "patch975": { + "offset": 10101484, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_283_spill" + }, + "patch976": { + "offset": 10101532, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_1" + }, + "patch977": { + "offset": 10113520, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_284_spill" + }, + "patch978": { + "offset": 10113580, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_284_spill" + }, + "patch979": { + "offset": 10116544, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_281_spill" + }, + "patch980": { + "offset": 10116604, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_281_spill" + }, + "patch981": { + "offset": 10116664, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_285_spill" + }, + "patch982": { + "offset": 10116724, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_285_spill" + }, + "patch983": { + "offset": 10117864, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_285_spill" + }, + "patch984": { + "offset": 10117912, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_1" + }, + "patch985": { + "offset": 10117972, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_285_spill" + }, + "patch986": { + "offset": 10118020, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ifm_ddr_1" + }, + "patch987": { + "offset": 10118080, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_288_wts_ddr" + }, + "patch988": { + "offset": 10118140, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_288_wts_ddr" + }, + "patch989": { + "offset": 10130224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287" + }, + "patch990": { + "offset": 10130284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287" + }, + "patch991": { + "offset": 10133224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287" + }, + "patch992": { + "offset": 10133284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287" + }, + "patch993": { + "offset": 10133344, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287" + }, + "patch994": { + "offset": 10133404, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287" + }, + "patch995": { + "offset": 10133464, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287" + }, + "patch996": { + "offset": 10133524, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287" + }, + "patch997": { + "offset": 10133992, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_288_spill" + }, + "patch998": { + "offset": 10134052, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_288_spill" + }, + "patch999": { + "offset": 10147984, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_288_spill" + }, + "patch1000": { + "offset": 10148044, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_288_spill" + }, + "patch1001": { + "offset": 10165816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_289_spill" + }, + "patch1002": { + "offset": 10165876, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_289_spill" + }, + "patch1003": { + "offset": 10166272, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_282_spill" + }, + "patch1004": { + "offset": 10166320, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_289_spill" + }, + "patch1005": { + "offset": 10166380, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_282_spill" + }, + "patch1006": { + "offset": 10166428, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_289_spill" + }, + "patch1007": { + "offset": 10178416, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_290_spill" + }, + "patch1008": { + "offset": 10178476, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_290_spill" + }, + "patch1009": { + "offset": 10181584, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_284_spill" + }, + "patch1010": { + "offset": 10181632, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_290_spill" + }, + "patch1011": { + "offset": 10181692, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_284_spill" + }, + "patch1012": { + "offset": 10181740, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_290_spill" + }, + "patch1013": { + "offset": 10191808, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_3_l2l3_291_spill" + }, + "patch1014": { + "offset": 10191856, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1015": { + "offset": 10191916, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_3_l2l3_291_spill" + }, + "patch1016": { + "offset": 10191964, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1017": { + "offset": 10195192, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_277_spill" + }, + "patch1018": { + "offset": 10195252, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_277_spill" + }, + "patch1019": { + "offset": 10195456, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1020": { + "offset": 10195516, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1021": { + "offset": 10195864, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1022": { + "offset": 10195924, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1023": { + "offset": 10195984, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1024": { + "offset": 10196044, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1025": { + "offset": 10196104, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1026": { + "offset": 10196164, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1027": { + "offset": 10196224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_295_wts_ddr" + }, + "patch1028": { + "offset": 10196284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_295_wts_ddr" + }, + "patch1029": { + "offset": 10208464, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1030": { + "offset": 10208524, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1031": { + "offset": 10208584, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1032": { + "offset": 10208644, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1033": { + "offset": 10208704, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1034": { + "offset": 10208764, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1035": { + "offset": 10209136, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1036": { + "offset": 10209196, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1037": { + "offset": 10209256, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1038": { + "offset": 10209316, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1039": { + "offset": 10209376, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1040": { + "offset": 10209436, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1041": { + "offset": 10209496, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1042": { + "offset": 10209556, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1043": { + "offset": 10209616, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1044": { + "offset": 10209676, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1045": { + "offset": 10209736, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1046": { + "offset": 10209796, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1047": { + "offset": 10210168, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1048": { + "offset": 10210228, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1049": { + "offset": 10210288, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1050": { + "offset": 10210348, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1051": { + "offset": 10210408, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1052": { + "offset": 10210468, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1053": { + "offset": 10210528, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1054": { + "offset": 10210588, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1055": { + "offset": 10210648, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1056": { + "offset": 10210708, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1057": { + "offset": 10210768, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1058": { + "offset": 10210828, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1059": { + "offset": 10211200, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1060": { + "offset": 10211260, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1061": { + "offset": 10211320, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1062": { + "offset": 10211380, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1063": { + "offset": 10211440, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1064": { + "offset": 10211500, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1065": { + "offset": 10211560, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1066": { + "offset": 10211620, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1067": { + "offset": 10211680, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1068": { + "offset": 10211740, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1069": { + "offset": 10211800, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1070": { + "offset": 10211860, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1071": { + "offset": 10212232, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1072": { + "offset": 10212292, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1073": { + "offset": 10212352, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1074": { + "offset": 10212412, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1075": { + "offset": 10212472, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1076": { + "offset": 10212532, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1077": { + "offset": 10212592, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1078": { + "offset": 10212652, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1079": { + "offset": 10212712, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1080": { + "offset": 10212772, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1081": { + "offset": 10212832, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1082": { + "offset": 10212892, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1083": { + "offset": 10213264, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1084": { + "offset": 10213324, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1085": { + "offset": 10213384, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1086": { + "offset": 10213444, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1087": { + "offset": 10213504, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1088": { + "offset": 10213564, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1089": { + "offset": 10213624, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1090": { + "offset": 10213684, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1091": { + "offset": 10213744, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1092": { + "offset": 10213804, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1093": { + "offset": 10213864, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1094": { + "offset": 10213924, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1095": { + "offset": 10214296, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1096": { + "offset": 10214356, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1097": { + "offset": 10214416, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1098": { + "offset": 10214476, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1099": { + "offset": 10214536, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1100": { + "offset": 10214596, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1101": { + "offset": 10214656, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1102": { + "offset": 10214716, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1103": { + "offset": 10214776, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1104": { + "offset": 10214836, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1105": { + "offset": 10214896, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1106": { + "offset": 10214956, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1107": { + "offset": 10215328, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1108": { + "offset": 10215388, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1109": { + "offset": 10215448, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1110": { + "offset": 10215508, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1111": { + "offset": 10215568, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1112": { + "offset": 10215628, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1113": { + "offset": 10215688, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1114": { + "offset": 10215748, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1115": { + "offset": 10215808, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1116": { + "offset": 10215868, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1117": { + "offset": 10215928, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1118": { + "offset": 10215988, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1119": { + "offset": 10216360, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1120": { + "offset": 10216420, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1121": { + "offset": 10216480, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1122": { + "offset": 10216540, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1123": { + "offset": 10216600, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1124": { + "offset": 10216660, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1125": { + "offset": 10216720, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1126": { + "offset": 10216780, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1127": { + "offset": 10216840, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1128": { + "offset": 10216900, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1129": { + "offset": 10216960, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1130": { + "offset": 10217020, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1131": { + "offset": 10217392, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1132": { + "offset": 10217452, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1133": { + "offset": 10217512, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1134": { + "offset": 10217572, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1135": { + "offset": 10217632, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1136": { + "offset": 10217692, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1137": { + "offset": 10217752, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1138": { + "offset": 10217812, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1139": { + "offset": 10217872, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1140": { + "offset": 10217932, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1141": { + "offset": 10217992, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1142": { + "offset": 10218052, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1143": { + "offset": 10218424, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1144": { + "offset": 10218484, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1145": { + "offset": 10218544, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1146": { + "offset": 10218604, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1147": { + "offset": 10218664, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1148": { + "offset": 10218724, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1149": { + "offset": 10218784, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1150": { + "offset": 10218844, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1151": { + "offset": 10218904, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1152": { + "offset": 10218964, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1153": { + "offset": 10219024, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1154": { + "offset": 10219084, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1155": { + "offset": 10219456, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1156": { + "offset": 10219516, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1157": { + "offset": 10219576, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1158": { + "offset": 10219636, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1159": { + "offset": 10219696, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1160": { + "offset": 10219756, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1161": { + "offset": 10219816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1162": { + "offset": 10219876, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1163": { + "offset": 10219936, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1164": { + "offset": 10219996, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1165": { + "offset": 10220056, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1166": { + "offset": 10220116, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1167": { + "offset": 10220488, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1168": { + "offset": 10220548, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1169": { + "offset": 10220608, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1170": { + "offset": 10220668, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1171": { + "offset": 10220728, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1172": { + "offset": 10220788, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1173": { + "offset": 10220848, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1174": { + "offset": 10220908, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1175": { + "offset": 10220968, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1176": { + "offset": 10221028, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1177": { + "offset": 10221088, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1178": { + "offset": 10221148, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1179": { + "offset": 10221520, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1180": { + "offset": 10221580, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1181": { + "offset": 10221640, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1182": { + "offset": 10221700, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1183": { + "offset": 10221760, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1184": { + "offset": 10221820, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1185": { + "offset": 10221880, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1186": { + "offset": 10221940, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1187": { + "offset": 10222000, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1188": { + "offset": 10222060, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1189": { + "offset": 10222120, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1190": { + "offset": 10222180, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1191": { + "offset": 10222552, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1192": { + "offset": 10222612, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1193": { + "offset": 10222672, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1194": { + "offset": 10222732, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1195": { + "offset": 10222792, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1196": { + "offset": 10222852, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1197": { + "offset": 10222912, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1198": { + "offset": 10222972, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1199": { + "offset": 10223032, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1200": { + "offset": 10223092, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1201": { + "offset": 10223152, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1202": { + "offset": 10223212, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1203": { + "offset": 10223584, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1204": { + "offset": 10223644, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1205": { + "offset": 10223704, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1206": { + "offset": 10223764, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1207": { + "offset": 10223824, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1208": { + "offset": 10223884, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1209": { + "offset": 10223944, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1210": { + "offset": 10224004, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1211": { + "offset": 10224064, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1212": { + "offset": 10224124, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1213": { + "offset": 10224184, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1214": { + "offset": 10224244, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1215": { + "offset": 10224616, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1216": { + "offset": 10224676, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1217": { + "offset": 10224736, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1218": { + "offset": 10224796, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1219": { + "offset": 10224856, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1220": { + "offset": 10224916, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1221": { + "offset": 10224976, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1222": { + "offset": 10225036, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1223": { + "offset": 10225096, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1224": { + "offset": 10225156, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1225": { + "offset": 10225216, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1226": { + "offset": 10225276, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1227": { + "offset": 10225648, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1228": { + "offset": 10225708, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1229": { + "offset": 10225768, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1230": { + "offset": 10225828, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1231": { + "offset": 10225888, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1232": { + "offset": 10225948, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1233": { + "offset": 10226008, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1234": { + "offset": 10226068, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1235": { + "offset": 10226128, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1236": { + "offset": 10226188, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1237": { + "offset": 10226248, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1238": { + "offset": 10226308, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1239": { + "offset": 10226680, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1240": { + "offset": 10226740, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1241": { + "offset": 10226800, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1242": { + "offset": 10226860, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1243": { + "offset": 10226920, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1244": { + "offset": 10226980, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1245": { + "offset": 10227040, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1246": { + "offset": 10227100, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1247": { + "offset": 10227160, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1248": { + "offset": 10227220, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1249": { + "offset": 10227280, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1250": { + "offset": 10227340, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1251": { + "offset": 10227712, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1252": { + "offset": 10227772, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1253": { + "offset": 10227832, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1254": { + "offset": 10227892, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1255": { + "offset": 10227952, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1256": { + "offset": 10228012, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1257": { + "offset": 10228072, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1258": { + "offset": 10228132, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1259": { + "offset": 10228192, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1260": { + "offset": 10228252, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1261": { + "offset": 10228312, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1262": { + "offset": 10228372, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1263": { + "offset": 10228744, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1264": { + "offset": 10228804, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1265": { + "offset": 10228864, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1266": { + "offset": 10228924, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1267": { + "offset": 10228984, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1268": { + "offset": 10229044, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1269": { + "offset": 10229104, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1270": { + "offset": 10229164, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1271": { + "offset": 10229224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1272": { + "offset": 10229284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1273": { + "offset": 10229344, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1274": { + "offset": 10229404, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1275": { + "offset": 10229776, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1276": { + "offset": 10229836, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1277": { + "offset": 10229896, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1278": { + "offset": 10229956, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1279": { + "offset": 10230016, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1280": { + "offset": 10230076, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1281": { + "offset": 10230136, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1282": { + "offset": 10230196, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1283": { + "offset": 10230256, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1284": { + "offset": 10230316, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1285": { + "offset": 10230376, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1286": { + "offset": 10230436, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1287": { + "offset": 10230808, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1288": { + "offset": 10230868, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1289": { + "offset": 10230928, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1290": { + "offset": 10230988, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1291": { + "offset": 10231048, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1292": { + "offset": 10231108, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292" + }, + "patch1293": { + "offset": 10231168, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1294": { + "offset": 10231228, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1295": { + "offset": 10231288, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1296": { + "offset": 10231348, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1297": { + "offset": 10231408, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1298": { + "offset": 10231468, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1299": { + "offset": 10232032, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1300": { + "offset": 10232092, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1301": { + "offset": 10232152, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1302": { + "offset": 10232212, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1303": { + "offset": 10232272, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1304": { + "offset": 10232332, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1305": { + "offset": 10232392, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1306": { + "offset": 10232452, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1307": { + "offset": 10232512, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_296_wts_ddr" + }, + "patch1308": { + "offset": 10232572, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_296_wts_ddr" + }, + "patch1309": { + "offset": 10245088, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1310": { + "offset": 10245148, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1311": { + "offset": 10246816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1312": { + "offset": 10246876, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1313": { + "offset": 10246936, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1314": { + "offset": 10246996, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1315": { + "offset": 10247056, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1316": { + "offset": 10247116, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1317": { + "offset": 10247176, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1318": { + "offset": 10247236, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1319": { + "offset": 10250104, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1320": { + "offset": 10250164, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1321": { + "offset": 10250224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1322": { + "offset": 10250284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1323": { + "offset": 10250344, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1324": { + "offset": 10250404, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1325": { + "offset": 10250464, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1326": { + "offset": 10250524, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294" + }, + "patch1327": { + "offset": 10253560, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1328": { + "offset": 10253620, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1329": { + "offset": 10253680, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1330": { + "offset": 10253740, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1331": { + "offset": 10253800, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1332": { + "offset": 10253860, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1333": { + "offset": 10253920, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_297_wts_ddr" + }, + "patch1334": { + "offset": 10253980, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.Layer_297_wts_ddr" + }, + "patch1335": { + "offset": 10265896, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_296_spill" + }, + "patch1336": { + "offset": 10265956, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_296_spill" + }, + "patch1337": { + "offset": 10267696, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1338": { + "offset": 10267756, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1339": { + "offset": 10267816, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1340": { + "offset": 10267876, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1341": { + "offset": 10267936, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1342": { + "offset": 10267996, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1343": { + "offset": 10270240, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1344": { + "offset": 10270300, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1345": { + "offset": 10270360, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1346": { + "offset": 10270420, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1347": { + "offset": 10270480, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1348": { + "offset": 10270540, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_295_spill" + }, + "patch1349": { + "offset": 10272880, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_296_spill" + }, + "patch1350": { + "offset": 10272940, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_296_spill" + }, + "patch1351": { + "offset": 10273504, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_297_spill" + }, + "patch1352": { + "offset": 10273564, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_297_spill" + }, + "patch1353": { + "offset": 10388104, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_297_spill" + }, + "patch1354": { + "offset": 10388164, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_297_spill" + }, + "patch1355": { + "offset": 10388224, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_298_spill" + }, + "patch1356": { + "offset": 10388284, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_298_spill" + }, + "patch1357": { + "offset": 10398424, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_4" + }, + "patch1358": { + "offset": 10398484, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_4" + }, + "patch1359": { + "offset": 10398544, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_298_spill" + }, + "patch1360": { + "offset": 10398604, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_298_spill" + }, + "patch1361": { + "offset": 10435768, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_297_spill" + }, + "patch1362": { + "offset": 10435828, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_297_spill" + }, + "patch1363": { + "offset": 10435888, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_300_spill" + }, + "patch1364": { + "offset": 10435948, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_300_spill" + }, + "patch1365": { + "offset": 10448056, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_300_spill" + }, + "patch1366": { + "offset": 10448104, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_5_spill" + }, + "patch1367": { + "offset": 10448164, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_300_spill" + }, + "patch1368": { + "offset": 10448212, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_5_spill" + }, + "patch1369": { + "offset": 10458736, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_301_spill" + }, + "patch1370": { + "offset": 10458796, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_301_spill" + }, + "patch1371": { + "offset": 10469440, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_5" + }, + "patch1372": { + "offset": 10469500, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.ofm_ddr_5" + }, + "patch1373": { + "offset": 10469560, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_301_spill" + }, + "patch1374": { + "offset": 10469620, + "size": 6, + "operation": "read_add_write", + "name": "compute_graph.l2l3_301_spill" + } + } +} \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/external_buffer_id.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/external_buffer_id.json new file mode 100644 index 0000000000000000000000000000000000000000..fda991fb500dd6e8fe60ad71d7a2d6d24551a4cd --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/external_buffer_id.json @@ -0,0 +1,8752 @@ +{ + "external_buffers": { + "buffer0": { + "xrt_id": 2, + "size_in_bytes": 69157168, + "name": "coalesed_spills", + "coalesed_buffers": [ + { + "logical_id": 134, + "offset_in_bytes": 21896304, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113", + "control_packet_patch_locations": [ + { + "offset": 3552912, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3552972, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4169040, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4169100, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 141, + "offset_in_bytes": 23279664, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123", + "control_packet_patch_locations": [ + { + "offset": 4191288, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4191348, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4955640, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4955700, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 142, + "offset_in_bytes": 23333424, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133", + "control_packet_patch_locations": [ + { + "offset": 4282848, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4282908, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4934616, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4934676, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 149, + "offset_in_bytes": 25270128, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153", + "control_packet_patch_locations": [ + { + "offset": 5058744, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5058804, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5675256, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5675316, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 156, + "offset_in_bytes": 27206832, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163", + "control_packet_patch_locations": [ + { + "offset": 5698464, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5698524, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6467952, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6468012, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 101, + "offset_in_bytes": 13818880, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818", + "control_packet_patch_locations": [ + { + "offset": 904188, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 904248, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 937680, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 937740, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 168, + "offset_in_bytes": 31893552, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183", + "control_packet_patch_locations": [ + { + "offset": 6480312, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6480372, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7193992, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7194052, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 196, + "offset_in_bytes": 38884528, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222", + "control_packet_patch_locations": [ + { + "offset": 7915800, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7915860, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7956792, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7956852, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 106, + "offset_in_bytes": 16076800, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424", + "control_packet_patch_locations": [ + { + "offset": 979608, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 979668, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1598412, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1598472, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 213, + "offset_in_bytes": 40766768, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245", + "control_packet_patch_locations": [ + { + "offset": 8730384, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8730444, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8773968, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8774028, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 229, + "offset_in_bytes": 44561968, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265", + "control_packet_patch_locations": [ + { + "offset": 9579264, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9579324, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9645360, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9645420, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 113, + "offset_in_bytes": 16871824, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434", + "control_packet_patch_locations": [ + { + "offset": 1621956, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1622016, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2312268, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2312328, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 114, + "offset_in_bytes": 16945424, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636", + "control_packet_patch_locations": [ + { + "offset": 1644012, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1644072, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2297052, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2297112, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 121, + "offset_in_bytes": 18270464, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646", + "control_packet_patch_locations": [ + { + "offset": 2323920, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2323980, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3014232, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3014292, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 122, + "offset_in_bytes": 18344064, + "name": "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848", + "control_packet_patch_locations": [ + { + "offset": 2345712, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2345772, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2999016, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2999076, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 92, + "offset_in_bytes": 8258560, + "name": "compute_graph.l2l3_10_spill", + "control_packet_patch_locations": [ + { + "offset": 506596, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 506656, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 521572, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 521632, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 135, + "offset_in_bytes": 22126704, + "name": "compute_graph.l2l3_113_spill", + "control_packet_patch_locations": [ + { + "offset": 3553176, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3553236, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3553536, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 136, + "offset_in_bytes": 22357104, + "name": "compute_graph.l2l3_114_spill", + "control_packet_patch_locations": [ + { + "offset": 3553596, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3802416, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 93, + "offset_in_bytes": 8719360, + "name": "compute_graph.l2l3_11_spill", + "control_packet_patch_locations": [ + { + "offset": 539404, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 539464, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 539980, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 540088, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 137, + "offset_in_bytes": 22587504, + "name": "compute_graph.l2l3_120_spill", + "control_packet_patch_locations": [ + { + "offset": 4167168, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4167408, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 140, + "offset_in_bytes": 23049264, + "name": "compute_graph.l2l3_121_spill", + "control_packet_patch_locations": [ + { + "offset": 4168524, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4169352, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4169412, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 94, + "offset_in_bytes": 9180160, + "name": "compute_graph.l2l3_12_spill", + "control_packet_patch_locations": [ + { + "offset": 552292, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 552352, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 807300, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 807360, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 836316, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 836424, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 143, + "offset_in_bytes": 23655984, + "name": "compute_graph.l2l3_133_spill", + "control_packet_patch_locations": [ + { + "offset": 4282584, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4282644, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4283208, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 144, + "offset_in_bytes": 23978544, + "name": "compute_graph.l2l3_134_spill", + "control_packet_patch_locations": [ + { + "offset": 4283268, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4532136, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 95, + "offset_in_bytes": 9640960, + "name": "compute_graph.l2l3_13_spill", + "control_packet_patch_locations": [ + { + "offset": 835836, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 835896, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 836268, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 836376, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 145, + "offset_in_bytes": 24301104, + "name": "compute_graph.l2l3_140_spill", + "control_packet_patch_locations": [ + { + "offset": 4932744, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4932984, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 148, + "offset_in_bytes": 24947568, + "name": "compute_graph.l2l3_141_spill", + "control_packet_patch_locations": [ + { + "offset": 4934100, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4934928, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4934988, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 96, + "offset_in_bytes": 10101760, + "name": "compute_graph.l2l3_14_spill", + "control_packet_patch_locations": [ + { + "offset": 848628, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 848736, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 855276, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 855336, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 150, + "offset_in_bytes": 25592688, + "name": "compute_graph.l2l3_153_spill", + "control_packet_patch_locations": [ + { + "offset": 5059008, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5059068, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5059368, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 151, + "offset_in_bytes": 25915248, + "name": "compute_graph.l2l3_154_spill", + "control_packet_patch_locations": [ + { + "offset": 5059428, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5308296, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 97, + "offset_in_bytes": 10562560, + "name": "compute_graph.l2l3_15_spill", + "control_packet_patch_locations": [ + { + "offset": 855948, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 856008, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 872604, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 872664, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 872724, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 872784, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 872844, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 872904, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 887460, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 887520, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 887580, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 887640, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 887700, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 887760, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 152, + "offset_in_bytes": 26237808, + "name": "compute_graph.l2l3_160_spill", + "control_packet_patch_locations": [ + { + "offset": 5673384, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5673624, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 155, + "offset_in_bytes": 26884272, + "name": "compute_graph.l2l3_161_spill", + "control_packet_patch_locations": [ + { + "offset": 5674740, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5675568, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5675628, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 157, + "offset_in_bytes": 27283632, + "name": "compute_graph.l2l3_164_spill", + "control_packet_patch_locations": [ + { + "offset": 5708160, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5708220, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5734752, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5734860, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 158, + "offset_in_bytes": 27744432, + "name": "compute_graph.l2l3_167_spill", + "control_packet_patch_locations": [ + { + "offset": 5734320, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5734380, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5734800, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5734908, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 159, + "offset_in_bytes": 28205232, + "name": "compute_graph.l2l3_168_spill", + "control_packet_patch_locations": [ + { + "offset": 5746344, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5746404, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5747592, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5747652, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 160, + "offset_in_bytes": 28666032, + "name": "compute_graph.l2l3_169_spill", + "control_packet_patch_locations": [ + { + "offset": 5757072, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5757132, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5783616, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5783724, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 98, + "offset_in_bytes": 12405760, + "name": "compute_graph.l2l3_16_spill", + "control_packet_patch_locations": [ + { + "offset": 884700, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 884760, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 891492, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 161, + "offset_in_bytes": 29126832, + "name": "compute_graph.l2l3_172_spill", + "control_packet_patch_locations": [ + { + "offset": 5783232, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5783292, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5783664, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5783772, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 162, + "offset_in_bytes": 29587632, + "name": "compute_graph.l2l3_173_spill", + "control_packet_patch_locations": [ + { + "offset": 5793048, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5793108, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5794368, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6445440, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6445500, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 163, + "offset_in_bytes": 30048432, + "name": "compute_graph.l2l3_174_spill", + "control_packet_patch_locations": [ + { + "offset": 5794428, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6043248, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 100, + "offset_in_bytes": 13358080, + "name": "compute_graph.l2l3_17_spill", + "control_packet_patch_locations": [ + { + "offset": 892224, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 892620, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 892680, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 164, + "offset_in_bytes": 30509232, + "name": "compute_graph.l2l3_180_spill", + "control_packet_patch_locations": [ + { + "offset": 6443568, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6443808, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 167, + "offset_in_bytes": 31432752, + "name": "compute_graph.l2l3_181_spill", + "control_packet_patch_locations": [ + { + "offset": 6444924, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6445752, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6445812, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 169, + "offset_in_bytes": 31970352, + "name": "compute_graph.l2l3_184_spill", + "control_packet_patch_locations": [ + { + "offset": 6490008, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6490068, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6516600, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6516708, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 170, + "offset_in_bytes": 32431152, + "name": "compute_graph.l2l3_187_spill", + "control_packet_patch_locations": [ + { + "offset": 6516168, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6516228, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6516648, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6516756, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 171, + "offset_in_bytes": 32891952, + "name": "compute_graph.l2l3_188_spill", + "control_packet_patch_locations": [ + { + "offset": 6528192, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6528252, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6529440, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6529500, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 172, + "offset_in_bytes": 33352752, + "name": "compute_graph.l2l3_189_spill", + "control_packet_patch_locations": [ + { + "offset": 6538920, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6538980, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6565464, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6565572, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 173, + "offset_in_bytes": 33813552, + "name": "compute_graph.l2l3_192_spill", + "control_packet_patch_locations": [ + { + "offset": 6565080, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6565140, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6565512, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6565620, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 174, + "offset_in_bytes": 34274352, + "name": "compute_graph.l2l3_193_spill", + "control_packet_patch_locations": [ + { + "offset": 6574896, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6574956, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6576216, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7171480, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7171540, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 175, + "offset_in_bytes": 34735152, + "name": "compute_graph.l2l3_194_spill", + "control_packet_patch_locations": [ + { + "offset": 6576276, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6825096, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 102, + "offset_in_bytes": 13991680, + "name": "compute_graph.l2l3_19_spill", + "control_packet_patch_locations": [ + { + "offset": 918300, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 918360, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 918780, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 918840, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 918900, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 79, + "offset_in_bytes": 0, + "name": "compute_graph.l2l3_1_spill", + "control_packet_patch_locations": [ + { + "offset": 415036, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 415096, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 415396, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 415456, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 176, + "offset_in_bytes": 35195952, + "name": "compute_graph.l2l3_200_spill", + "control_packet_patch_locations": [ + { + "offset": 7169608, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7169848, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 179, + "offset_in_bytes": 36119472, + "name": "compute_graph.l2l3_201_spill", + "control_packet_patch_locations": [ + { + "offset": 7170964, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7171792, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7171852, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 180, + "offset_in_bytes": 36580272, + "name": "compute_graph.l2l3_204_spill", + "control_packet_patch_locations": [ + { + "offset": 7215784, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7215844, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7242328, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7242436, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 181, + "offset_in_bytes": 37041072, + "name": "compute_graph.l2l3_207_spill", + "control_packet_patch_locations": [ + { + "offset": 7241944, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7242004, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7242376, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7242484, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 182, + "offset_in_bytes": 37501872, + "name": "compute_graph.l2l3_208_spill", + "control_packet_patch_locations": [ + { + "offset": 7251760, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7251820, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7253080, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7859064, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7859124, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 183, + "offset_in_bytes": 37962672, + "name": "compute_graph.l2l3_209_spill", + "control_packet_patch_locations": [ + { + "offset": 7253140, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7501960, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 103, + "offset_in_bytes": 14510080, + "name": "compute_graph.l2l3_20_spill", + "control_packet_patch_locations": [ + { + "offset": 918660, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 918720, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 936552, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 184, + "offset_in_bytes": 38423472, + "name": "compute_graph.l2l3_212_spill", + "control_packet_patch_locations": [ + { + "offset": 7856592, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7856832, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 187, + "offset_in_bytes": 38546608, + "name": "compute_graph.l2l3_213_spill", + "control_packet_patch_locations": [ + { + "offset": 7857948, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7858536, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7858596, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 188, + "offset_in_bytes": 38608048, + "name": "compute_graph.l2l3_214_spill", + "control_packet_patch_locations": [ + { + "offset": 7869672, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7869732, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7869984, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7870044, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7870584, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7870644, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 189, + "offset_in_bytes": 38669488, + "name": "compute_graph.l2l3_215_spill", + "control_packet_patch_locations": [ + { + "offset": 7870272, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7870332, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7872888, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7872948, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 190, + "offset_in_bytes": 38700208, + "name": "compute_graph.l2l3_216_spill", + "control_packet_patch_locations": [ + { + "offset": 7870872, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7870932, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7873464, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7873524, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7874592, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7874652, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 193, + "offset_in_bytes": 38792368, + "name": "compute_graph.l2l3_219_spill", + "control_packet_patch_locations": [ + { + "offset": 7893840, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7893900, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7894152, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7894212, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7916112, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7916172, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 105, + "offset_in_bytes": 15558400, + "name": "compute_graph.l2l3_21_spill", + "control_packet_patch_locations": [ + { + "offset": 937284, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 938232, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 938292, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 194, + "offset_in_bytes": 38853808, + "name": "compute_graph.l2l3_220_spill", + "control_packet_patch_locations": [ + { + "offset": 7894440, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7894500, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7894752, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7894812, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7946400, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7946460, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 197, + "offset_in_bytes": 38915248, + "name": "compute_graph.l2l3_223_spill", + "control_packet_patch_locations": [ + { + "offset": 7916400, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7916460, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7916712, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7916772, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 201, + "offset_in_bytes": 39068848, + "name": "compute_graph.l2l3_231_spill", + "control_packet_patch_locations": [ + { + "offset": 8312824, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8312884, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8313160, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8313220, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 202, + "offset_in_bytes": 39314608, + "name": "compute_graph.l2l3_232_spill", + "control_packet_patch_locations": [ + { + "offset": 8313448, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8313508, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8387416, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8387476, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 203, + "offset_in_bytes": 39550128, + "name": "compute_graph.l2l3_233_spill", + "control_packet_patch_locations": [ + { + "offset": 8324008, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8324116, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8359288, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8359348, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 205, + "offset_in_bytes": 40104368, + "name": "compute_graph.l2l3_237_spill", + "control_packet_patch_locations": [ + { + "offset": 8399008, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8399068, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8399320, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8399380, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8399920, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8399980, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 206, + "offset_in_bytes": 40251568, + "name": "compute_graph.l2l3_238_spill", + "control_packet_patch_locations": [ + { + "offset": 8399608, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8399668, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8783760, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8783820, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 207, + "offset_in_bytes": 40325168, + "name": "compute_graph.l2l3_239_spill", + "control_packet_patch_locations": [ + { + "offset": 8400208, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8400268, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8400520, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8400580, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8401096, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8401156, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 210, + "offset_in_bytes": 40545968, + "name": "compute_graph.l2l3_242_spill", + "control_packet_patch_locations": [ + { + "offset": 8706888, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8706948, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8707200, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8707260, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8730696, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8730756, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 211, + "offset_in_bytes": 40693168, + "name": "compute_graph.l2l3_243_spill", + "control_packet_patch_locations": [ + { + "offset": 8707488, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8707548, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8707800, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8707860, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8762808, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8762868, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 214, + "offset_in_bytes": 40840368, + "name": "compute_graph.l2l3_246_spill", + "control_packet_patch_locations": [ + { + "offset": 8730984, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8731044, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8731296, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8731356, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 107, + "offset_in_bytes": 16209280, + "name": "compute_graph.l2l3_24_spill", + "control_packet_patch_locations": [ + { + "offset": 979872, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 979932, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 980232, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 218, + "offset_in_bytes": 41208368, + "name": "compute_graph.l2l3_254_spill", + "control_packet_patch_locations": [ + { + "offset": 9124600, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9124660, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9137224, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9137284, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 219, + "offset_in_bytes": 41797168, + "name": "compute_graph.l2l3_255_spill", + "control_packet_patch_locations": [ + { + "offset": 9137344, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9137404, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9138136, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9138196, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 221, + "offset_in_bytes": 43179568, + "name": "compute_graph.l2l3_257_spill", + "control_packet_patch_locations": [ + { + "offset": 9147256, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9147316, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9147688, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9147748, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9156808, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9156868, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 222, + "offset_in_bytes": 43467568, + "name": "compute_graph.l2l3_258_spill", + "control_packet_patch_locations": [ + { + "offset": 9147808, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9147868, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9655896, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 223, + "offset_in_bytes": 43640368, + "name": "compute_graph.l2l3_259_spill", + "control_packet_patch_locations": [ + { + "offset": 9156928, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9156988, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9168016, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9602112, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 108, + "offset_in_bytes": 16341760, + "name": "compute_graph.l2l3_25_spill", + "control_packet_patch_locations": [ + { + "offset": 980292, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 980352, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1229172, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 225, + "offset_in_bytes": 43813168, + "name": "compute_graph.l2l3_260_spill", + "control_packet_patch_locations": [ + { + "offset": 9168076, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9180040, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9180100, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 226, + "offset_in_bytes": 44101168, + "name": "compute_graph.l2l3_262_spill", + "control_packet_patch_locations": [ + { + "offset": 9548112, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9548172, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9548520, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9548580, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9579672, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9579732, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 227, + "offset_in_bytes": 44389168, + "name": "compute_graph.l2l3_263_spill", + "control_packet_patch_locations": [ + { + "offset": 9548640, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9548700, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9560112, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9560172, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9634584, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9634644, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 230, + "offset_in_bytes": 44734768, + "name": "compute_graph.l2l3_266_spill", + "control_packet_patch_locations": [ + { + "offset": 9579792, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9579852, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9591264, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9591324, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 231, + "offset_in_bytes": 44907568, + "name": "compute_graph.l2l3_267_spill", + "control_packet_patch_locations": [ + { + "offset": 9601584, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9601644, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9602232, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 232, + "offset_in_bytes": 45080368, + "name": "compute_graph.l2l3_268_spill", + "control_packet_patch_locations": [ + { + "offset": 9602172, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9614160, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9614220, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 234, + "offset_in_bytes": 45368368, + "name": "compute_graph.l2l3_273_spill", + "control_packet_patch_locations": [ + { + "offset": 9655956, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10010032, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10010092, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10010152, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10010212, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10010272, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10010332, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10010392, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10010452, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023136, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023196, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023256, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023316, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023376, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023436, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023496, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023556, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10024504, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10024564, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10024624, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10024684, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10024744, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10024804, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10024864, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10024924, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10025872, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10025932, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10025992, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026052, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026112, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026172, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026232, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026292, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027240, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027300, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027360, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027420, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027480, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027540, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027600, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027660, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10028608, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10028668, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10028728, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10028788, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10028848, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10028908, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10028968, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10029028, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10029976, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030036, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030096, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030156, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030216, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030276, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030336, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030396, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031344, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031404, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031464, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031524, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031584, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031644, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031704, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031764, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10032712, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10032772, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10032832, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10032892, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10032952, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033012, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033072, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033132, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034080, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034140, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034200, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034260, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034320, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034380, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034440, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034500, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10035448, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10035508, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10035568, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10035628, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10035688, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10035748, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10035808, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10035868, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10036816, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10036876, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10036936, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10036996, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037056, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037116, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037176, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037236, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 236, + "offset_in_bytes": 47499568, + "name": "compute_graph.l2l3_276_spill", + "control_packet_patch_locations": [ + { + "offset": 10050088, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10050148, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10054624, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10054684, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10055824, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10055884, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 237, + "offset_in_bytes": 48421168, + "name": "compute_graph.l2l3_277_spill", + "control_packet_patch_locations": [ + { + "offset": 10054744, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10054804, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10195192, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10195252, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 238, + "offset_in_bytes": 48881968, + "name": "compute_graph.l2l3_278_spill", + "control_packet_patch_locations": [ + { + "offset": 10055944, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10056004, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10056952, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10057012, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10057528, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10057588, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 241, + "offset_in_bytes": 50264368, + "name": "compute_graph.l2l3_280_spill", + "control_packet_patch_locations": [ + { + "offset": 10059136, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10059196, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10073320, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10073380, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 242, + "offset_in_bytes": 51185968, + "name": "compute_graph.l2l3_281_spill", + "control_packet_patch_locations": [ + { + "offset": 10073200, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10073260, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10084816, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10084876, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10116544, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10116604, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 243, + "offset_in_bytes": 52107568, + "name": "compute_graph.l2l3_282_spill", + "control_packet_patch_locations": [ + { + "offset": 10084936, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10084996, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10086112, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10086220, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10166272, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10166380, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 245, + "offset_in_bytes": 52568368, + "name": "compute_graph.l2l3_283_spill", + "control_packet_patch_locations": [ + { + "offset": 10098208, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10098268, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10101376, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10101484, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 246, + "offset_in_bytes": 53029168, + "name": "compute_graph.l2l3_284_spill", + "control_packet_patch_locations": [ + { + "offset": 10113520, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10113580, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10181584, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10181692, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 247, + "offset_in_bytes": 53489968, + "name": "compute_graph.l2l3_285_spill", + "control_packet_patch_locations": [ + { + "offset": 10116664, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10116724, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10117864, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10117972, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 249, + "offset_in_bytes": 54872368, + "name": "compute_graph.l2l3_288_spill", + "control_packet_patch_locations": [ + { + "offset": 10133992, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10134052, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10147984, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10148044, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 250, + "offset_in_bytes": 55333168, + "name": "compute_graph.l2l3_289_spill", + "control_packet_patch_locations": [ + { + "offset": 10165816, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10165876, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10166320, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10166428, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 251, + "offset_in_bytes": 55793968, + "name": "compute_graph.l2l3_290_spill", + "control_packet_patch_locations": [ + { + "offset": 10178416, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10178476, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10181632, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10181740, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 255, + "offset_in_bytes": 61784368, + "name": "compute_graph.l2l3_295_spill", + "control_packet_patch_locations": [ + { + "offset": 10245088, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10245148, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10253560, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10253620, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10253680, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10253740, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10253800, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10253860, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10267696, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10267756, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10267816, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10267876, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10267936, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10267996, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10270240, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10270300, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10270360, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10270420, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10270480, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10270540, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 256, + "offset_in_bytes": 63627568, + "name": "compute_graph.l2l3_296_spill", + "control_packet_patch_locations": [ + { + "offset": 10265896, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10265956, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10272880, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10272940, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 257, + "offset_in_bytes": 65470768, + "name": "compute_graph.l2l3_297_spill", + "control_packet_patch_locations": [ + { + "offset": 10273504, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10273564, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10388104, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10388164, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10435768, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10435828, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 258, + "offset_in_bytes": 66392368, + "name": "compute_graph.l2l3_298_spill", + "control_packet_patch_locations": [ + { + "offset": 10388224, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10388284, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10398544, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10398604, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 80, + "offset_in_bytes": 471040, + "name": "compute_graph.l2l3_2_spill", + "control_packet_patch_locations": [ + { + "offset": 415516, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 415576, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 416212, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 260, + "offset_in_bytes": 67313968, + "name": "compute_graph.l2l3_300_spill", + "control_packet_patch_locations": [ + { + "offset": 10435888, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10435948, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10448056, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10448164, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 261, + "offset_in_bytes": 68235568, + "name": "compute_graph.l2l3_301_spill", + "control_packet_patch_locations": [ + { + "offset": 10458736, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10458796, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10469560, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10469620, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 109, + "offset_in_bytes": 16474240, + "name": "compute_graph.l2l3_31_spill", + "control_packet_patch_locations": [ + { + "offset": 1596516, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1596756, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 112, + "offset_in_bytes": 16739344, + "name": "compute_graph.l2l3_32_spill", + "control_packet_patch_locations": [ + { + "offset": 1597872, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1598724, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1598784, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 115, + "offset_in_bytes": 17166224, + "name": "compute_graph.l2l3_36_spill", + "control_packet_patch_locations": [ + { + "offset": 1643748, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1643808, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1644372, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 116, + "offset_in_bytes": 17387024, + "name": "compute_graph.l2l3_37_spill", + "control_packet_patch_locations": [ + { + "offset": 1644432, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1644480, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1893300, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 82, + "offset_in_bytes": 1766400, + "name": "compute_graph.l2l3_3_spill", + "control_packet_patch_locations": [ + { + "offset": 416944, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 417364, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 117, + "offset_in_bytes": 17607824, + "name": "compute_graph.l2l3_43_spill", + "control_packet_patch_locations": [ + { + "offset": 2295156, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2295396, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 120, + "offset_in_bytes": 18049664, + "name": "compute_graph.l2l3_44_spill", + "control_packet_patch_locations": [ + { + "offset": 2296512, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2297364, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2297424, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 123, + "offset_in_bytes": 18564864, + "name": "compute_graph.l2l3_48_spill", + "control_packet_patch_locations": [ + { + "offset": 2345976, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2346036, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2346336, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 124, + "offset_in_bytes": 18785664, + "name": "compute_graph.l2l3_49_spill", + "control_packet_patch_locations": [ + { + "offset": 2346396, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2346444, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2595264, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 83, + "offset_in_bytes": 2708480, + "name": "compute_graph.l2l3_4_spill", + "control_packet_patch_locations": [ + { + "offset": 417424, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 428236, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 125, + "offset_in_bytes": 19006464, + "name": "compute_graph.l2l3_55_spill", + "control_packet_patch_locations": [ + { + "offset": 2997120, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2997360, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 128, + "offset_in_bytes": 19448304, + "name": "compute_graph.l2l3_56_spill", + "control_packet_patch_locations": [ + { + "offset": 2998476, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2999328, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2999388, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 129, + "offset_in_bytes": 19669104, + "name": "compute_graph.l2l3_59_spill", + "control_packet_patch_locations": [ + { + "offset": 3037452, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3037512, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3065724, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3065832, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3065940, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 85, + "offset_in_bytes": 4572160, + "name": "compute_graph.l2l3_5_spill", + "control_packet_patch_locations": [ + { + "offset": 428944, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 429412, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 429520, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7871160, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7871220, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8313808, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8313868, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10448104, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10448212, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 130, + "offset_in_bytes": 20110704, + "name": "compute_graph.l2l3_62_spill", + "control_packet_patch_locations": [ + { + "offset": 3065340, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3065400, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3065772, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3065880, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3065988, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 131, + "offset_in_bytes": 20552304, + "name": "compute_graph.l2l3_63_spill", + "control_packet_patch_locations": [ + { + "offset": 3077376, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3077436, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3082296, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 133, + "offset_in_bytes": 21454704, + "name": "compute_graph.l2l3_64_spill", + "control_packet_patch_locations": [ + { + "offset": 3083028, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3083424, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3083484, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 87, + "offset_in_bytes": 5493760, + "name": "compute_graph.l2l3_6_spill", + "control_packet_patch_locations": [ + { + "offset": 442012, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 442072, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 452812, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 452920, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 89, + "offset_in_bytes": 6415360, + "name": "compute_graph.l2l3_7_spill", + "control_packet_patch_locations": [ + { + "offset": 465652, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 465712, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 476284, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 476344, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 476404, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 476464, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 476524, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 476584, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 90, + "offset_in_bytes": 7336960, + "name": "compute_graph.l2l3_8_spill", + "control_packet_patch_locations": [ + { + "offset": 477052, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 477112, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 490324, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 490384, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 539932, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 540040, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 91, + "offset_in_bytes": 7797760, + "name": "compute_graph.l2l3_9_spill", + "control_packet_patch_locations": [ + { + "offset": 506236, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 506296, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 506716, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 506776, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 138, + "offset_in_bytes": 22588464, + "name": "compute_graph.l2l3_scratch_0_121_spill", + "control_packet_patch_locations": [ + { + "offset": 4167468, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4168032, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 146, + "offset_in_bytes": 24302448, + "name": "compute_graph.l2l3_scratch_0_141_spill", + "control_packet_patch_locations": [ + { + "offset": 4933044, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4933608, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 154, + "offset_in_bytes": 26561712, + "name": "compute_graph.l2l3_scratch_0_161_spill", + "control_packet_patch_locations": [ + { + "offset": 5673684, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5674248, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 99, + "offset_in_bytes": 12897280, + "name": "compute_graph.l2l3_scratch_0_17_spill", + "control_packet_patch_locations": [ + { + "offset": 891552, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 892164, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 165, + "offset_in_bytes": 30511152, + "name": "compute_graph.l2l3_scratch_0_181_spill", + "control_packet_patch_locations": [ + { + "offset": 6443868, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6444432, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 177, + "offset_in_bytes": 35197872, + "name": "compute_graph.l2l3_scratch_0_201_spill", + "control_packet_patch_locations": [ + { + "offset": 7169908, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7170472, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 185, + "offset_in_bytes": 38423728, + "name": "compute_graph.l2l3_scratch_0_213_spill", + "control_packet_patch_locations": [ + { + "offset": 7856892, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7857456, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 104, + "offset_in_bytes": 15040000, + "name": "compute_graph.l2l3_scratch_0_21_spill", + "control_packet_patch_locations": [ + { + "offset": 936612, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 937224, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 110, + "offset_in_bytes": 16474384, + "name": "compute_graph.l2l3_scratch_0_32_spill", + "control_packet_patch_locations": [ + { + "offset": 1596816, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1597380, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 81, + "offset_in_bytes": 824320, + "name": "compute_graph.l2l3_scratch_0_3_spill", + "control_packet_patch_locations": [ + { + "offset": 416272, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 416884, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 118, + "offset_in_bytes": 17608064, + "name": "compute_graph.l2l3_scratch_0_44_spill", + "control_packet_patch_locations": [ + { + "offset": 2295456, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2296020, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 126, + "offset_in_bytes": 19006704, + "name": "compute_graph.l2l3_scratch_0_56_spill", + "control_packet_patch_locations": [ + { + "offset": 2997420, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2997984, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 84, + "offset_in_bytes": 3650560, + "name": "compute_graph.l2l3_scratch_0_5_spill", + "control_packet_patch_locations": [ + { + "offset": 428296, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 428884, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 132, + "offset_in_bytes": 21013104, + "name": "compute_graph.l2l3_scratch_0_64_spill", + "control_packet_patch_locations": [ + { + "offset": 3082356, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3082968, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 139, + "offset_in_bytes": 22818864, + "name": "compute_graph.l2l3_scratch_1_121_spill", + "control_packet_patch_locations": [ + { + "offset": 4168200, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4168464, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 147, + "offset_in_bytes": 24625008, + "name": "compute_graph.l2l3_scratch_1_141_spill", + "control_packet_patch_locations": [ + { + "offset": 4933776, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4934040, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 153, + "offset_in_bytes": 26239152, + "name": "compute_graph.l2l3_scratch_1_161_spill", + "control_packet_patch_locations": [ + { + "offset": 5674416, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5674680, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 166, + "offset_in_bytes": 30971952, + "name": "compute_graph.l2l3_scratch_1_181_spill", + "control_packet_patch_locations": [ + { + "offset": 6444600, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6444864, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 178, + "offset_in_bytes": 35658672, + "name": "compute_graph.l2l3_scratch_1_201_spill", + "control_packet_patch_locations": [ + { + "offset": 7170640, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7170904, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 186, + "offset_in_bytes": 38485168, + "name": "compute_graph.l2l3_scratch_1_213_spill", + "control_packet_patch_locations": [ + { + "offset": 7857624, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7857888, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 111, + "offset_in_bytes": 16606864, + "name": "compute_graph.l2l3_scratch_1_32_spill", + "control_packet_patch_locations": [ + { + "offset": 1597548, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1597812, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 119, + "offset_in_bytes": 17828864, + "name": "compute_graph.l2l3_scratch_1_44_spill", + "control_packet_patch_locations": [ + { + "offset": 2296188, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2296452, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 127, + "offset_in_bytes": 19227504, + "name": "compute_graph.l2l3_scratch_1_56_spill", + "control_packet_patch_locations": [ + { + "offset": 2998152, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2998416, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 192, + "offset_in_bytes": 38730928, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_217", + "control_packet_patch_locations": [ + { + "offset": 7874280, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7874340, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7874856, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7874916, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7875192, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7875252, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 198, + "offset_in_bytes": 38945968, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_225", + "control_packet_patch_locations": [ + { + "offset": 7873728, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7873788, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7927656, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7927716, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7927944, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7928004, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 200, + "offset_in_bytes": 39007408, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_230", + "control_packet_patch_locations": [ + { + "offset": 7873152, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7873212, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7965240, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7965300, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8304544, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8304604, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 204, + "offset_in_bytes": 39780528, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_236", + "control_packet_patch_locations": [ + { + "offset": 3025836, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3025896, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8387104, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8387164, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8387680, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8387740, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8388016, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8388076, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 209, + "offset_in_bytes": 40398768, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_240", + "control_packet_patch_locations": [ + { + "offset": 7872576, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7872636, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8401360, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8401420, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8401696, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8401756, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 215, + "offset_in_bytes": 40913968, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_248", + "control_packet_patch_locations": [ + { + "offset": 8400784, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8400844, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8743008, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8743068, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8743296, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8743356, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 217, + "offset_in_bytes": 41061168, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_253", + "control_packet_patch_locations": [ + { + "offset": 8783448, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8783508, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8784024, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8784084, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9124480, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9124540, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 220, + "offset_in_bytes": 42373168, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_256", + "control_packet_patch_locations": [ + { + "offset": 951720, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 951780, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8375224, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8375284, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9138424, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9138484, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9138784, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9138844, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 235, + "offset_in_bytes": 45656368, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_275", + "control_packet_patch_locations": [ + { + "offset": 848676, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 848784, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8324056, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8324164, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10022248, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10022308, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10022368, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10022428, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10022488, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10022548, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10022608, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10022668, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023616, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023676, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023736, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023796, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023856, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023916, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10023976, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10024036, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10024984, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10025044, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10025104, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10025164, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10025224, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10025284, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10025344, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10025404, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026352, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026412, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026472, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026532, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026592, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026652, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026712, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10026772, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027720, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027780, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027840, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027900, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10027960, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10028020, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10028080, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10028140, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10029088, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10029148, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10029208, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10029268, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10029328, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10029388, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10029448, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10029508, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030456, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030516, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030576, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030636, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030696, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030756, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030816, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10030876, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031824, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031884, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10031944, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10032004, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10032064, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10032124, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10032184, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10032244, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033192, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033252, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033312, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033372, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033432, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033492, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033552, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10033612, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034560, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034620, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034680, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034740, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034800, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034860, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034920, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10034980, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10035928, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10035988, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10036048, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10036108, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10036168, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10036228, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10036288, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10036348, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037296, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037356, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037416, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037476, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037536, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037596, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037656, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10037716, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10038328, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10038388, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10050640, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10050700, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10051600, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10051660, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10052560, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10052620, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10053520, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10053580, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 240, + "offset_in_bytes": 49342768, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_279", + "control_packet_patch_locations": [ + { + "offset": 7872024, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7872084, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10057792, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10057852, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10058152, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10058212, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10058272, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10058332, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10058392, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10058452, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 248, + "offset_in_bytes": 53950768, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_287", + "control_packet_patch_locations": [ + { + "offset": 10057216, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10057276, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10130224, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10130284, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10133224, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10133284, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10133344, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10133404, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10133464, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10133524, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 253, + "offset_in_bytes": 56254768, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_292", + "control_packet_patch_locations": [ + { + "offset": 10191856, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10191964, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10195456, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10195516, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10195864, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10195924, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10195984, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10196044, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10196104, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10196164, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209136, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209196, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209256, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209316, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209376, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209436, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210168, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210228, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210288, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210348, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210408, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210468, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211200, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211260, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211320, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211380, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211440, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211500, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212232, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212292, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212352, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212412, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212472, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212532, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213264, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213324, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213384, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213444, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213504, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213564, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214296, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214356, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214416, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214476, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214536, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214596, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215328, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215388, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215448, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215508, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215568, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215628, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216360, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216420, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216480, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216540, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216600, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216660, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217392, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217452, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217512, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217572, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217632, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217692, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218424, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218484, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218544, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218604, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218664, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218724, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219456, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219516, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219576, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219636, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219696, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219756, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220488, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220548, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220608, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220668, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220728, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220788, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221520, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221580, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221640, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221700, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221760, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221820, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222552, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222612, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222672, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222732, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222792, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222852, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223584, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223644, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223704, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223764, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223824, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223884, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224616, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224676, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224736, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224796, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224856, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224916, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225648, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225708, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225768, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225828, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225888, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225948, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226680, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226740, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226800, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226860, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226920, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226980, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227712, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227772, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227832, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227892, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227952, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228012, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228744, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228804, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228864, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228924, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228984, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229044, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229776, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229836, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229896, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229956, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230016, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230076, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230808, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230868, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230928, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230988, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10231048, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10231108, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 254, + "offset_in_bytes": 57176368, + "name": "compute_graph.spill_L3_Concat_Buffer_layer_294", + "control_packet_patch_locations": [ + { + "offset": 7871448, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7871508, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10208464, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10208524, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10208584, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10208644, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10208704, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10208764, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209496, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209556, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209616, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209676, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209736, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10209796, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210528, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210588, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210648, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210708, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210768, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10210828, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211560, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211620, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211680, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211740, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211800, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10211860, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212592, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212652, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212712, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212772, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212832, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10212892, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213624, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213684, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213744, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213804, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213864, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10213924, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214656, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214716, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214776, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214836, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214896, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10214956, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215688, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215748, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215808, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215868, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215928, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10215988, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216720, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216780, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216840, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216900, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10216960, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217020, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217752, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217812, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217872, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217932, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10217992, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218052, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218784, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218844, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218904, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10218964, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219024, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219084, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219816, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219876, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219936, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10219996, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220056, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220116, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220848, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220908, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10220968, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221028, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221088, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221148, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221880, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10221940, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222000, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222060, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222120, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222180, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222912, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10222972, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223032, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223092, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223152, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223212, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10223944, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224004, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224064, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224124, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224184, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224244, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10224976, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225036, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225096, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225156, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225216, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10225276, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226008, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226068, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226128, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226188, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226248, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10226308, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227040, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227100, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227160, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227220, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227280, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10227340, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228072, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228132, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228192, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228252, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228312, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10228372, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229104, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229164, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229224, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229284, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229344, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10229404, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230136, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230196, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230256, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230316, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230376, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10230436, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10231168, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10231228, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10231288, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10231348, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10231408, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10231468, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10232032, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10232092, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10232152, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10232212, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10232272, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10232332, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10232392, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10232452, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10246816, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10246876, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10246936, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10246996, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10247056, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10247116, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10247176, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10247236, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10250104, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10250164, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10250224, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10250284, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10250344, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10250404, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10250464, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10250524, + "size": 6, + "operation": "read_add_write" + } + ] + } + ] + }, + "buffer1": { + "xrt_id": 0, + "size_in_bytes": 11809920, + "name": "coalesed_weights", + "coalesed_buffers": [ + { + "logical_id": 35, + "offset_in_bytes": 570880, + "name": "compute_graph.Layer_103_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3437184, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3437244, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 36, + "offset_in_bytes": 609280, + "name": "compute_graph.Layer_104_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3451848, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3451908, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 37, + "offset_in_bytes": 689920, + "name": "compute_graph.Layer_109_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3498048, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3498108, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 38, + "offset_in_bytes": 705280, + "name": "compute_graph.Layer_116_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3802608, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3802668, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 39, + "offset_in_bytes": 832256, + "name": "compute_graph.Layer_117_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 4122504, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4122564, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 40, + "offset_in_bytes": 959232, + "name": "compute_graph.Layer_123_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 4169664, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4169724, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 41, + "offset_in_bytes": 1087232, + "name": "compute_graph.Layer_124_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 4180872, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4180932, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 42, + "offset_in_bytes": 1270016, + "name": "compute_graph.Layer_129_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 4227720, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4227780, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 43, + "offset_in_bytes": 1291520, + "name": "compute_graph.Layer_136_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 4532328, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4532388, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 44, + "offset_in_bytes": 1560320, + "name": "compute_graph.Layer_137_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 4887984, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4888044, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 1, + "offset_in_bytes": 4864, + "name": "compute_graph.Layer_13_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 540148, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 540208, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 45, + "offset_in_bytes": 1802496, + "name": "compute_graph.Layer_143_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 4935240, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4935300, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 46, + "offset_in_bytes": 1981696, + "name": "compute_graph.Layer_144_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 4955952, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 4956012, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 47, + "offset_in_bytes": 2164480, + "name": "compute_graph.Layer_149_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 5002920, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5002980, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 2, + "offset_in_bytes": 5888, + "name": "compute_graph.Layer_14_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 807612, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 807672, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 48, + "offset_in_bytes": 2315008, + "name": "compute_graph.Layer_156_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 5308488, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5308548, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 49, + "offset_in_bytes": 2583808, + "name": "compute_graph.Layer_157_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 5628624, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5628684, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 3, + "offset_in_bytes": 10240, + "name": "compute_graph.Layer_15_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 836484, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 836544, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 50, + "offset_in_bytes": 2825984, + "name": "compute_graph.Layer_163_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 5675880, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5675940, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 51, + "offset_in_bytes": 3067904, + "name": "compute_graph.Layer_164_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 5698176, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5698236, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 52, + "offset_in_bytes": 3390464, + "name": "compute_graph.Layer_169_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 5734968, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 5735028, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 4, + "offset_in_bytes": 18944, + "name": "compute_graph.Layer_16_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 856068, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 856128, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 53, + "offset_in_bytes": 3605504, + "name": "compute_graph.Layer_176_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 6043440, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6043500, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 54, + "offset_in_bytes": 4107264, + "name": "compute_graph.Layer_177_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 6408984, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6409044, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 55, + "offset_in_bytes": 4591104, + "name": "compute_graph.Layer_183_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 6446064, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6446124, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 56, + "offset_in_bytes": 4992000, + "name": "compute_graph.Layer_184_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 6480024, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6480084, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 57, + "offset_in_bytes": 5314560, + "name": "compute_graph.Layer_189_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 6516816, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6516876, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 5, + "offset_in_bytes": 20992, + "name": "compute_graph.Layer_18_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 892740, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 892800, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 58, + "offset_in_bytes": 5529600, + "name": "compute_graph.Layer_196_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 6825288, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 6825348, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 59, + "offset_in_bytes": 6031360, + "name": "compute_graph.Layer_197_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 7135024, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7135084, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 6, + "offset_in_bytes": 29696, + "name": "compute_graph.Layer_19_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 893172, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 893232, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 60, + "offset_in_bytes": 6515200, + "name": "compute_graph.Layer_203_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 7172104, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7172164, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 61, + "offset_in_bytes": 6916096, + "name": "compute_graph.Layer_204_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 7206064, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7206124, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 7, + "offset_in_bytes": 42752, + "name": "compute_graph.Layer_20_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 904308, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 904368, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 62, + "offset_in_bytes": 7238656, + "name": "compute_graph.Layer_211_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 7502152, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7502212, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 63, + "offset_in_bytes": 7489536, + "name": "compute_graph.Layer_214_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 7858656, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7858716, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 64, + "offset_in_bytes": 7745536, + "name": "compute_graph.Layer_218_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 7875312, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7875372, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 65, + "offset_in_bytes": 8048640, + "name": "compute_graph.Layer_226_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 7917312, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7917372, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 8, + "offset_in_bytes": 45824, + "name": "compute_graph.Layer_22_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 937800, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 937860, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 66, + "offset_in_bytes": 8200192, + "name": "compute_graph.Layer_237_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 8388136, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8388196, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 9, + "offset_in_bytes": 55552, + "name": "compute_graph.Layer_23_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 938544, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 938604, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 67, + "offset_in_bytes": 8616960, + "name": "compute_graph.Layer_241_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 8401816, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8401876, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 68, + "offset_in_bytes": 8806400, + "name": "compute_graph.Layer_249_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 8731896, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8731956, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 10, + "offset_in_bytes": 68608, + "name": "compute_graph.Layer_24_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 951840, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 951900, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 69, + "offset_in_bytes": 8901120, + "name": "compute_graph.Layer_257_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 9138904, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9138964, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 70, + "offset_in_bytes": 9033728, + "name": "compute_graph.Layer_261_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 9168136, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9168196, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 71, + "offset_in_bytes": 9090560, + "name": "compute_graph.Layer_269_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 9614280, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9614340, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 72, + "offset_in_bytes": 9114880, + "name": "compute_graph.Layer_276_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 10010512, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10010572, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 11, + "offset_in_bytes": 77056, + "name": "compute_graph.Layer_27_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 1229364, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1229424, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 73, + "offset_in_bytes": 9152768, + "name": "compute_graph.Layer_280_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 10058512, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10058572, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 74, + "offset_in_bytes": 9171712, + "name": "compute_graph.Layer_288_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 10118080, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10118140, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 12, + "offset_in_bytes": 86784, + "name": "compute_graph.Layer_28_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 1551756, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1551816, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 75, + "offset_in_bytes": 9190656, + "name": "compute_graph.Layer_295_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 10196224, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10196284, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 76, + "offset_in_bytes": 9214976, + "name": "compute_graph.Layer_296_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 10232512, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10232572, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 77, + "offset_in_bytes": 9224448, + "name": "compute_graph.Layer_297_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 10253920, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10253980, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 13, + "offset_in_bytes": 104192, + "name": "compute_graph.Layer_34_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 1599036, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1599096, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 14, + "offset_in_bytes": 113920, + "name": "compute_graph.Layer_35_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 1611492, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1611552, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 15, + "offset_in_bytes": 131328, + "name": "compute_graph.Layer_36_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 1622076, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1622136, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 16, + "offset_in_bytes": 142592, + "name": "compute_graph.Layer_39_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 1893492, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 1893552, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 17, + "offset_in_bytes": 158464, + "name": "compute_graph.Layer_40_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 2250492, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2250552, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 18, + "offset_in_bytes": 175872, + "name": "compute_graph.Layer_46_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 2297676, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2297736, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 19, + "offset_in_bytes": 193280, + "name": "compute_graph.Layer_47_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 2312580, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2312640, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 20, + "offset_in_bytes": 210688, + "name": "compute_graph.Layer_48_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 2324040, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2324100, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 21, + "offset_in_bytes": 221952, + "name": "compute_graph.Layer_51_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 2595456, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2595516, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 22, + "offset_in_bytes": 237824, + "name": "compute_graph.Layer_52_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 2952456, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2952516, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 23, + "offset_in_bytes": 255232, + "name": "compute_graph.Layer_58_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 2999640, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 2999700, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 24, + "offset_in_bytes": 272640, + "name": "compute_graph.Layer_59_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3014544, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3014604, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 25, + "offset_in_bytes": 307456, + "name": "compute_graph.Layer_65_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3083544, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3083604, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 26, + "offset_in_bytes": 315648, + "name": "compute_graph.Layer_70_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3119976, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3120036, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 27, + "offset_in_bytes": 364032, + "name": "compute_graph.Layer_71_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3131184, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3131244, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 28, + "offset_in_bytes": 407040, + "name": "compute_graph.Layer_76_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3177144, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3177204, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 29, + "offset_in_bytes": 414208, + "name": "compute_graph.Layer_81_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3223824, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3223884, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 30, + "offset_in_bytes": 455680, + "name": "compute_graph.Layer_82_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3238488, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3238548, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 31, + "offset_in_bytes": 487936, + "name": "compute_graph.Layer_87_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3284400, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3284460, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 0, + "offset_in_bytes": 0, + "name": "compute_graph.Layer_8_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 453028, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 453088, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 32, + "offset_in_bytes": 494080, + "name": "compute_graph.Layer_92_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3330504, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3330564, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 33, + "offset_in_bytes": 532480, + "name": "compute_graph.Layer_93_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3345168, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3345228, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 34, + "offset_in_bytes": 564736, + "name": "compute_graph.Layer_98_wts_ddr", + "control_packet_patch_locations": [ + { + "offset": 3391080, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 3391140, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 244, + "offset_in_bytes": 11349120, + "name": "compute_graph.const_ifm_ddr", + "control_packet_patch_locations": [ + { + "offset": 10086064, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10086172, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 228, + "offset_in_bytes": 11176320, + "name": "compute_graph.const_ifm_ddr_1", + "control_packet_patch_locations": [ + { + "offset": 9560400, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9560460, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 212, + "offset_in_bytes": 11102720, + "name": "compute_graph.const_ifm_ddr_2", + "control_packet_patch_locations": [ + { + "offset": 8708088, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8708148, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 195, + "offset_in_bytes": 11072000, + "name": "compute_graph.const_ifm_ddr_3", + "control_packet_patch_locations": [ + { + "offset": 7895040, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7895100, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 88, + "offset_in_bytes": 10150400, + "name": "compute_graph.const_ifm_ddr_4", + "control_packet_patch_locations": [ + { + "offset": 452860, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 452968, + "size": 6, + "operation": "read_add_write" + } + ] + }, + { + "logical_id": 86, + "offset_in_bytes": 9228800, + "name": "compute_graph.const_ifm_ddr_5", + "control_packet_patch_locations": [ + { + "offset": 429460, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 429568, + "size": 6, + "operation": "read_add_write" + } + ] + } + ] + }, + "buffer2": { + "xrt_id": 1, + "logical_id": 78, + "size_in_bytes": 471040, + "name": "compute_graph.ifm_ddr", + "control_packet_patch_locations": [ + { + "offset": 405436, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 405496, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer3": { + "xrt_id": 10, + "logical_id": 239, + "size_in_bytes": 460800, + "name": "compute_graph.ifm_ddr_1", + "control_packet_patch_locations": [ + { + "offset": 7871760, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7871820, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10101424, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10101532, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10117912, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10118020, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer4": { + "xrt_id": 8, + "logical_id": 224, + "size_in_bytes": 172800, + "name": "compute_graph.ifm_ddr_2", + "control_packet_patch_locations": [ + { + "offset": 9157048, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9571200, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9571260, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9591576, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9591636, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer5": { + "xrt_id": 6, + "logical_id": 208, + "size_in_bytes": 73600, + "name": "compute_graph.ifm_ddr_3", + "control_packet_patch_locations": [ + { + "offset": 7872312, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7872372, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8719248, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8719308, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8731584, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8731644, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer6": { + "xrt_id": 4, + "logical_id": 191, + "size_in_bytes": 30720, + "name": "compute_graph.ifm_ddr_4", + "control_packet_patch_locations": [ + { + "offset": 7874016, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7874076, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7905432, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7905492, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7917000, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7917060, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer7": { + "xrt_id": 5, + "logical_id": 199, + "size_in_bytes": 30720, + "name": "compute_graph.ofm_ddr_0_l2l3_229_spill", + "control_packet_patch_locations": [ + { + "offset": 7965504, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 7965564, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer8": { + "xrt_id": 7, + "logical_id": 216, + "size_in_bytes": 73600, + "name": "compute_graph.ofm_ddr_1_l2l3_252_spill", + "control_packet_patch_locations": [ + { + "offset": 8783184, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 8783244, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer9": { + "xrt_id": 9, + "logical_id": 233, + "size_in_bytes": 172800, + "name": "compute_graph.ofm_ddr_2_l2l3_272_spill", + "control_packet_patch_locations": [ + { + "offset": 9655368, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9655428, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 9656016, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer10": { + "xrt_id": 11, + "logical_id": 252, + "size_in_bytes": 460800, + "name": "compute_graph.ofm_ddr_3_l2l3_291_spill", + "control_packet_patch_locations": [ + { + "offset": 10191808, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10191916, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer11": { + "xrt_id": 12, + "logical_id": 259, + "size_in_bytes": 921600, + "name": "compute_graph.ofm_ddr_4", + "control_packet_patch_locations": [ + { + "offset": 10398424, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10398484, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer12": { + "xrt_id": 13, + "logical_id": 262, + "size_in_bytes": 921600, + "name": "compute_graph.ofm_ddr_5", + "control_packet_patch_locations": [ + { + "offset": 10469440, + "size": 6, + "operation": "read_add_write" + }, + { + "offset": 10469500, + "size": 6, + "operation": "read_add_write" + } + ] + }, + "buffer13": { + "xrt_id": 3, + "logical_id": 4294967295, + "size_in_bytes": 10499908, + "ctrl_pkt_buffer": true, + "name": "runtime_control_packet" + } + } +} \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/flexmlrt-hsi.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/flexmlrt-hsi.json new file mode 100644 index 0000000000000000000000000000000000000000..3f869465582054c5842c70d0707eded21657b0ac --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/flexmlrt-hsi.json @@ -0,0 +1,1253 @@ +{ + "inputs" : [ + { + "name" : "compute_graph.ifm_ddr", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 180, + 320, + 4 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 320, + 23, + 4, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + }, + { + "name" : "compute_graph.ifm_ddr_1", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 16, + 90, + 160 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 90, + 2, + 160, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + }, + { + "name" : "compute_graph.ifm_ddr_2", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 20, + 45, + 80 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 45, + 3, + 80, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + }, + { + "name" : "compute_graph.ifm_ddr_3", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 40, + 23, + 40 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 23, + 5, + 40, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + }, + { + "name" : "compute_graph.ifm_ddr_4", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 64, + 12, + 20 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 12, + 8, + 20, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + } + ], + "outputs" : [ + { + "name" : "compute_graph.ofm_ddr_3_l2l3_291_spill", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 16, + 90, + 160 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 90, + 2, + 160, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + }, + { + "name" : "compute_graph.ofm_ddr_2_l2l3_272_spill", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 20, + 45, + 80 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 45, + 3, + 80, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + }, + { + "name" : "compute_graph.ofm_ddr_1_l2l3_252_spill", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 40, + 23, + 40 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 23, + 5, + 40, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + }, + { + "name" : "compute_graph.ofm_ddr_0_l2l3_229_spill", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 64, + 12, + 20 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 12, + 8, + 20, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + }, + { + "name" : "compute_graph.ofm_ddr_5", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 3, + 180, + 320 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 180, + 1, + 320, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + }, + { + "name" : "compute_graph.ofm_ddr_4", + "scale_factor" : -1, + "cpu_shape" : [ + 1, + 1, + 180, + 320 + ], + "cpu_format" : "NCHW", + "cpu_dtype" : "fp32", + "hw_shape" : [ + 180, + 1, + 320, + 1, + 8 + ], + "hw_format" : "HCWNC8", + "hw_dtype" : "bf16" + } + ], + "weights" : { + "layers" : [ + { + "name" : "compute_graph.Layer_8_wts_ddr", + "offset" : 0 + }, + { + "name" : "compute_graph.Layer_13_wts_ddr", + "offset" : 1216 + }, + { + "name" : "compute_graph.Layer_14_wts_ddr", + "offset" : 1472 + }, + { + "name" : "compute_graph.Layer_15_wts_ddr", + "offset" : 2560 + }, + { + "name" : "compute_graph.Layer_16_wts_ddr", + "offset" : 4736 + }, + { + "name" : "compute_graph.Layer_18_wts_ddr", + "offset" : 5248 + }, + { + "name" : "compute_graph.Layer_19_wts_ddr", + "offset" : 7424 + }, + { + "name" : "compute_graph.Layer_20_wts_ddr", + "offset" : 10688 + }, + { + "name" : "compute_graph.Layer_22_wts_ddr", + "offset" : 11456 + }, + { + "name" : "compute_graph.Layer_23_wts_ddr", + "offset" : 13888 + }, + { + "name" : "compute_graph.Layer_24_wts_ddr", + "offset" : 17152 + }, + { + "name" : "compute_graph.Layer_27_wts_ddr", + "offset" : 19264 + }, + { + "name" : "compute_graph.Layer_28_wts_ddr", + "offset" : 21696 + }, + { + "name" : "compute_graph.Layer_34_wts_ddr", + "offset" : 26048 + }, + { + "name" : "compute_graph.Layer_35_wts_ddr", + "offset" : 28480 + }, + { + "name" : "compute_graph.Layer_36_wts_ddr", + "offset" : 32832 + }, + { + "name" : "compute_graph.Layer_39_wts_ddr", + "offset" : 35648 + }, + { + "name" : "compute_graph.Layer_40_wts_ddr", + "offset" : 39616 + }, + { + "name" : "compute_graph.Layer_46_wts_ddr", + "offset" : 43968 + }, + { + "name" : "compute_graph.Layer_47_wts_ddr", + "offset" : 48320 + }, + { + "name" : "compute_graph.Layer_48_wts_ddr", + "offset" : 52672 + }, + { + "name" : "compute_graph.Layer_51_wts_ddr", + "offset" : 55488 + }, + { + "name" : "compute_graph.Layer_52_wts_ddr", + "offset" : 59456 + }, + { + "name" : "compute_graph.Layer_58_wts_ddr", + "offset" : 63808 + }, + { + "name" : "compute_graph.Layer_59_wts_ddr", + "offset" : 68160 + }, + { + "name" : "compute_graph.Layer_65_wts_ddr", + "offset" : 76864 + }, + { + "name" : "compute_graph.Layer_70_wts_ddr", + "offset" : 78912 + }, + { + "name" : "compute_graph.Layer_71_wts_ddr", + "offset" : 91008 + }, + { + "name" : "compute_graph.Layer_76_wts_ddr", + "offset" : 101760 + }, + { + "name" : "compute_graph.Layer_81_wts_ddr", + "offset" : 103552 + }, + { + "name" : "compute_graph.Layer_82_wts_ddr", + "offset" : 113920 + }, + { + "name" : "compute_graph.Layer_87_wts_ddr", + "offset" : 121984 + }, + { + "name" : "compute_graph.Layer_92_wts_ddr", + "offset" : 123520 + }, + { + "name" : "compute_graph.Layer_93_wts_ddr", + "offset" : 133120 + }, + { + "name" : "compute_graph.Layer_98_wts_ddr", + "offset" : 141184 + }, + { + "name" : "compute_graph.Layer_103_wts_ddr", + "offset" : 142720 + }, + { + "name" : "compute_graph.Layer_104_wts_ddr", + "offset" : 152320 + }, + { + "name" : "compute_graph.Layer_109_wts_ddr", + "offset" : 172480 + }, + { + "name" : "compute_graph.Layer_116_wts_ddr", + "offset" : 176320 + }, + { + "name" : "compute_graph.Layer_117_wts_ddr", + "offset" : 208064 + }, + { + "name" : "compute_graph.Layer_123_wts_ddr", + "offset" : 239808 + }, + { + "name" : "compute_graph.Layer_124_wts_ddr", + "offset" : 271808 + }, + { + "name" : "compute_graph.Layer_129_wts_ddr", + "offset" : 317504 + }, + { + "name" : "compute_graph.Layer_136_wts_ddr", + "offset" : 322880 + }, + { + "name" : "compute_graph.Layer_137_wts_ddr", + "offset" : 390080 + }, + { + "name" : "compute_graph.Layer_143_wts_ddr", + "offset" : 450624 + }, + { + "name" : "compute_graph.Layer_144_wts_ddr", + "offset" : 495424 + }, + { + "name" : "compute_graph.Layer_149_wts_ddr", + "offset" : 541120 + }, + { + "name" : "compute_graph.Layer_156_wts_ddr", + "offset" : 578752 + }, + { + "name" : "compute_graph.Layer_157_wts_ddr", + "offset" : 645952 + }, + { + "name" : "compute_graph.Layer_163_wts_ddr", + "offset" : 706496 + }, + { + "name" : "compute_graph.Layer_164_wts_ddr", + "offset" : 766976 + }, + { + "name" : "compute_graph.Layer_169_wts_ddr", + "offset" : 847616 + }, + { + "name" : "compute_graph.Layer_176_wts_ddr", + "offset" : 901376 + }, + { + "name" : "compute_graph.Layer_177_wts_ddr", + "offset" : 1026816 + }, + { + "name" : "compute_graph.Layer_183_wts_ddr", + "offset" : 1147776 + }, + { + "name" : "compute_graph.Layer_184_wts_ddr", + "offset" : 1248000 + }, + { + "name" : "compute_graph.Layer_189_wts_ddr", + "offset" : 1328640 + }, + { + "name" : "compute_graph.Layer_196_wts_ddr", + "offset" : 1382400 + }, + { + "name" : "compute_graph.Layer_197_wts_ddr", + "offset" : 1507840 + }, + { + "name" : "compute_graph.Layer_203_wts_ddr", + "offset" : 1628800 + }, + { + "name" : "compute_graph.Layer_204_wts_ddr", + "offset" : 1729024 + }, + { + "name" : "compute_graph.Layer_211_wts_ddr", + "offset" : 1809664 + }, + { + "name" : "compute_graph.Layer_214_wts_ddr", + "offset" : 1872384 + }, + { + "name" : "compute_graph.Layer_218_wts_ddr", + "offset" : 1936384 + }, + { + "name" : "compute_graph.Layer_226_wts_ddr", + "offset" : 2012160 + }, + { + "name" : "compute_graph.Layer_237_wts_ddr", + "offset" : 2050048 + }, + { + "name" : "compute_graph.Layer_241_wts_ddr", + "offset" : 2154240 + }, + { + "name" : "compute_graph.Layer_249_wts_ddr", + "offset" : 2201600 + }, + { + "name" : "compute_graph.Layer_257_wts_ddr", + "offset" : 2225280 + }, + { + "name" : "compute_graph.Layer_261_wts_ddr", + "offset" : 2258432 + }, + { + "name" : "compute_graph.Layer_269_wts_ddr", + "offset" : 2272640 + }, + { + "name" : "compute_graph.Layer_276_wts_ddr", + "offset" : 2278720 + }, + { + "name" : "compute_graph.Layer_280_wts_ddr", + "offset" : 2288192 + }, + { + "name" : "compute_graph.Layer_288_wts_ddr", + "offset" : 2292928 + }, + { + "name" : "compute_graph.Layer_295_wts_ddr", + "offset" : 2297664 + }, + { + "name" : "compute_graph.Layer_296_wts_ddr", + "offset" : 2303744 + }, + { + "name" : "compute_graph.Layer_297_wts_ddr", + "offset" : 2306112 + }, + { + "name" : "compute_graph.const_ifm_ddr_5", + "offset" : 2307200 + }, + { + "name" : "compute_graph.const_ifm_ddr_4", + "offset" : 2537600 + }, + { + "name" : "compute_graph.const_ifm_ddr_3", + "offset" : 2768000 + }, + { + "name" : "compute_graph.const_ifm_ddr_2", + "offset" : 2775680 + }, + { + "name" : "compute_graph.const_ifm_ddr_1", + "offset" : 2794080 + }, + { + "name" : "compute_graph.const_ifm_ddr", + "offset" : 2837280 + } + ] + }, + "spills" : { + "layers" : [ + { + "name" : "compute_graph.l2l3_1_spill", + "offset" : 0 + }, + { + "name" : "compute_graph.l2l3_2_spill", + "offset" : 117760 + }, + { + "name" : "compute_graph.l2l3_scratch_0_3_spill", + "offset" : 206080 + }, + { + "name" : "compute_graph.l2l3_3_spill", + "offset" : 441600 + }, + { + "name" : "compute_graph.l2l3_4_spill", + "offset" : 677120 + }, + { + "name" : "compute_graph.l2l3_scratch_0_5_spill", + "offset" : 912640 + }, + { + "name" : "compute_graph.l2l3_5_spill", + "offset" : 1143040 + }, + { + "name" : "compute_graph.l2l3_6_spill", + "offset" : 1373440 + }, + { + "name" : "compute_graph.l2l3_7_spill", + "offset" : 1603840 + }, + { + "name" : "compute_graph.l2l3_8_spill", + "offset" : 1834240 + }, + { + "name" : "compute_graph.l2l3_9_spill", + "offset" : 1949440 + }, + { + "name" : "compute_graph.l2l3_10_spill", + "offset" : 2064640 + }, + { + "name" : "compute_graph.l2l3_11_spill", + "offset" : 2179840 + }, + { + "name" : "compute_graph.l2l3_12_spill", + "offset" : 2295040 + }, + { + "name" : "compute_graph.l2l3_13_spill", + "offset" : 2410240 + }, + { + "name" : "compute_graph.l2l3_14_spill", + "offset" : 2525440 + }, + { + "name" : "compute_graph.l2l3_15_spill", + "offset" : 2640640 + }, + { + "name" : "compute_graph.l2l3_16_spill", + "offset" : 3101440 + }, + { + "name" : "compute_graph.l2l3_scratch_0_17_spill", + "offset" : 3224320 + }, + { + "name" : "compute_graph.l2l3_17_spill", + "offset" : 3339520 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_1818", + "offset" : 3454720 + }, + { + "name" : "compute_graph.l2l3_19_spill", + "offset" : 3497920 + }, + { + "name" : "compute_graph.l2l3_20_spill", + "offset" : 3627520 + }, + { + "name" : "compute_graph.l2l3_scratch_0_21_spill", + "offset" : 3760000 + }, + { + "name" : "compute_graph.l2l3_21_spill", + "offset" : 3889600 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_2424", + "offset" : 4019200 + }, + { + "name" : "compute_graph.l2l3_24_spill", + "offset" : 4052320 + }, + { + "name" : "compute_graph.l2l3_25_spill", + "offset" : 4085440 + }, + { + "name" : "compute_graph.l2l3_31_spill", + "offset" : 4118560 + }, + { + "name" : "compute_graph.l2l3_scratch_0_32_spill", + "offset" : 4118596 + }, + { + "name" : "compute_graph.l2l3_scratch_1_32_spill", + "offset" : 4151716 + }, + { + "name" : "compute_graph.l2l3_32_spill", + "offset" : 4184836 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3434", + "offset" : 4217956 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_3636", + "offset" : 4236356 + }, + { + "name" : "compute_graph.l2l3_36_spill", + "offset" : 4291556 + }, + { + "name" : "compute_graph.l2l3_37_spill", + "offset" : 4346756 + }, + { + "name" : "compute_graph.l2l3_43_spill", + "offset" : 4401956 + }, + { + "name" : "compute_graph.l2l3_scratch_0_44_spill", + "offset" : 4402016 + }, + { + "name" : "compute_graph.l2l3_scratch_1_44_spill", + "offset" : 4457216 + }, + { + "name" : "compute_graph.l2l3_44_spill", + "offset" : 4512416 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4646", + "offset" : 4567616 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_4848", + "offset" : 4586016 + }, + { + "name" : "compute_graph.l2l3_48_spill", + "offset" : 4641216 + }, + { + "name" : "compute_graph.l2l3_49_spill", + "offset" : 4696416 + }, + { + "name" : "compute_graph.l2l3_55_spill", + "offset" : 4751616 + }, + { + "name" : "compute_graph.l2l3_scratch_0_56_spill", + "offset" : 4751676 + }, + { + "name" : "compute_graph.l2l3_scratch_1_56_spill", + "offset" : 4806876 + }, + { + "name" : "compute_graph.l2l3_56_spill", + "offset" : 4862076 + }, + { + "name" : "compute_graph.l2l3_59_spill", + "offset" : 4917276 + }, + { + "name" : "compute_graph.l2l3_62_spill", + "offset" : 5027676 + }, + { + "name" : "compute_graph.l2l3_63_spill", + "offset" : 5138076 + }, + { + "name" : "compute_graph.l2l3_scratch_0_64_spill", + "offset" : 5253276 + }, + { + "name" : "compute_graph.l2l3_64_spill", + "offset" : 5363676 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_113113", + "offset" : 5474076 + }, + { + "name" : "compute_graph.l2l3_113_spill", + "offset" : 5531676 + }, + { + "name" : "compute_graph.l2l3_114_spill", + "offset" : 5589276 + }, + { + "name" : "compute_graph.l2l3_120_spill", + "offset" : 5646876 + }, + { + "name" : "compute_graph.l2l3_scratch_0_121_spill", + "offset" : 5647116 + }, + { + "name" : "compute_graph.l2l3_scratch_1_121_spill", + "offset" : 5704716 + }, + { + "name" : "compute_graph.l2l3_121_spill", + "offset" : 5762316 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_123123", + "offset" : 5819916 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_133133", + "offset" : 5833356 + }, + { + "name" : "compute_graph.l2l3_133_spill", + "offset" : 5913996 + }, + { + "name" : "compute_graph.l2l3_134_spill", + "offset" : 5994636 + }, + { + "name" : "compute_graph.l2l3_140_spill", + "offset" : 6075276 + }, + { + "name" : "compute_graph.l2l3_scratch_0_141_spill", + "offset" : 6075612 + }, + { + "name" : "compute_graph.l2l3_scratch_1_141_spill", + "offset" : 6156252 + }, + { + "name" : "compute_graph.l2l3_141_spill", + "offset" : 6236892 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_153153", + "offset" : 6317532 + }, + { + "name" : "compute_graph.l2l3_153_spill", + "offset" : 6398172 + }, + { + "name" : "compute_graph.l2l3_154_spill", + "offset" : 6478812 + }, + { + "name" : "compute_graph.l2l3_160_spill", + "offset" : 6559452 + }, + { + "name" : "compute_graph.l2l3_scratch_1_161_spill", + "offset" : 6559788 + }, + { + "name" : "compute_graph.l2l3_scratch_0_161_spill", + "offset" : 6640428 + }, + { + "name" : "compute_graph.l2l3_161_spill", + "offset" : 6721068 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_163163", + "offset" : 6801708 + }, + { + "name" : "compute_graph.l2l3_164_spill", + "offset" : 6820908 + }, + { + "name" : "compute_graph.l2l3_167_spill", + "offset" : 6936108 + }, + { + "name" : "compute_graph.l2l3_168_spill", + "offset" : 7051308 + }, + { + "name" : "compute_graph.l2l3_169_spill", + "offset" : 7166508 + }, + { + "name" : "compute_graph.l2l3_172_spill", + "offset" : 7281708 + }, + { + "name" : "compute_graph.l2l3_173_spill", + "offset" : 7396908 + }, + { + "name" : "compute_graph.l2l3_174_spill", + "offset" : 7512108 + }, + { + "name" : "compute_graph.l2l3_180_spill", + "offset" : 7627308 + }, + { + "name" : "compute_graph.l2l3_scratch_0_181_spill", + "offset" : 7627788 + }, + { + "name" : "compute_graph.l2l3_scratch_1_181_spill", + "offset" : 7742988 + }, + { + "name" : "compute_graph.l2l3_181_spill", + "offset" : 7858188 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_183183", + "offset" : 7973388 + }, + { + "name" : "compute_graph.l2l3_184_spill", + "offset" : 7992588 + }, + { + "name" : "compute_graph.l2l3_187_spill", + "offset" : 8107788 + }, + { + "name" : "compute_graph.l2l3_188_spill", + "offset" : 8222988 + }, + { + "name" : "compute_graph.l2l3_189_spill", + "offset" : 8338188 + }, + { + "name" : "compute_graph.l2l3_192_spill", + "offset" : 8453388 + }, + { + "name" : "compute_graph.l2l3_193_spill", + "offset" : 8568588 + }, + { + "name" : "compute_graph.l2l3_194_spill", + "offset" : 8683788 + }, + { + "name" : "compute_graph.l2l3_200_spill", + "offset" : 8798988 + }, + { + "name" : "compute_graph.l2l3_scratch_0_201_spill", + "offset" : 8799468 + }, + { + "name" : "compute_graph.l2l3_scratch_1_201_spill", + "offset" : 8914668 + }, + { + "name" : "compute_graph.l2l3_201_spill", + "offset" : 9029868 + }, + { + "name" : "compute_graph.l2l3_204_spill", + "offset" : 9145068 + }, + { + "name" : "compute_graph.l2l3_207_spill", + "offset" : 9260268 + }, + { + "name" : "compute_graph.l2l3_208_spill", + "offset" : 9375468 + }, + { + "name" : "compute_graph.l2l3_209_spill", + "offset" : 9490668 + }, + { + "name" : "compute_graph.l2l3_212_spill", + "offset" : 9605868 + }, + { + "name" : "compute_graph.l2l3_scratch_0_213_spill", + "offset" : 9605932 + }, + { + "name" : "compute_graph.l2l3_scratch_1_213_spill", + "offset" : 9621292 + }, + { + "name" : "compute_graph.l2l3_213_spill", + "offset" : 9636652 + }, + { + "name" : "compute_graph.l2l3_214_spill", + "offset" : 9652012 + }, + { + "name" : "compute_graph.l2l3_215_spill", + "offset" : 9667372 + }, + { + "name" : "compute_graph.l2l3_216_spill", + "offset" : 9675052 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_217", + "offset" : 9682732 + }, + { + "name" : "compute_graph.l2l3_219_spill", + "offset" : 9698092 + }, + { + "name" : "compute_graph.l2l3_220_spill", + "offset" : 9713452 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_222222", + "offset" : 9721132 + }, + { + "name" : "compute_graph.l2l3_223_spill", + "offset" : 9728812 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_225", + "offset" : 9736492 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_230", + "offset" : 9751852 + }, + { + "name" : "compute_graph.l2l3_231_spill", + "offset" : 9767212 + }, + { + "name" : "compute_graph.l2l3_232_spill", + "offset" : 9828652 + }, + { + "name" : "compute_graph.l2l3_233_spill", + "offset" : 9887532 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_236", + "offset" : 9945132 + }, + { + "name" : "compute_graph.l2l3_237_spill", + "offset" : 10026092 + }, + { + "name" : "compute_graph.l2l3_238_spill", + "offset" : 10062892 + }, + { + "name" : "compute_graph.l2l3_239_spill", + "offset" : 10081292 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_240", + "offset" : 10099692 + }, + { + "name" : "compute_graph.l2l3_242_spill", + "offset" : 10136492 + }, + { + "name" : "compute_graph.l2l3_243_spill", + "offset" : 10173292 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_245245", + "offset" : 10191692 + }, + { + "name" : "compute_graph.l2l3_246_spill", + "offset" : 10210092 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_248", + "offset" : 10228492 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_253", + "offset" : 10265292 + }, + { + "name" : "compute_graph.l2l3_254_spill", + "offset" : 10302092 + }, + { + "name" : "compute_graph.l2l3_255_spill", + "offset" : 10449292 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_256", + "offset" : 10593292 + }, + { + "name" : "compute_graph.l2l3_257_spill", + "offset" : 10794892 + }, + { + "name" : "compute_graph.l2l3_258_spill", + "offset" : 10866892 + }, + { + "name" : "compute_graph.l2l3_259_spill", + "offset" : 10910092 + }, + { + "name" : "compute_graph.l2l3_260_spill", + "offset" : 10953292 + }, + { + "name" : "compute_graph.l2l3_262_spill", + "offset" : 11025292 + }, + { + "name" : "compute_graph.l2l3_263_spill", + "offset" : 11097292 + }, + { + "name" : "compute_graph.L3_OFM_Buffer_layer_TGSpilling_265265", + "offset" : 11140492 + }, + { + "name" : "compute_graph.l2l3_266_spill", + "offset" : 11183692 + }, + { + "name" : "compute_graph.l2l3_267_spill", + "offset" : 11226892 + }, + { + "name" : "compute_graph.l2l3_268_spill", + "offset" : 11270092 + }, + { + "name" : "compute_graph.l2l3_273_spill", + "offset" : 11342092 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_275", + "offset" : 11414092 + }, + { + "name" : "compute_graph.l2l3_276_spill", + "offset" : 11874892 + }, + { + "name" : "compute_graph.l2l3_277_spill", + "offset" : 12105292 + }, + { + "name" : "compute_graph.l2l3_278_spill", + "offset" : 12220492 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_279", + "offset" : 12335692 + }, + { + "name" : "compute_graph.l2l3_280_spill", + "offset" : 12566092 + }, + { + "name" : "compute_graph.l2l3_281_spill", + "offset" : 12796492 + }, + { + "name" : "compute_graph.l2l3_282_spill", + "offset" : 13026892 + }, + { + "name" : "compute_graph.l2l3_283_spill", + "offset" : 13142092 + }, + { + "name" : "compute_graph.l2l3_284_spill", + "offset" : 13257292 + }, + { + "name" : "compute_graph.l2l3_285_spill", + "offset" : 13372492 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_287", + "offset" : 13487692 + }, + { + "name" : "compute_graph.l2l3_288_spill", + "offset" : 13718092 + }, + { + "name" : "compute_graph.l2l3_289_spill", + "offset" : 13833292 + }, + { + "name" : "compute_graph.l2l3_290_spill", + "offset" : 13948492 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_292", + "offset" : 14063692 + }, + { + "name" : "compute_graph.spill_L3_Concat_Buffer_layer_294", + "offset" : 14294092 + }, + { + "name" : "compute_graph.l2l3_295_spill", + "offset" : 15446092 + }, + { + "name" : "compute_graph.l2l3_296_spill", + "offset" : 15906892 + }, + { + "name" : "compute_graph.l2l3_297_spill", + "offset" : 16367692 + }, + { + "name" : "compute_graph.l2l3_298_spill", + "offset" : 16598092 + }, + { + "name" : "compute_graph.l2l3_300_spill", + "offset" : 16828492 + }, + { + "name" : "compute_graph.l2l3_301_spill", + "offset" : 17058892 + } + ] + } +} diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/ml_txn.bin b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/ml_txn.bin new file mode 100644 index 0000000000000000000000000000000000000000..ca7829007db81560493be09fb1d280a92d823d86 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/ml_txn.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e8c431abc3fb04c2333b756f029894d8440645eeb714fb234525160657d575e +size 744144 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/model_top.cpp b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/model_top.cpp new file mode 100644 index 0000000000000000000000000000000000000000..75a9e51705667137de80b63105f6adef7015241a --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/model_top.cpp @@ -0,0 +1,6 @@ +#include "multi_layer_overlay.h" +FlexMLGraph compute_graph; +#if defined(__AIESIM__) || defined(__X86SIM__) +#include "../aie_runtime_control.cpp" +int main() { return 0; } +#endif diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/par.mlopslib.tosa.mlir b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/par.mlopslib.tosa.mlir new file mode 100644 index 0000000000000000000000000000000000000000..9a979c4e508c9ab88cd7b3fdf401508c13c46b38 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/par.mlopslib.tosa.mlir @@ -0,0 +1,24900 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func private @forward(%arg0: tensor<1x180x320x4xbf16> loc(unknown), %arg1: tensor<1x16x90x160xbf16> loc(unknown), %arg2: tensor<1x20x45x80xbf16> loc(unknown), %arg3: tensor<1x40x23x40xbf16> loc(unknown), %arg4: tensor<1x64x12x20xbf16> loc(unknown)) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) { + %0 = xten_nn.subgraph (%arg5 = %arg0: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.906250e-03 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.906250e-03> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc1) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_2", + OutputName = "Div_2", + shift = 0 : i8} : (tensor<1x180x320x4xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %463 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %461 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + %1 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/biases"} -> tensor<4xbf16> loc(#loc) + %2 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/weights"} -> tensor<4x16x1x1xbf16> loc(#loc) + %3 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/biases"} -> tensor<16xbf16> loc(#loc) + %4 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/weights"} -> tensor<16x16x3x3xbf16> loc(#loc) + %5 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/biases"} -> tensor<16xbf16> loc(#loc) + %6 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/weights"} -> tensor<16x35x3x3xbf16> loc(#loc) + %7 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_431/Constant_0_0"} -> tensor<1x16x90x160xbf16> loc(#loc2) + %8 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/biases"} -> tensor<16xbf16> loc(#loc) + %9 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/weights"} -> tensor<16x32x3x3xbf16> loc(#loc) + %10 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/biases"} -> tensor<32xbf16> loc(#loc) + %11 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/weights"} -> tensor<32x32x3x3xbf16> loc(#loc) + %12 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/biases"} -> tensor<32xbf16> loc(#loc) + %13 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/weights"} -> tensor<32x59x3x3xbf16> loc(#loc) + %14 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_411/Constant_0_0"} -> tensor<1x20x45x80xbf16> loc(#loc3) + %15 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/biases"} -> tensor<20xbf16> loc(#loc) + %16 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/weights"} -> tensor<20x40x3x3xbf16> loc(#loc) + %17 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/biases"} -> tensor<40xbf16> loc(#loc) + %18 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/weights"} -> tensor<40x40x3x3xbf16> loc(#loc) + %19 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/biases"} -> tensor<40xbf16> loc(#loc) + %20 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/weights"} -> tensor<40x107x3x3xbf16> loc(#loc) + %21 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_385/Constant_0_0"} -> tensor<1x40x23x40xbf16> loc(#loc4) + %22 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/biases"} -> tensor<40xbf16> loc(#loc) + %23 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/weights"} -> tensor<40x80x3x3xbf16> loc(#loc) + %24 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/biases"} -> tensor<80xbf16> loc(#loc) + %25 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/weights"} -> tensor<80x80x3x3xbf16> loc(#loc) + %26 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/biases"} -> tensor<80xbf16> loc(#loc) + %27 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/weights"} -> tensor<80x171x3x3xbf16> loc(#loc) + %28 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_359/Constant_0_0"} -> tensor<1x64x12x20xbf16> loc(#loc5) + %29 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/biases"} -> tensor<64xbf16> loc(#loc) + %30 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/weights"} -> tensor<64x128x3x3xbf16> loc(#loc) + %31 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/biases"} -> tensor<128xbf16> loc(#loc) + %32 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/weights"} -> tensor<128x128x3x3xbf16> loc(#loc) + %33 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/biases"} -> tensor<128xbf16> loc(#loc) + %34 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %35 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/biases"} -> tensor<128xbf16> loc(#loc) + %36 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %37 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/biases"} -> tensor<960xbf16> loc(#loc) + %38 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %39 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/biases"} -> tensor<160xbf16> loc(#loc) + %40 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %41 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/biases"} -> tensor<960xbf16> loc(#loc) + %42 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %43 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/biases"} -> tensor<240xbf16> loc(#loc) + %44 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %45 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/biases"} -> tensor<960xbf16> loc(#loc) + %46 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %47 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/biases"} -> tensor<960xbf16> loc(#loc) + %48 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %49 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/biases"} -> tensor<160xbf16> loc(#loc) + %50 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %51 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/biases"} -> tensor<960xbf16> loc(#loc) + %52 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %53 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/biases"} -> tensor<240xbf16> loc(#loc) + %54 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %55 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/biases"} -> tensor<960xbf16> loc(#loc) + %56 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %57 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/biases"} -> tensor<960xbf16> loc(#loc) + %58 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %59 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/biases"} -> tensor<160xbf16> loc(#loc) + %60 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/weights"} -> tensor<160x672x1x1xbf16> loc(#loc) + %61 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/biases"} -> tensor<672xbf16> loc(#loc) + %62 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %63 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/biases"} -> tensor<168xbf16> loc(#loc) + %64 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %65 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/biases"} -> tensor<672xbf16> loc(#loc) + %66 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/weights"} -> tensor<672x1x9x9xbf16> loc(#loc) + %67 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/biases"} -> tensor<672xbf16> loc(#loc) + %68 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %69 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/biases"} -> tensor<112xbf16> loc(#loc) + %70 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/weights"} -> tensor<112x672x1x1xbf16> loc(#loc) + %71 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/biases"} -> tensor<672xbf16> loc(#loc) + %72 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %73 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/biases"} -> tensor<168xbf16> loc(#loc) + %74 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %75 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/biases"} -> tensor<672xbf16> loc(#loc) + %76 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/weights"} -> tensor<672x1x3x3xbf16> loc(#loc) + %77 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/biases"} -> tensor<672xbf16> loc(#loc) + %78 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %79 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/biases"} -> tensor<112xbf16> loc(#loc) + %80 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/weights"} -> tensor<112x480x1x1xbf16> loc(#loc) + %81 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/biases"} -> tensor<480xbf16> loc(#loc) + %82 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/weights"} -> tensor<480x120x1x1xbf16> loc(#loc) + %83 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/biases"} -> tensor<120xbf16> loc(#loc) + %84 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/weights"} -> tensor<120x480x1x1xbf16> loc(#loc) + %85 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/biases"} -> tensor<480xbf16> loc(#loc) + %86 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/weights"} -> tensor<480x1x3x3xbf16> loc(#loc) + %87 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/biases"} -> tensor<480xbf16> loc(#loc) + %88 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/weights"} -> tensor<480x80x1x1xbf16> loc(#loc) + %89 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/biases"} -> tensor<80xbf16> loc(#loc) + %90 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %91 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/biases"} -> tensor<184xbf16> loc(#loc) + %92 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %93 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/biases"} -> tensor<184xbf16> loc(#loc) + %94 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %95 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/biases"} -> tensor<80xbf16> loc(#loc) + %96 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %97 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/biases"} -> tensor<184xbf16> loc(#loc) + %98 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %99 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/biases"} -> tensor<184xbf16> loc(#loc) + %100 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %101 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/biases"} -> tensor<80xbf16> loc(#loc) + %102 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/weights"} -> tensor<80x200x1x1xbf16> loc(#loc) + %103 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/biases"} -> tensor<200xbf16> loc(#loc) + %104 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/weights"} -> tensor<200x1x3x3xbf16> loc(#loc) + %105 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/biases"} -> tensor<200xbf16> loc(#loc) + %106 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/weights"} -> tensor<200x80x1x1xbf16> loc(#loc) + %107 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/biases"} -> tensor<80xbf16> loc(#loc) + %108 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/weights"} -> tensor<80x240x1x1xbf16> loc(#loc) + %109 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/biases"} -> tensor<240xbf16> loc(#loc) + %110 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/weights"} -> tensor<240x1x3x3xbf16> loc(#loc) + %111 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/biases"} -> tensor<240xbf16> loc(#loc) + %112 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/weights"} -> tensor<240x40x1x1xbf16> loc(#loc) + %113 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/biases"} -> tensor<40xbf16> loc(#loc) + %114 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %115 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/biases"} -> tensor<120xbf16> loc(#loc) + %116 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %117 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/biases"} -> tensor<32xbf16> loc(#loc) + %118 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %119 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/biases"} -> tensor<120xbf16> loc(#loc) + %120 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %121 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/biases"} -> tensor<120xbf16> loc(#loc) + %122 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %123 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/biases"} -> tensor<40xbf16> loc(#loc) + %124 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %125 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/biases"} -> tensor<120xbf16> loc(#loc) + %126 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %127 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/biases"} -> tensor<32xbf16> loc(#loc) + %128 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %129 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/biases"} -> tensor<120xbf16> loc(#loc) + %130 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %131 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/biases"} -> tensor<120xbf16> loc(#loc) + %132 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %133 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/biases"} -> tensor<40xbf16> loc(#loc) + %134 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/weights"} -> tensor<40x72x1x1xbf16> loc(#loc) + %135 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/biases"} -> tensor<72xbf16> loc(#loc) + %136 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %137 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/biases"} -> tensor<24xbf16> loc(#loc) + %138 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %139 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/biases"} -> tensor<72xbf16> loc(#loc) + %140 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/weights"} -> tensor<72x1x5x5xbf16> loc(#loc) + %141 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/biases"} -> tensor<72xbf16> loc(#loc) + %142 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %143 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/biases"} -> tensor<24xbf16> loc(#loc) + %144 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %145 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/biases"} -> tensor<72xbf16> loc(#loc) + %146 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/weights"} -> tensor<72x1x3x3xbf16> loc(#loc) + %147 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/biases"} -> tensor<72xbf16> loc(#loc) + %148 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %149 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/biases"} -> tensor<24xbf16> loc(#loc) + %150 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/weights"} -> tensor<24x64x1x1xbf16> loc(#loc) + %151 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/biases"} -> tensor<64xbf16> loc(#loc) + %152 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/weights"} -> tensor<64x1x3x3xbf16> loc(#loc) + %153 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/biases"} -> tensor<64xbf16> loc(#loc) + %154 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/weights"} -> tensor<64x16x1x1xbf16> loc(#loc) + %155 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/biases"} -> tensor<16xbf16> loc(#loc) + %156 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/weights"} -> tensor<16x16x1x1xbf16> loc(#loc) + %157 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/biases"} -> tensor<16xbf16> loc(#loc) + %158 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/weights"} -> tensor<16x1x3x3xbf16> loc(#loc) + %159 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/biases"} -> tensor<16xbf16> loc(#loc) + %160 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/weights"} -> tensor<16x3x3x3xbf16> loc(#loc) + %161 = xten_nn.load_external_const {file = "constants.h5", key = "Div_16/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc6) + %162 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_14/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc309) + %163 = xten_nn.subgraph (%arg5 = %0: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Slice_7", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Slice_7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "W", + config.dim_c = 184 : ui32, + config.dim_h = 320 : ui32, + config.dim_w = 4 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_7", + OutputName = "Slice_7", + size = array, + start = array} : (tensor<1x180x320x4xbf16>) -> tensor<1x180x320x3xbf16> loc(#loc9) + xten_nn.output %461 : tensor<1x180x320x3xbf16> loc(#loc9) + } -> tensor<1x180x320x3xbf16> loc(#loc9) + %164 = xten_nn.subgraph (%arg5 = %163: tensor<1x180x320x3xbf16>) attributes { + LayerName = "Generated-#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Generated-#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "BufferPadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_0_padded = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_1_padded = 23 : ui32, + config.dim_2 = 3 : ui32, + config.dim_2_padded = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_padded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x180x320x3xbf16> loc(#loc10) + } -> tensor<1x180x320x3xbf16> loc(#loc10) + %165 = xten_nn.subgraph (%arg5 = %164: tensor<1x180x320x3xbf16>) attributes { + LayerName = "Generated-#2", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Generated-#3", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_2 = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 10 : ui32 + }} { + %461 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc11) + %462 = tosa.transpose %arg5, %461 : (tensor<1x180x320x3xbf16>, tensor<4xi32>) -> tensor<1x3x180x320xbf16> loc(#loc311) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc311) + } -> tensor<1x3x180x320xbf16> loc(#loc310) + %166 = xten_nn.subgraph (%arg5 = %165: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Generated-#4", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Generated-#5", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 184 : ui32, + config.dim_0_unpadded = 180 : ui32, + config.dim_1 = 1 : ui32, + config.dim_1_unpadded = 1 : ui32, + config.dim_2 = 320 : ui32, + config.dim_2_unpadded = 320 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x3x180x320xbf16> loc(#loc10) + } -> tensor<1x3x180x320xbf16> loc(#loc10) + %167 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>, %arg6 = %162: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Initializer_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Initializer_398", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Sub_14", OutputName = "Initializer_398"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc309) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc309) + } -> tensor<1x3x180x320xbf16> loc(#loc309) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc309) + } -> tensor<1x3x180x320xbf16> loc(#loc309) + %168 = xten_nn.subgraph (%arg5 = %167: tensor<1x3x180x320xbf16>, %arg6 = %161: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + OutputName = "Div_16", + PartOfLayerName = "Div_16", + shift = 0 : i8} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + %169 = xten_nn.subgraph (%arg5 = %168: tensor<1x3x180x320xbf16>, %arg6 = %160: tensor<16x3x3x3xbf16>, %arg7 = %159: tensor<16xbf16>) attributes { + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x3x180x320xbf16>, %arg9 = %arg6: tensor<16x3x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 2 : ui8, + config.stride_w = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x3x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x3xbf16> loc(#loc13) + %465 = tosa.transpose %arg8, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc13) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_17", + PartOfOutputName = "Conv_17", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>, tensor<16x3x3x3xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc13) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc13) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc13) + } -> tensor<1x16x90x160xbf16> loc(#loc13) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc13) + } -> tensor<1x16x90x160xbf16> loc(#loc13) + %170 = xten_nn.subgraph (%arg5 = %169: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_19", OutputName = "Add_19"} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc14) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc14) + } -> tensor<1x16x90x160xbf16> loc(#loc14) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc14) + } -> tensor<1x16x90x160xbf16> loc(#loc14) + %171 = xten_nn.subgraph (%arg5 = %170: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_22", + OutputName = "Clip_22", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + %172 = xten_nn.subgraph (%arg5 = %171: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_24", + OutputName = "Div_24", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + %173 = xten_nn.subgraph (%arg5 = %169: tensor<1x16x90x160xbf16>, %arg6 = %172: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_25", + OutputName = "Mul_25", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + %174 = xten_nn.subgraph (%arg5 = %173: tensor<1x16x90x160xbf16>, %arg6 = %158: tensor<16x1x3x3xbf16>, %arg7 = %157: tensor<16xbf16>) attributes { + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<16x1x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc312) + %465 = tosa.transpose %arg9, %464 : (tensor<16x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x16x1xbf16> loc(#loc312) + %466 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc312) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_26", + PartOfOutputName = "Conv_26", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<3x3x16x1xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc18) + %468 = tosa.clamp %467 { + LayerName = "Relu_27", + OutputName = "Relu_27", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc19) + %469 = tosa.transpose %468, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc312) + xten_nn.output %469 : tensor<1x16x90x160xbf16> loc(#loc19) + } -> tensor<1x16x90x160xbf16> loc(#loc312) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc312) + } -> tensor<1x16x90x160xbf16> loc(#loc312) + %175 = xten_nn.subgraph (%arg5 = %174: tensor<1x16x90x160xbf16>, %arg6 = %156: tensor<16x16x1x1xbf16>, %arg7 = %155: tensor<16xbf16>, %arg8 = %173: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Conv_28", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x16x90x160xbf16>, %arg10 = %arg6: tensor<16x16x1x1xbf16>, %arg11 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_28", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_28", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc20) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<16x16x1x1xbf16>) -> tensor<16x1x1x16xbf16> loc(#loc20) + %466 = tosa.transpose %arg9, %464 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_28", + PartOfOutputName = "Conv_28", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<16x1x1x16xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %468 = tosa.transpose %467, %463 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc20) + xten_nn.output %468 : tensor<1x16x90x160xbf16> loc(#loc20) + } -> tensor<1x16x90x160xbf16> loc(#loc20) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x16x90x160xbf16>, %arg10 = %arg8: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_29", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_29", OutputName = "Add_29"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc313) + %176 = xten_nn.subgraph (%arg5 = %175: tensor<1x16x90x160xbf16>, %arg6 = %154: tensor<64x16x1x1xbf16>, %arg7 = %153: tensor<64xbf16>) attributes { + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<64x16x1x1xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc314) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<64x16x1x1xbf16>) -> tensor<64x1x1x16xbf16> loc(#loc314) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc314) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_30", + PartOfOutputName = "Conv_30", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<64x1x1x16xbf16>, tensor<64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc22) + %467 = tosa.clamp %466 { + LayerName = "Relu_31", + OutputName = "Relu_31", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc23) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x64xbf16>, tensor<4xi32>) -> tensor<1x64x90x160xbf16> loc(#loc314) + xten_nn.output %468 : tensor<1x64x90x160xbf16> loc(#loc23) + } -> tensor<1x64x90x160xbf16> loc(#loc314) + xten_nn.output %461 : tensor<1x64x90x160xbf16> loc(#loc314) + } -> tensor<1x64x90x160xbf16> loc(#loc314) + %177 = xten_nn.subgraph (%arg5 = %176: tensor<1x64x90x160xbf16>, %arg6 = %152: tensor<64x1x3x3xbf16>, %arg7 = %151: tensor<64xbf16>) attributes { + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x90x160xbf16>, %arg9 = %arg6: tensor<64x1x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc315) + %465 = tosa.transpose %arg9, %464 : (tensor<64x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x64x1xbf16> loc(#loc315) + %466 = tosa.transpose %arg8, %463 : (tensor<1x64x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x64xbf16> loc(#loc315) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_32", + PartOfOutputName = "Conv_32", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x64xbf16>, tensor<3x3x64x1xbf16>, tensor<64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc24) + %468 = tosa.clamp %467 { + LayerName = "Relu_33", + OutputName = "Relu_33", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc25) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x64xbf16>, tensor<4xi32>) -> tensor<1x64x45x80xbf16> loc(#loc315) + xten_nn.output %469 : tensor<1x64x45x80xbf16> loc(#loc25) + } -> tensor<1x64x45x80xbf16> loc(#loc315) + xten_nn.output %461 : tensor<1x64x45x80xbf16> loc(#loc315) + } -> tensor<1x64x45x80xbf16> loc(#loc315) + %178 = xten_nn.subgraph (%arg5 = %177: tensor<1x64x45x80xbf16>, %arg6 = %150: tensor<24x64x1x1xbf16>, %arg7 = %149: tensor<24xbf16>) attributes { + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x45x80xbf16>, %arg9 = %arg6: tensor<24x64x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc26) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x64x1x1xbf16>) -> tensor<24x1x1x64xbf16> loc(#loc26) + %465 = tosa.transpose %arg8, %463 : (tensor<1x64x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x64xbf16> loc(#loc26) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_34", + PartOfOutputName = "Conv_34", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x64xbf16>, tensor<24x1x1x64xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc26) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc26) + xten_nn.output %467 : tensor<1x24x45x80xbf16> loc(#loc26) + } -> tensor<1x24x45x80xbf16> loc(#loc26) + xten_nn.output %461 : tensor<1x24x45x80xbf16> loc(#loc26) + } -> tensor<1x24x45x80xbf16> loc(#loc26) + %179 = xten_nn.subgraph (%arg5 = %178: tensor<1x24x45x80xbf16>, %arg6 = %148: tensor<72x24x1x1xbf16>, %arg7 = %147: tensor<72xbf16>) attributes { + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc316) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc316) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc316) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_35", + PartOfOutputName = "Conv_35", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc27) + %467 = tosa.clamp %466 { + LayerName = "Relu_36", + OutputName = "Relu_36", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc28) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc316) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc28) + } -> tensor<1x72x45x80xbf16> loc(#loc316) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc316) + } -> tensor<1x72x45x80xbf16> loc(#loc316) + %180 = xten_nn.subgraph (%arg5 = %179: tensor<1x72x45x80xbf16>, %arg6 = %146: tensor<72x1x3x3xbf16>, %arg7 = %145: tensor<72xbf16>) attributes { + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x3x3xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc317) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x72x1xbf16> loc(#loc317) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc317) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_37", + PartOfOutputName = "Conv_37", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<3x3x72x1xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc29) + %468 = tosa.clamp %467 { + LayerName = "Relu_38", + OutputName = "Relu_38", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc30) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc317) + xten_nn.output %469 : tensor<1x72x45x80xbf16> loc(#loc30) + } -> tensor<1x72x45x80xbf16> loc(#loc317) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc317) + } -> tensor<1x72x45x80xbf16> loc(#loc317) + %181 = xten_nn.subgraph (%arg5 = %180: tensor<1x72x45x80xbf16>, %arg6 = %144: tensor<24x72x1x1xbf16>, %arg7 = %143: tensor<24xbf16>, %arg8 = %178: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Conv_39", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x72x45x80xbf16>, %arg10 = %arg6: tensor<24x72x1x1xbf16>, %arg11 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_39", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_39", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc31) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc31) + %466 = tosa.transpose %arg9, %464 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc31) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_39", + PartOfOutputName = "Conv_39", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc31) + %468 = tosa.transpose %467, %463 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc31) + xten_nn.output %468 : tensor<1x24x45x80xbf16> loc(#loc31) + } -> tensor<1x24x45x80xbf16> loc(#loc31) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x24x45x80xbf16>, %arg10 = %arg8: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Add_40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_40", OutputName = "Add_40"} : (tensor<1x24x45x80xbf16>, tensor<1x24x45x80xbf16>) -> tensor<1x24x45x80xbf16> loc(#loc32) + xten_nn.output %463 : tensor<1x24x45x80xbf16> loc(#loc32) + } -> tensor<1x24x45x80xbf16> loc(#loc32) + xten_nn.output %462 : tensor<1x24x45x80xbf16> loc(#loc32) + } -> tensor<1x24x45x80xbf16> loc(#loc318) + %182 = xten_nn.subgraph (%arg5 = %181: tensor<1x24x45x80xbf16>, %arg6 = %142: tensor<72x24x1x1xbf16>, %arg7 = %141: tensor<72xbf16>) attributes { + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc319) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc319) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc319) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_41", + PartOfOutputName = "Conv_41", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc33) + %467 = tosa.clamp %466 { + LayerName = "Relu_42", + OutputName = "Relu_42", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc34) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc319) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc34) + } -> tensor<1x72x45x80xbf16> loc(#loc319) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc319) + } -> tensor<1x72x45x80xbf16> loc(#loc319) + %183 = xten_nn.subgraph (%arg5 = %182: tensor<1x72x45x80xbf16>, %arg6 = %140: tensor<72x1x5x5xbf16>, %arg7 = %139: tensor<72xbf16>) attributes { + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x5x5xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 1]], + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc320) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x72x1xbf16> loc(#loc320) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc320) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_43", + PartOfOutputName = "Conv_43", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<5x5x72x1xbf16>, tensor<72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc35) + %468 = tosa.clamp %467 { + LayerName = "Relu_44", + OutputName = "Relu_44", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc36) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x72xbf16>, tensor<4xi32>) -> tensor<1x72x23x40xbf16> loc(#loc320) + xten_nn.output %469 : tensor<1x72x23x40xbf16> loc(#loc36) + } -> tensor<1x72x23x40xbf16> loc(#loc320) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc320) + } -> tensor<1x72x23x40xbf16> loc(#loc320) + %184 = xten_nn.subgraph (%arg5 = %183: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Generated-#6", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 9 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x72x23x40xbf16>) -> tensor<1x72x1x920xbf16> loc(#loc37) + xten_nn.output %461 : tensor<1x72x1x920xbf16> loc(#loc37) + } -> tensor<1x72x1x920xbf16> loc(#loc37) + %185 = xten_nn.subgraph (%arg5 = %184: tensor<1x72x1x920xbf16>) attributes { + LayerName = "Generated-#8", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#9", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x920xbf16>) attributes { + LayerName = "Generated-#8", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#9", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 72 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x72x1x920xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc37) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc37) + } -> tensor<1x72x1x1xbf16> loc(#loc37) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc37) + } -> tensor<1x72x1x1xbf16> loc(#loc37) + %186 = xten_nn.subgraph (%arg5 = %185: tensor<1x72x1x1xbf16>, %arg6 = %138: tensor<24x72x1x1xbf16>, %arg7 = %137: tensor<24xbf16>) attributes { + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x1x1xbf16>, %arg9 = %arg6: tensor<24x72x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc321) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc321) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_46", + PartOfOutputName = "Conv_46", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc38) + %465 = tosa.clamp %464 { + LayerName = "Relu_47", + OutputName = "Relu_47", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc39) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x24xbf16>) -> tensor<1x24x1x1xbf16> loc(#loc321) + xten_nn.output %466 : tensor<1x24x1x1xbf16> loc(#loc39) + } -> tensor<1x24x1x1xbf16> loc(#loc321) + xten_nn.output %461 : tensor<1x24x1x1xbf16> loc(#loc321) + } -> tensor<1x24x1x1xbf16> loc(#loc321) + %187 = xten_nn.subgraph (%arg5 = %186: tensor<1x24x1x1xbf16>, %arg6 = %136: tensor<72x24x1x1xbf16>, %arg7 = %135: tensor<72xbf16>) attributes { + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x1x1xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc40) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x24x1x1xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc40) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_48", + PartOfOutputName = "Conv_48", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc40) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x72xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc40) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc40) + } -> tensor<1x72x1x1xbf16> loc(#loc40) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc40) + } -> tensor<1x72x1x1xbf16> loc(#loc40) + %188 = xten_nn.subgraph (%arg5 = %187: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_50", OutputName = "Add_50"} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc41) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc41) + } -> tensor<1x72x1x1xbf16> loc(#loc41) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc41) + } -> tensor<1x72x1x1xbf16> loc(#loc41) + %189 = xten_nn.subgraph (%arg5 = %188: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_53", + OutputName = "Clip_53", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + %190 = xten_nn.subgraph (%arg5 = %189: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_55", + OutputName = "Div_55", + shift = 0 : i8} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + %191 = xten_nn.subgraph (%arg5 = %190: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Generated-#10", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#11", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 72 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + %192 = xten_nn.subgraph (%arg5 = %191: tensor<1x72x23x40xbf16>, %arg6 = %183: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x72x23x40xbf16>, %arg8 = %arg6: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_56", + OutputName = "Mul_56", + shift = 0 : i8} : (tensor<1x72x23x40xbf16>, tensor<1x72x23x40xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %462 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + %193 = xten_nn.subgraph (%arg5 = %192: tensor<1x72x23x40xbf16>, %arg6 = %134: tensor<40x72x1x1xbf16>, %arg7 = %133: tensor<40xbf16>) attributes { + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x23x40xbf16>, %arg9 = %arg6: tensor<40x72x1x1xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc45) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<40x72x1x1xbf16>) -> tensor<40x1x1x72xbf16> loc(#loc45) + %465 = tosa.transpose %arg8, %463 : (tensor<1x72x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x72xbf16> loc(#loc45) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_57", + PartOfOutputName = "Conv_57", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x72xbf16>, tensor<40x1x1x72xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc45) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc45) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc45) + } -> tensor<1x40x23x40xbf16> loc(#loc45) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc45) + } -> tensor<1x40x23x40xbf16> loc(#loc45) + %194 = xten_nn.subgraph (%arg5 = %193: tensor<1x40x23x40xbf16>, %arg6 = %132: tensor<120x40x1x1xbf16>, %arg7 = %131: tensor<120xbf16>) attributes { + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc322) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc322) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc322) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_58", + PartOfOutputName = "Conv_58", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc46) + %467 = tosa.clamp %466 { + LayerName = "Relu_59", + OutputName = "Relu_59", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc47) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc322) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc47) + } -> tensor<1x120x23x40xbf16> loc(#loc322) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc322) + } -> tensor<1x120x23x40xbf16> loc(#loc322) + %195 = xten_nn.subgraph (%arg5 = %194: tensor<1x120x23x40xbf16>, %arg6 = %130: tensor<120x1x5x5xbf16>, %arg7 = %129: tensor<120xbf16>) attributes { + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc323) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc323) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc323) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_60", + PartOfOutputName = "Conv_60", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc48) + %468 = tosa.clamp %467 { + LayerName = "Relu_61", + OutputName = "Relu_61", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc49) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc323) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc49) + } -> tensor<1x120x23x40xbf16> loc(#loc323) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc323) + } -> tensor<1x120x23x40xbf16> loc(#loc323) + %196 = xten_nn.subgraph (%arg5 = %195: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Generated-#12", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#13", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc50) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc50) + } -> tensor<1x120x1x920xbf16> loc(#loc50) + %197 = xten_nn.subgraph (%arg5 = %196: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#15", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#15", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc50) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc50) + } -> tensor<1x120x1x1xbf16> loc(#loc50) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc50) + } -> tensor<1x120x1x1xbf16> loc(#loc50) + %198 = xten_nn.subgraph (%arg5 = %197: tensor<1x120x1x1xbf16>, %arg6 = %128: tensor<32x120x1x1xbf16>, %arg7 = %127: tensor<32xbf16>) attributes { + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc324) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc324) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_63", + PartOfOutputName = "Conv_63", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc51) + %465 = tosa.clamp %464 { + LayerName = "Relu_64", + OutputName = "Relu_64", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc52) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc324) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc52) + } -> tensor<1x32x1x1xbf16> loc(#loc324) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc324) + } -> tensor<1x32x1x1xbf16> loc(#loc324) + %199 = xten_nn.subgraph (%arg5 = %198: tensor<1x32x1x1xbf16>, %arg6 = %126: tensor<120x32x1x1xbf16>, %arg7 = %125: tensor<120xbf16>) attributes { + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc53) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc53) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_65", + PartOfOutputName = "Conv_65", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc53) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc53) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc53) + } -> tensor<1x120x1x1xbf16> loc(#loc53) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc53) + } -> tensor<1x120x1x1xbf16> loc(#loc53) + %200 = xten_nn.subgraph (%arg5 = %199: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_67", OutputName = "Add_67"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc54) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc54) + } -> tensor<1x120x1x1xbf16> loc(#loc54) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc54) + } -> tensor<1x120x1x1xbf16> loc(#loc54) + %201 = xten_nn.subgraph (%arg5 = %200: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_70", + OutputName = "Clip_70", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + %202 = xten_nn.subgraph (%arg5 = %201: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_72", + OutputName = "Div_72", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + %203 = xten_nn.subgraph (%arg5 = %202: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Generated-#16", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#17", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + %204 = xten_nn.subgraph (%arg5 = %203: tensor<1x120x23x40xbf16>, %arg6 = %195: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_73", + OutputName = "Mul_73", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + %205 = xten_nn.subgraph (%arg5 = %204: tensor<1x120x23x40xbf16>, %arg6 = %124: tensor<40x120x1x1xbf16>, %arg7 = %123: tensor<40xbf16>, %arg8 = %193: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_74", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_74", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_74", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc58) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc58) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc58) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_74", + PartOfOutputName = "Conv_74", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc58) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc58) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc58) + } -> tensor<1x40x23x40xbf16> loc(#loc58) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_75", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_75", OutputName = "Add_75"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc59) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc59) + } -> tensor<1x40x23x40xbf16> loc(#loc59) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc59) + } -> tensor<1x40x23x40xbf16> loc(#loc325) + %206 = xten_nn.subgraph (%arg5 = %205: tensor<1x40x23x40xbf16>, %arg6 = %122: tensor<120x40x1x1xbf16>, %arg7 = %121: tensor<120xbf16>) attributes { + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc326) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc326) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc326) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_76", + PartOfOutputName = "Conv_76", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc60) + %467 = tosa.clamp %466 { + LayerName = "Relu_77", + OutputName = "Relu_77", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc61) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc326) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc61) + } -> tensor<1x120x23x40xbf16> loc(#loc326) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc326) + } -> tensor<1x120x23x40xbf16> loc(#loc326) + %207 = xten_nn.subgraph (%arg5 = %206: tensor<1x120x23x40xbf16>, %arg6 = %120: tensor<120x1x5x5xbf16>, %arg7 = %119: tensor<120xbf16>) attributes { + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc327) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc327) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc327) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_78", + PartOfOutputName = "Conv_78", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc62) + %468 = tosa.clamp %467 { + LayerName = "Relu_79", + OutputName = "Relu_79", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc63) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc327) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc63) + } -> tensor<1x120x23x40xbf16> loc(#loc327) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc327) + } -> tensor<1x120x23x40xbf16> loc(#loc327) + %208 = xten_nn.subgraph (%arg5 = %207: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Generated-#18", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#19", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc64) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc64) + } -> tensor<1x120x1x920xbf16> loc(#loc64) + %209 = xten_nn.subgraph (%arg5 = %208: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#20", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#21", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#20", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#21", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc64) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc64) + } -> tensor<1x120x1x1xbf16> loc(#loc64) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc64) + } -> tensor<1x120x1x1xbf16> loc(#loc64) + %210 = xten_nn.subgraph (%arg5 = %209: tensor<1x120x1x1xbf16>, %arg6 = %118: tensor<32x120x1x1xbf16>, %arg7 = %117: tensor<32xbf16>) attributes { + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc328) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc328) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_81", + PartOfOutputName = "Conv_81", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc65) + %465 = tosa.clamp %464 { + LayerName = "Relu_82", + OutputName = "Relu_82", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc66) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc328) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc66) + } -> tensor<1x32x1x1xbf16> loc(#loc328) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc328) + } -> tensor<1x32x1x1xbf16> loc(#loc328) + %211 = xten_nn.subgraph (%arg5 = %210: tensor<1x32x1x1xbf16>, %arg6 = %116: tensor<120x32x1x1xbf16>, %arg7 = %115: tensor<120xbf16>) attributes { + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc67) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc67) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_83", + PartOfOutputName = "Conv_83", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc67) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc67) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc67) + } -> tensor<1x120x1x1xbf16> loc(#loc67) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc67) + } -> tensor<1x120x1x1xbf16> loc(#loc67) + %212 = xten_nn.subgraph (%arg5 = %211: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_85", OutputName = "Add_85"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc68) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc68) + } -> tensor<1x120x1x1xbf16> loc(#loc68) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc68) + } -> tensor<1x120x1x1xbf16> loc(#loc68) + %213 = xten_nn.subgraph (%arg5 = %212: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_88", + OutputName = "Clip_88", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + %214 = xten_nn.subgraph (%arg5 = %213: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_90", + OutputName = "Div_90", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + %215 = xten_nn.subgraph (%arg5 = %214: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Generated-#22", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#23", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + %216 = xten_nn.subgraph (%arg5 = %215: tensor<1x120x23x40xbf16>, %arg6 = %207: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_91", + OutputName = "Mul_91", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + %217 = xten_nn.subgraph (%arg5 = %216: tensor<1x120x23x40xbf16>, %arg6 = %114: tensor<40x120x1x1xbf16>, %arg7 = %113: tensor<40xbf16>, %arg8 = %205: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_92", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_92", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_92", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc72) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc72) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc72) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_92", + PartOfOutputName = "Conv_92", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc72) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc72) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc72) + } -> tensor<1x40x23x40xbf16> loc(#loc72) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_93", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_93", OutputName = "Add_93"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc73) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc73) + } -> tensor<1x40x23x40xbf16> loc(#loc73) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc73) + } -> tensor<1x40x23x40xbf16> loc(#loc329) + %218 = xten_nn.subgraph (%arg5 = %217: tensor<1x40x23x40xbf16>, %arg6 = %112: tensor<240x40x1x1xbf16>, %arg7 = %111: tensor<240xbf16>) attributes { + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<240x40x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc74) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x40x1x1xbf16>) -> tensor<240x1x1x40xbf16> loc(#loc74) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc74) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_94", + PartOfOutputName = "Conv_94", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<240x1x1x40xbf16>, tensor<240xbf16>) -> tensor<1x23x40x240xbf16> loc(#loc74) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x240xbf16>, tensor<4xi32>) -> tensor<1x240x23x40xbf16> loc(#loc74) + xten_nn.output %467 : tensor<1x240x23x40xbf16> loc(#loc74) + } -> tensor<1x240x23x40xbf16> loc(#loc74) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc74) + } -> tensor<1x240x23x40xbf16> loc(#loc74) + %219 = xten_nn.subgraph (%arg5 = %218: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_96", OutputName = "Add_96"} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc75) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc75) + } -> tensor<1x240x23x40xbf16> loc(#loc75) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc75) + } -> tensor<1x240x23x40xbf16> loc(#loc75) + %220 = xten_nn.subgraph (%arg5 = %219: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_99", + OutputName = "Clip_99", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + %221 = xten_nn.subgraph (%arg5 = %220: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_101", + OutputName = "Div_101", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + %222 = xten_nn.subgraph (%arg5 = %218: tensor<1x240x23x40xbf16>, %arg6 = %221: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x23x40xbf16>, %arg8 = %arg6: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_102", + OutputName = "Mul_102", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + %223 = xten_nn.subgraph (%arg5 = %222: tensor<1x240x23x40xbf16>, %arg6 = %110: tensor<240x1x3x3xbf16>, %arg7 = %109: tensor<240xbf16>) attributes { + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x23x40xbf16>, %arg9 = %arg6: tensor<240x1x3x3xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 0]], + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc79) + %465 = tosa.transpose %arg9, %464 : (tensor<240x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x240x1xbf16> loc(#loc79) + %466 = tosa.transpose %arg8, %463 : (tensor<1x240x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x240xbf16> loc(#loc79) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_103", + PartOfOutputName = "Conv_103", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x240xbf16>, tensor<3x3x240x1xbf16>, tensor<240xbf16>) -> tensor<1x12x20x240xbf16> loc(#loc79) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x240xbf16>, tensor<4xi32>) -> tensor<1x240x12x20xbf16> loc(#loc79) + xten_nn.output %468 : tensor<1x240x12x20xbf16> loc(#loc79) + } -> tensor<1x240x12x20xbf16> loc(#loc79) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc79) + } -> tensor<1x240x12x20xbf16> loc(#loc79) + %224 = xten_nn.subgraph (%arg5 = %223: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_105", OutputName = "Add_105"} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc80) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc80) + } -> tensor<1x240x12x20xbf16> loc(#loc80) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc80) + } -> tensor<1x240x12x20xbf16> loc(#loc80) + %225 = xten_nn.subgraph (%arg5 = %224: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_108", + OutputName = "Clip_108", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + %226 = xten_nn.subgraph (%arg5 = %225: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_110", + OutputName = "Div_110", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + %227 = xten_nn.subgraph (%arg5 = %223: tensor<1x240x12x20xbf16>, %arg6 = %226: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x12x20xbf16>, %arg8 = %arg6: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_111", + OutputName = "Mul_111", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + %228 = xten_nn.subgraph (%arg5 = %227: tensor<1x240x12x20xbf16>, %arg6 = %108: tensor<80x240x1x1xbf16>, %arg7 = %107: tensor<80xbf16>) attributes { + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x12x20xbf16>, %arg9 = %arg6: tensor<80x240x1x1xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc84) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<80x240x1x1xbf16>) -> tensor<80x1x1x240xbf16> loc(#loc84) + %465 = tosa.transpose %arg8, %463 : (tensor<1x240x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x240xbf16> loc(#loc84) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_112", + PartOfOutputName = "Conv_112", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x240xbf16>, tensor<80x1x1x240xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc84) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc84) + xten_nn.output %467 : tensor<1x80x12x20xbf16> loc(#loc84) + } -> tensor<1x80x12x20xbf16> loc(#loc84) + xten_nn.output %461 : tensor<1x80x12x20xbf16> loc(#loc84) + } -> tensor<1x80x12x20xbf16> loc(#loc84) + %229 = xten_nn.subgraph (%arg5 = %228: tensor<1x80x12x20xbf16>, %arg6 = %106: tensor<200x80x1x1xbf16>, %arg7 = %105: tensor<200xbf16>) attributes { + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<200x80x1x1xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc85) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<200x80x1x1xbf16>) -> tensor<200x1x1x80xbf16> loc(#loc85) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc85) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_113", + PartOfOutputName = "Conv_113", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<200x1x1x80xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc85) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc85) + xten_nn.output %467 : tensor<1x200x12x20xbf16> loc(#loc85) + } -> tensor<1x200x12x20xbf16> loc(#loc85) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc85) + } -> tensor<1x200x12x20xbf16> loc(#loc85) + %230 = xten_nn.subgraph (%arg5 = %229: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_115", OutputName = "Add_115"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc86) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc86) + } -> tensor<1x200x12x20xbf16> loc(#loc86) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc86) + } -> tensor<1x200x12x20xbf16> loc(#loc86) + %231 = xten_nn.subgraph (%arg5 = %230: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_118", + OutputName = "Clip_118", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + %232 = xten_nn.subgraph (%arg5 = %231: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_120", + OutputName = "Div_120", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + %233 = xten_nn.subgraph (%arg5 = %229: tensor<1x200x12x20xbf16>, %arg6 = %232: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_121", + OutputName = "Mul_121", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + %234 = xten_nn.subgraph (%arg5 = %233: tensor<1x200x12x20xbf16>, %arg6 = %104: tensor<200x1x3x3xbf16>, %arg7 = %103: tensor<200xbf16>) attributes { + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x200x12x20xbf16>, %arg9 = %arg6: tensor<200x1x3x3xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc90) + %465 = tosa.transpose %arg9, %464 : (tensor<200x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x200x1xbf16> loc(#loc90) + %466 = tosa.transpose %arg8, %463 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc90) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_122", + PartOfOutputName = "Conv_122", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<3x3x200x1xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc90) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %468 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + %235 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_124", OutputName = "Add_124"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + %236 = xten_nn.subgraph (%arg5 = %235: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_127", + OutputName = "Clip_127", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + %237 = xten_nn.subgraph (%arg5 = %236: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_129", + OutputName = "Div_129", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + %238 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>, %arg6 = %237: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_130", + OutputName = "Mul_130", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + %239 = xten_nn.subgraph (%arg5 = %238: tensor<1x200x12x20xbf16>, %arg6 = %102: tensor<80x200x1x1xbf16>, %arg7 = %101: tensor<80xbf16>, %arg8 = %228: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_131", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x200x12x20xbf16>, %arg10 = %arg6: tensor<80x200x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_131", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_131", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc95) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x200x1x1xbf16>) -> tensor<80x1x1x200xbf16> loc(#loc95) + %466 = tosa.transpose %arg9, %464 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc95) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_131", + PartOfOutputName = "Conv_131", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<80x1x1x200xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc95) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc95) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc95) + } -> tensor<1x80x12x20xbf16> loc(#loc95) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_132", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_132", OutputName = "Add_132"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc96) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc96) + } -> tensor<1x80x12x20xbf16> loc(#loc96) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc96) + } -> tensor<1x80x12x20xbf16> loc(#loc330) + %240 = xten_nn.subgraph (%arg5 = %239: tensor<1x80x12x20xbf16>, %arg6 = %100: tensor<184x80x1x1xbf16>, %arg7 = %99: tensor<184xbf16>) attributes { + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc97) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc97) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc97) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_133", + PartOfOutputName = "Conv_133", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc97) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc97) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc97) + } -> tensor<1x184x12x20xbf16> loc(#loc97) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc97) + } -> tensor<1x184x12x20xbf16> loc(#loc97) + %241 = xten_nn.subgraph (%arg5 = %240: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_135", OutputName = "Add_135"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc98) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc98) + } -> tensor<1x184x12x20xbf16> loc(#loc98) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc98) + } -> tensor<1x184x12x20xbf16> loc(#loc98) + %242 = xten_nn.subgraph (%arg5 = %241: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_138", + OutputName = "Clip_138", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + %243 = xten_nn.subgraph (%arg5 = %242: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_140", + OutputName = "Div_140", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + %244 = xten_nn.subgraph (%arg5 = %240: tensor<1x184x12x20xbf16>, %arg6 = %243: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_141", + OutputName = "Mul_141", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + %245 = xten_nn.subgraph (%arg5 = %244: tensor<1x184x12x20xbf16>, %arg6 = %98: tensor<184x1x3x3xbf16>, %arg7 = %97: tensor<184xbf16>) attributes { + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc102) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc102) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc102) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_142", + PartOfOutputName = "Conv_142", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc102) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + %246 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_144", OutputName = "Add_144"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + %247 = xten_nn.subgraph (%arg5 = %246: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_147", + OutputName = "Clip_147", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + %248 = xten_nn.subgraph (%arg5 = %247: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_149", + OutputName = "Div_149", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + %249 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>, %arg6 = %248: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_150", + OutputName = "Mul_150", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + %250 = xten_nn.subgraph (%arg5 = %249: tensor<1x184x12x20xbf16>, %arg6 = %96: tensor<80x184x1x1xbf16>, %arg7 = %95: tensor<80xbf16>, %arg8 = %239: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_151", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_151", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_151", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc107) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc107) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc107) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_151", + PartOfOutputName = "Conv_151", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc107) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc107) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc107) + } -> tensor<1x80x12x20xbf16> loc(#loc107) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_152", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_152", OutputName = "Add_152"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc108) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc108) + } -> tensor<1x80x12x20xbf16> loc(#loc108) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc108) + } -> tensor<1x80x12x20xbf16> loc(#loc331) + %251 = xten_nn.subgraph (%arg5 = %250: tensor<1x80x12x20xbf16>, %arg6 = %94: tensor<184x80x1x1xbf16>, %arg7 = %93: tensor<184xbf16>) attributes { + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc109) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc109) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc109) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_153", + PartOfOutputName = "Conv_153", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc109) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + %252 = xten_nn.subgraph (%arg5 = %251: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_155", OutputName = "Add_155"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + %253 = xten_nn.subgraph (%arg5 = %252: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_158", + OutputName = "Clip_158", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + %254 = xten_nn.subgraph (%arg5 = %253: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_160", + OutputName = "Div_160", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + %255 = xten_nn.subgraph (%arg5 = %251: tensor<1x184x12x20xbf16>, %arg6 = %254: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_161", + OutputName = "Mul_161", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + %256 = xten_nn.subgraph (%arg5 = %255: tensor<1x184x12x20xbf16>, %arg6 = %92: tensor<184x1x3x3xbf16>, %arg7 = %91: tensor<184xbf16>) attributes { + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc114) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc114) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc114) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_162", + PartOfOutputName = "Conv_162", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc114) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + %257 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_164", OutputName = "Add_164"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + %258 = xten_nn.subgraph (%arg5 = %257: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_167", + OutputName = "Clip_167", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + %259 = xten_nn.subgraph (%arg5 = %258: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_169", + OutputName = "Div_169", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + %260 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>, %arg6 = %259: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_170", + OutputName = "Mul_170", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + %261 = xten_nn.subgraph (%arg5 = %260: tensor<1x184x12x20xbf16>, %arg6 = %90: tensor<80x184x1x1xbf16>, %arg7 = %89: tensor<80xbf16>, %arg8 = %250: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_171", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_171", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_171", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc119) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc119) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc119) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_171", + PartOfOutputName = "Conv_171", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc119) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc119) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc119) + } -> tensor<1x80x12x20xbf16> loc(#loc119) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_172", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_172", OutputName = "Add_172"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc120) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc120) + } -> tensor<1x80x12x20xbf16> loc(#loc120) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc120) + } -> tensor<1x80x12x20xbf16> loc(#loc332) + %262 = xten_nn.subgraph (%arg5 = %261: tensor<1x80x12x20xbf16>, %arg6 = %88: tensor<480x80x1x1xbf16>, %arg7 = %87: tensor<480xbf16>) attributes { + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<480x80x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc121) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x80x1x1xbf16>) -> tensor<480x1x1x80xbf16> loc(#loc121) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc121) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_173", + PartOfOutputName = "Conv_173", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<480x1x1x80xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc121) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc121) + xten_nn.output %467 : tensor<1x480x12x20xbf16> loc(#loc121) + } -> tensor<1x480x12x20xbf16> loc(#loc121) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc121) + } -> tensor<1x480x12x20xbf16> loc(#loc121) + %263 = xten_nn.subgraph (%arg5 = %262: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_175", OutputName = "Add_175"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc122) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc122) + } -> tensor<1x480x12x20xbf16> loc(#loc122) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc122) + } -> tensor<1x480x12x20xbf16> loc(#loc122) + %264 = xten_nn.subgraph (%arg5 = %263: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_178", + OutputName = "Clip_178", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + %265 = xten_nn.subgraph (%arg5 = %264: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_180", + OutputName = "Div_180", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + %266 = xten_nn.subgraph (%arg5 = %262: tensor<1x480x12x20xbf16>, %arg6 = %265: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_181", + OutputName = "Mul_181", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + %267 = xten_nn.subgraph (%arg5 = %266: tensor<1x480x12x20xbf16>, %arg6 = %86: tensor<480x1x3x3xbf16>, %arg7 = %85: tensor<480xbf16>) attributes { + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<480x1x3x3xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc126) + %465 = tosa.transpose %arg9, %464 : (tensor<480x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x480x1xbf16> loc(#loc126) + %466 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc126) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_182", + PartOfOutputName = "Conv_182", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<3x3x480x1xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc126) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %468 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + %268 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_184", OutputName = "Add_184"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + %269 = xten_nn.subgraph (%arg5 = %268: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_187", + OutputName = "Clip_187", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + %270 = xten_nn.subgraph (%arg5 = %269: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_189", + OutputName = "Div_189", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + %271 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>, %arg6 = %270: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_190", + OutputName = "Mul_190", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + %272 = xten_nn.subgraph (%arg5 = %271: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Generated-#24", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#25", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 60 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x1x240xbf16> loc(#loc333) + xten_nn.output %461 : tensor<1x480x1x240xbf16> loc(#loc333) + } -> tensor<1x480x1x240xbf16> loc(#loc333) + %273 = xten_nn.subgraph (%arg5 = %272: tensor<1x480x1x240xbf16>) attributes { + LayerName = "Generated-#26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x240xbf16>) attributes { + LayerName = "Generated-#26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#27", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 480 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x480x1x240xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc131) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc131) + } -> tensor<1x480x1x1xbf16> loc(#loc131) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc131) + } -> tensor<1x480x1x1xbf16> loc(#loc131) + %274 = xten_nn.subgraph (%arg5 = %273: tensor<1x480x1x1xbf16>, %arg6 = %84: tensor<120x480x1x1xbf16>, %arg7 = %83: tensor<120xbf16>) attributes { + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x1x1xbf16>, %arg9 = %arg6: tensor<120x480x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x480x1x1xbf16>) -> tensor<120x1x1x480xbf16> loc(#loc334) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc334) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_192", + PartOfOutputName = "Conv_192", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x480xbf16>, tensor<120x1x1x480xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc132) + %465 = tosa.clamp %464 { + LayerName = "Relu_193", + OutputName = "Relu_193", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc133) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc334) + xten_nn.output %466 : tensor<1x120x1x1xbf16> loc(#loc133) + } -> tensor<1x120x1x1xbf16> loc(#loc334) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc334) + } -> tensor<1x120x1x1xbf16> loc(#loc334) + %275 = xten_nn.subgraph (%arg5 = %274: tensor<1x120x1x1xbf16>, %arg6 = %82: tensor<480x120x1x1xbf16>, %arg7 = %81: tensor<480xbf16>) attributes { + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<480x120x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x120x1x1xbf16>) -> tensor<480x1x1x120xbf16> loc(#loc134) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc134) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_194", + PartOfOutputName = "Conv_194", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<480x1x1x120xbf16>, tensor<480xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc134) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x480xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc134) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc134) + } -> tensor<1x480x1x1xbf16> loc(#loc134) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc134) + } -> tensor<1x480x1x1xbf16> loc(#loc134) + %276 = xten_nn.subgraph (%arg5 = %275: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_196", OutputName = "Add_196"} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc135) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc135) + } -> tensor<1x480x1x1xbf16> loc(#loc135) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc135) + } -> tensor<1x480x1x1xbf16> loc(#loc135) + %277 = xten_nn.subgraph (%arg5 = %276: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_199", + OutputName = "Clip_199", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + %278 = xten_nn.subgraph (%arg5 = %277: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_201", + OutputName = "Div_201", + shift = 0 : i8} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + %279 = xten_nn.subgraph (%arg5 = %278: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Generated-#28", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#29", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 480 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + %280 = xten_nn.subgraph (%arg5 = %279: tensor<1x480x12x20xbf16>, %arg6 = %271: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_202", + OutputName = "Mul_202", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + %281 = xten_nn.subgraph (%arg5 = %280: tensor<1x480x12x20xbf16>, %arg6 = %80: tensor<112x480x1x1xbf16>, %arg7 = %79: tensor<112xbf16>) attributes { + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<112x480x1x1xbf16>, %arg10 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc139) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<112x480x1x1xbf16>) -> tensor<112x1x1x480xbf16> loc(#loc139) + %465 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc139) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_203", + PartOfOutputName = "Conv_203", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<112x1x1x480xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc139) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc139) + xten_nn.output %467 : tensor<1x112x12x20xbf16> loc(#loc139) + } -> tensor<1x112x12x20xbf16> loc(#loc139) + xten_nn.output %461 : tensor<1x112x12x20xbf16> loc(#loc139) + } -> tensor<1x112x12x20xbf16> loc(#loc139) + %282 = xten_nn.subgraph (%arg5 = %281: tensor<1x112x12x20xbf16>, %arg6 = %78: tensor<672x112x1x1xbf16>, %arg7 = %77: tensor<672xbf16>) attributes { + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc140) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc140) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc140) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_204", + PartOfOutputName = "Conv_204", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc140) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc140) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc140) + } -> tensor<1x672x12x20xbf16> loc(#loc140) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc140) + } -> tensor<1x672x12x20xbf16> loc(#loc140) + %283 = xten_nn.subgraph (%arg5 = %282: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_206", OutputName = "Add_206"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc141) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc141) + } -> tensor<1x672x12x20xbf16> loc(#loc141) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc141) + } -> tensor<1x672x12x20xbf16> loc(#loc141) + %284 = xten_nn.subgraph (%arg5 = %283: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_209", + OutputName = "Clip_209", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + %285 = xten_nn.subgraph (%arg5 = %284: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_211", + OutputName = "Div_211", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + %286 = xten_nn.subgraph (%arg5 = %282: tensor<1x672x12x20xbf16>, %arg6 = %285: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_212", + OutputName = "Mul_212", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + %287 = xten_nn.subgraph (%arg5 = %286: tensor<1x672x12x20xbf16>, %arg6 = %76: tensor<672x1x3x3xbf16>, %arg7 = %75: tensor<672xbf16>) attributes { + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x3x3xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc145) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x672x1xbf16> loc(#loc145) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc145) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_213", + PartOfOutputName = "Conv_213", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<3x3x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc145) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + %288 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_215", OutputName = "Add_215"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + %289 = xten_nn.subgraph (%arg5 = %288: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_218", + OutputName = "Clip_218", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + %290 = xten_nn.subgraph (%arg5 = %289: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_220", + OutputName = "Div_220", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + %291 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>, %arg6 = %290: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_221", + OutputName = "Mul_221", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + %292 = xten_nn.subgraph (%arg5 = %291: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Generated-#30", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#31", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc335) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc335) + } -> tensor<1x672x1x240xbf16> loc(#loc335) + %293 = xten_nn.subgraph (%arg5 = %292: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#33", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc150) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc150) + } -> tensor<1x672x1x1xbf16> loc(#loc150) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc150) + } -> tensor<1x672x1x1xbf16> loc(#loc150) + %294 = xten_nn.subgraph (%arg5 = %293: tensor<1x672x1x1xbf16>, %arg6 = %74: tensor<168x672x1x1xbf16>, %arg7 = %73: tensor<168xbf16>) attributes { + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc336) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc336) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_223", + PartOfOutputName = "Conv_223", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc151) + %465 = tosa.clamp %464 { + LayerName = "Relu_224", + OutputName = "Relu_224", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc152) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc336) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc152) + } -> tensor<1x168x1x1xbf16> loc(#loc336) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc336) + } -> tensor<1x168x1x1xbf16> loc(#loc336) + %295 = xten_nn.subgraph (%arg5 = %294: tensor<1x168x1x1xbf16>, %arg6 = %72: tensor<672x168x1x1xbf16>, %arg7 = %71: tensor<672xbf16>) attributes { + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc153) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc153) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_225", + PartOfOutputName = "Conv_225", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc153) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc153) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc153) + } -> tensor<1x672x1x1xbf16> loc(#loc153) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc153) + } -> tensor<1x672x1x1xbf16> loc(#loc153) + %296 = xten_nn.subgraph (%arg5 = %295: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_227", OutputName = "Add_227"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc154) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc154) + } -> tensor<1x672x1x1xbf16> loc(#loc154) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc154) + } -> tensor<1x672x1x1xbf16> loc(#loc154) + %297 = xten_nn.subgraph (%arg5 = %296: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_230", + OutputName = "Clip_230", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + %298 = xten_nn.subgraph (%arg5 = %297: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_232", + OutputName = "Div_232", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + %299 = xten_nn.subgraph (%arg5 = %298: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Generated-#34", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#35", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + %300 = xten_nn.subgraph (%arg5 = %299: tensor<1x672x12x20xbf16>, %arg6 = %291: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_233", + OutputName = "Mul_233", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + %301 = xten_nn.subgraph (%arg5 = %300: tensor<1x672x12x20xbf16>, %arg6 = %70: tensor<112x672x1x1xbf16>, %arg7 = %69: tensor<112xbf16>, %arg8 = %281: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Conv_234", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x672x12x20xbf16>, %arg10 = %arg6: tensor<112x672x1x1xbf16>, %arg11 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_234", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_234", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc158) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<112x672x1x1xbf16>) -> tensor<112x1x1x672xbf16> loc(#loc158) + %466 = tosa.transpose %arg9, %464 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc158) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_234", + PartOfOutputName = "Conv_234", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<112x1x1x672xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc158) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc158) + xten_nn.output %468 : tensor<1x112x12x20xbf16> loc(#loc158) + } -> tensor<1x112x12x20xbf16> loc(#loc158) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x112x12x20xbf16>, %arg10 = %arg8: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Add_235", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_235", OutputName = "Add_235"} : (tensor<1x112x12x20xbf16>, tensor<1x112x12x20xbf16>) -> tensor<1x112x12x20xbf16> loc(#loc159) + xten_nn.output %463 : tensor<1x112x12x20xbf16> loc(#loc159) + } -> tensor<1x112x12x20xbf16> loc(#loc159) + xten_nn.output %462 : tensor<1x112x12x20xbf16> loc(#loc159) + } -> tensor<1x112x12x20xbf16> loc(#loc337) + %302 = xten_nn.subgraph (%arg5 = %301: tensor<1x112x12x20xbf16>, %arg6 = %68: tensor<672x112x1x1xbf16>, %arg7 = %67: tensor<672xbf16>) attributes { + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc160) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc160) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc160) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_236", + PartOfOutputName = "Conv_236", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc160) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc160) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc160) + } -> tensor<1x672x12x20xbf16> loc(#loc160) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc160) + } -> tensor<1x672x12x20xbf16> loc(#loc160) + %303 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_238", OutputName = "Add_238"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc161) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc161) + } -> tensor<1x672x12x20xbf16> loc(#loc161) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc161) + } -> tensor<1x672x12x20xbf16> loc(#loc161) + %304 = xten_nn.subgraph (%arg5 = %303: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_241", + OutputName = "Clip_241", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + %305 = xten_nn.subgraph (%arg5 = %304: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_243", + OutputName = "Div_243", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + %306 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>, %arg6 = %305: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_244", + OutputName = "Mul_244", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + %307 = xten_nn.subgraph (%arg5 = %306: tensor<1x672x12x20xbf16>, %arg6 = %66: tensor<672x1x9x9xbf16>, %arg7 = %65: tensor<672xbf16>) attributes { + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x9x9xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc165) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x672x1xbf16> loc(#loc165) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc165) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_245", + PartOfOutputName = "Conv_245", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<9x9x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc165) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + %308 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_247", OutputName = "Add_247"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + %309 = xten_nn.subgraph (%arg5 = %308: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_250", + OutputName = "Clip_250", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + %310 = xten_nn.subgraph (%arg5 = %309: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_252", + OutputName = "Div_252", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + %311 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>, %arg6 = %310: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_253", + OutputName = "Mul_253", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + %312 = xten_nn.subgraph (%arg5 = %311: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Generated-#36", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#37", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc338) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc338) + } -> tensor<1x672x1x240xbf16> loc(#loc338) + %313 = xten_nn.subgraph (%arg5 = %312: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#38", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#39", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#38", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#39", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc170) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc170) + } -> tensor<1x672x1x1xbf16> loc(#loc170) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc170) + } -> tensor<1x672x1x1xbf16> loc(#loc170) + %314 = xten_nn.subgraph (%arg5 = %313: tensor<1x672x1x1xbf16>, %arg6 = %64: tensor<168x672x1x1xbf16>, %arg7 = %63: tensor<168xbf16>) attributes { + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc339) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc339) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_255", + PartOfOutputName = "Conv_255", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc171) + %465 = tosa.clamp %464 { + LayerName = "Relu_256", + OutputName = "Relu_256", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc172) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc339) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc172) + } -> tensor<1x168x1x1xbf16> loc(#loc339) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc339) + } -> tensor<1x168x1x1xbf16> loc(#loc339) + %315 = xten_nn.subgraph (%arg5 = %314: tensor<1x168x1x1xbf16>, %arg6 = %62: tensor<672x168x1x1xbf16>, %arg7 = %61: tensor<672xbf16>) attributes { + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc173) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc173) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_257", + PartOfOutputName = "Conv_257", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc173) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc173) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc173) + } -> tensor<1x672x1x1xbf16> loc(#loc173) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc173) + } -> tensor<1x672x1x1xbf16> loc(#loc173) + %316 = xten_nn.subgraph (%arg5 = %315: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_259", OutputName = "Add_259"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc174) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc174) + } -> tensor<1x672x1x1xbf16> loc(#loc174) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc174) + } -> tensor<1x672x1x1xbf16> loc(#loc174) + %317 = xten_nn.subgraph (%arg5 = %316: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_262", + OutputName = "Clip_262", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + %318 = xten_nn.subgraph (%arg5 = %317: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_264", + OutputName = "Div_264", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + %319 = xten_nn.subgraph (%arg5 = %318: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Generated-#40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#41", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + %320 = xten_nn.subgraph (%arg5 = %319: tensor<1x672x12x20xbf16>, %arg6 = %311: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_265", + OutputName = "Mul_265", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + %321 = xten_nn.subgraph (%arg5 = %320: tensor<1x672x12x20xbf16>, %arg6 = %60: tensor<160x672x1x1xbf16>, %arg7 = %59: tensor<160xbf16>) attributes { + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<160x672x1x1xbf16>, %arg10 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc178) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<160x672x1x1xbf16>) -> tensor<160x1x1x672xbf16> loc(#loc178) + %465 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc178) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_266", + PartOfOutputName = "Conv_266", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<160x1x1x672xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc178) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc178) + xten_nn.output %467 : tensor<1x160x12x20xbf16> loc(#loc178) + } -> tensor<1x160x12x20xbf16> loc(#loc178) + xten_nn.output %461 : tensor<1x160x12x20xbf16> loc(#loc178) + } -> tensor<1x160x12x20xbf16> loc(#loc178) + %322 = xten_nn.subgraph (%arg5 = %321: tensor<1x160x12x20xbf16>, %arg6 = %58: tensor<960x160x1x1xbf16>, %arg7 = %57: tensor<960xbf16>) attributes { + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc179) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc179) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc179) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_267", + PartOfOutputName = "Conv_267", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc179) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc179) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc179) + } -> tensor<1x960x12x20xbf16> loc(#loc179) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc179) + } -> tensor<1x960x12x20xbf16> loc(#loc179) + %323 = xten_nn.subgraph (%arg5 = %322: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_269", OutputName = "Add_269"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc180) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc180) + } -> tensor<1x960x12x20xbf16> loc(#loc180) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc180) + } -> tensor<1x960x12x20xbf16> loc(#loc180) + %324 = xten_nn.subgraph (%arg5 = %323: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_272", + OutputName = "Clip_272", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + %325 = xten_nn.subgraph (%arg5 = %324: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_274", + OutputName = "Div_274", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + %326 = xten_nn.subgraph (%arg5 = %322: tensor<1x960x12x20xbf16>, %arg6 = %325: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_275", + OutputName = "Mul_275", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + %327 = xten_nn.subgraph (%arg5 = %326: tensor<1x960x12x20xbf16>, %arg6 = %56: tensor<960x1x9x9xbf16>, %arg7 = %55: tensor<960xbf16>) attributes { + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc184) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc184) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc184) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_276", + PartOfOutputName = "Conv_276", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc184) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + %328 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_278", OutputName = "Add_278"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + %329 = xten_nn.subgraph (%arg5 = %328: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_281", + OutputName = "Clip_281", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + %330 = xten_nn.subgraph (%arg5 = %329: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_283", + OutputName = "Div_283", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + %331 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>, %arg6 = %330: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_284", + OutputName = "Mul_284", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + %332 = xten_nn.subgraph (%arg5 = %331: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Generated-#42", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#43", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc340) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc340) + } -> tensor<1x960x1x240xbf16> loc(#loc340) + %333 = xten_nn.subgraph (%arg5 = %332: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#44", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#45", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#44", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#45", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc189) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc189) + } -> tensor<1x960x1x1xbf16> loc(#loc189) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc189) + } -> tensor<1x960x1x1xbf16> loc(#loc189) + %334 = xten_nn.subgraph (%arg5 = %333: tensor<1x960x1x1xbf16>, %arg6 = %54: tensor<240x960x1x1xbf16>, %arg7 = %53: tensor<240xbf16>) attributes { + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc341) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc341) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_286", + PartOfOutputName = "Conv_286", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc190) + %465 = tosa.clamp %464 { + LayerName = "Relu_287", + OutputName = "Relu_287", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc191) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc341) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc191) + } -> tensor<1x240x1x1xbf16> loc(#loc341) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc341) + } -> tensor<1x240x1x1xbf16> loc(#loc341) + %335 = xten_nn.subgraph (%arg5 = %334: tensor<1x240x1x1xbf16>, %arg6 = %52: tensor<960x240x1x1xbf16>, %arg7 = %51: tensor<960xbf16>) attributes { + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc192) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc192) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_288", + PartOfOutputName = "Conv_288", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc192) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc192) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc192) + } -> tensor<1x960x1x1xbf16> loc(#loc192) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc192) + } -> tensor<1x960x1x1xbf16> loc(#loc192) + %336 = xten_nn.subgraph (%arg5 = %335: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_290", OutputName = "Add_290"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc193) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc193) + } -> tensor<1x960x1x1xbf16> loc(#loc193) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc193) + } -> tensor<1x960x1x1xbf16> loc(#loc193) + %337 = xten_nn.subgraph (%arg5 = %336: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_293", + OutputName = "Clip_293", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + %338 = xten_nn.subgraph (%arg5 = %337: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_295", + OutputName = "Div_295", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + %339 = xten_nn.subgraph (%arg5 = %338: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Generated-#46", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#47", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + %340 = xten_nn.subgraph (%arg5 = %339: tensor<1x960x12x20xbf16>, %arg6 = %331: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_296", + OutputName = "Mul_296", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + %341 = xten_nn.subgraph (%arg5 = %340: tensor<1x960x12x20xbf16>, %arg6 = %50: tensor<160x960x1x1xbf16>, %arg7 = %49: tensor<160xbf16>, %arg8 = %321: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_297", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_297", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_297", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc197) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc197) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc197) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_297", + PartOfOutputName = "Conv_297", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc197) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc197) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc197) + } -> tensor<1x160x12x20xbf16> loc(#loc197) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_298", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_298", OutputName = "Add_298"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc198) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc198) + } -> tensor<1x160x12x20xbf16> loc(#loc198) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc198) + } -> tensor<1x160x12x20xbf16> loc(#loc342) + %342 = xten_nn.subgraph (%arg5 = %341: tensor<1x160x12x20xbf16>, %arg6 = %48: tensor<960x160x1x1xbf16>, %arg7 = %47: tensor<960xbf16>) attributes { + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc199) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc199) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc199) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_299", + PartOfOutputName = "Conv_299", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc199) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc199) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc199) + } -> tensor<1x960x12x20xbf16> loc(#loc199) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc199) + } -> tensor<1x960x12x20xbf16> loc(#loc199) + %343 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_301", OutputName = "Add_301"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc200) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc200) + } -> tensor<1x960x12x20xbf16> loc(#loc200) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc200) + } -> tensor<1x960x12x20xbf16> loc(#loc200) + %344 = xten_nn.subgraph (%arg5 = %343: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_304", + OutputName = "Clip_304", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + %345 = xten_nn.subgraph (%arg5 = %344: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_306", + OutputName = "Div_306", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + %346 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>, %arg6 = %345: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_307", + OutputName = "Mul_307", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + %347 = xten_nn.subgraph (%arg5 = %346: tensor<1x960x12x20xbf16>, %arg6 = %46: tensor<960x1x9x9xbf16>, %arg7 = %45: tensor<960xbf16>) attributes { + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc204) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc204) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc204) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_308", + PartOfOutputName = "Conv_308", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc204) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + %348 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_310", OutputName = "Add_310"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + %349 = xten_nn.subgraph (%arg5 = %348: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_313", + OutputName = "Clip_313", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + %350 = xten_nn.subgraph (%arg5 = %349: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_315", + OutputName = "Div_315", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + %351 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>, %arg6 = %350: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_316", + OutputName = "Mul_316", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + %352 = xten_nn.subgraph (%arg5 = %351: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Generated-#48", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#49", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc343) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc343) + } -> tensor<1x960x1x240xbf16> loc(#loc343) + %353 = xten_nn.subgraph (%arg5 = %352: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#51", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#51", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc209) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc209) + } -> tensor<1x960x1x1xbf16> loc(#loc209) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc209) + } -> tensor<1x960x1x1xbf16> loc(#loc209) + %354 = xten_nn.subgraph (%arg5 = %353: tensor<1x960x1x1xbf16>, %arg6 = %44: tensor<240x960x1x1xbf16>, %arg7 = %43: tensor<240xbf16>) attributes { + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc344) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc344) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_318", + PartOfOutputName = "Conv_318", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc210) + %465 = tosa.clamp %464 { + LayerName = "Relu_319", + OutputName = "Relu_319", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc211) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc344) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc211) + } -> tensor<1x240x1x1xbf16> loc(#loc344) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc344) + } -> tensor<1x240x1x1xbf16> loc(#loc344) + %355 = xten_nn.subgraph (%arg5 = %354: tensor<1x240x1x1xbf16>, %arg6 = %42: tensor<960x240x1x1xbf16>, %arg7 = %41: tensor<960xbf16>) attributes { + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc212) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc212) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_320", + PartOfOutputName = "Conv_320", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc212) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc212) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc212) + } -> tensor<1x960x1x1xbf16> loc(#loc212) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc212) + } -> tensor<1x960x1x1xbf16> loc(#loc212) + %356 = xten_nn.subgraph (%arg5 = %355: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_322", OutputName = "Add_322"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc213) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc213) + } -> tensor<1x960x1x1xbf16> loc(#loc213) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc213) + } -> tensor<1x960x1x1xbf16> loc(#loc213) + %357 = xten_nn.subgraph (%arg5 = %356: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_325", + OutputName = "Clip_325", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + %358 = xten_nn.subgraph (%arg5 = %357: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_327", + OutputName = "Div_327", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + %359 = xten_nn.subgraph (%arg5 = %358: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Generated-#52", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#53", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + %360 = xten_nn.subgraph (%arg5 = %359: tensor<1x960x12x20xbf16>, %arg6 = %351: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_328", + OutputName = "Mul_328", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + %361 = xten_nn.subgraph (%arg5 = %360: tensor<1x960x12x20xbf16>, %arg6 = %40: tensor<160x960x1x1xbf16>, %arg7 = %39: tensor<160xbf16>, %arg8 = %341: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_329", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_329", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_329", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc217) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc217) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc217) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_329", + PartOfOutputName = "Conv_329", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc217) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc217) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc217) + } -> tensor<1x160x12x20xbf16> loc(#loc217) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_330", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_330", OutputName = "Add_330"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc218) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc218) + } -> tensor<1x160x12x20xbf16> loc(#loc218) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc218) + } -> tensor<1x160x12x20xbf16> loc(#loc345) + %362 = xten_nn.subgraph (%arg5 = %361: tensor<1x160x12x20xbf16>, %arg6 = %38: tensor<960x160x1x1xbf16>, %arg7 = %37: tensor<960xbf16>) attributes { + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc219) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc219) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc219) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_331", + PartOfOutputName = "Conv_331", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc219) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc219) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc219) + } -> tensor<1x960x12x20xbf16> loc(#loc219) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc219) + } -> tensor<1x960x12x20xbf16> loc(#loc219) + %363 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_333", OutputName = "Add_333"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc220) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc220) + } -> tensor<1x960x12x20xbf16> loc(#loc220) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc220) + } -> tensor<1x960x12x20xbf16> loc(#loc220) + %364 = xten_nn.subgraph (%arg5 = %363: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_336", + OutputName = "Clip_336", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + %365 = xten_nn.subgraph (%arg5 = %364: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_338", + OutputName = "Div_338", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + %366 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>, %arg6 = %365: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_339", + OutputName = "Mul_339", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + %367 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Generated-#54", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#55", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc346) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc346) + } -> tensor<1x960x1x240xbf16> loc(#loc346) + %368 = xten_nn.subgraph (%arg5 = %367: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#57", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc224) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc224) + } -> tensor<1x960x1x1xbf16> loc(#loc224) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc224) + } -> tensor<1x960x1x1xbf16> loc(#loc224) + %369 = xten_nn.subgraph (%arg5 = %368: tensor<1x960x1x1xbf16>, %arg6 = %36: tensor<128x960x1x1xbf16>, %arg7 = %35: tensor<128xbf16>) attributes { + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<128x960x1x1xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc225) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc225) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_343", + PartOfOutputName = "Conv_343", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x1x1x128xbf16> loc(#loc225) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x128xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc225) + xten_nn.output %465 : tensor<1x128x1x1xbf16> loc(#loc225) + } -> tensor<1x128x1x1xbf16> loc(#loc225) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc225) + } -> tensor<1x128x1x1xbf16> loc(#loc225) + %370 = xten_nn.subgraph (%arg5 = %369: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_344", OutputName = "Sigmoid_344"} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc226) + xten_nn.output %462 : tensor<1x128x1x1xbf16> loc(#loc226) + } -> tensor<1x128x1x1xbf16> loc(#loc226) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc226) + } -> tensor<1x128x1x1xbf16> loc(#loc226) + %371 = xten_nn.subgraph (%arg5 = %370: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Generated-#58", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#59", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 128 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc227) + %372 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>, %arg6 = %34: tensor<128x960x1x1xbf16>, %arg7 = %33: tensor<128xbf16>, %arg8 = %371: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Conv_340", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<128x960x1x1xbf16>, %arg11 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_340", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_341", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc348) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc348) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc348) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_340", + PartOfOutputName = "Conv_340", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc228) + %468 = tosa.clamp %467 { + LayerName = "Relu_341", + OutputName = "Relu_341", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x12x20x128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc229) + %469 = tosa.transpose %468, %463 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc348) + xten_nn.output %469 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc348) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x128x12x20xbf16>, %arg10 = %arg8: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Mul_345", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.mul %arg9, %arg10 { + LayerName = "Mul_345", + OutputName = "Mul_345", + shift = 0 : i8} : (tensor<1x128x12x20xbf16>, tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %463 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc347) + %373 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc230) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc230) + } -> tensor<1x64x12x20xbf16> loc(#loc230) + %374 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc230) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc230) + } -> tensor<1x64x12x20xbf16> loc(#loc230) + %375 = xten_nn.subgraph (%arg5 = %374: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_350", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_350", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_350", + OutputName = "Concat_350", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc231) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc231) + } -> tensor<1x128x12x20xbf16> loc(#loc231) + %376 = xten_nn.subgraph (%arg5 = %375: tensor<1x128x12x20xbf16>, %arg6 = %32: tensor<128x128x3x3xbf16>, %arg7 = %31: tensor<128xbf16>) attributes { + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<128x128x3x3xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<128x128x3x3xbf16>, tensor<4xi32>) -> tensor<128x3x3x128xbf16> loc(#loc232) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc232) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_351", + PartOfOutputName = "Conv_351", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<128x3x3x128xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc232) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc232) + xten_nn.output %467 : tensor<1x128x12x20xbf16> loc(#loc232) + } -> tensor<1x128x12x20xbf16> loc(#loc232) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc232) + } -> tensor<1x128x12x20xbf16> loc(#loc232) + %377 = xten_nn.subgraph (%arg5 = %376: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_352", OutputName = "Sigmoid_352"} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + %378 = xten_nn.subgraph (%arg5 = %377: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc234) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc234) + } -> tensor<1x64x12x20xbf16> loc(#loc234) + %379 = xten_nn.subgraph (%arg5 = %377: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc234) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc234) + } -> tensor<1x64x12x20xbf16> loc(#loc234) + %380 = xten_nn.subgraph (%arg5 = %378: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_354", + OutputName = "Mul_354", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc235) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc235) + } -> tensor<1x64x12x20xbf16> loc(#loc235) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc235) + } -> tensor<1x64x12x20xbf16> loc(#loc235) + %381 = xten_nn.subgraph (%arg5 = %374: tensor<1x64x12x20xbf16>, %arg6 = %380: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_355", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_355", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_355", + OutputName = "Concat_355", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc236) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc236) + } -> tensor<1x128x12x20xbf16> loc(#loc236) + %382 = xten_nn.subgraph (%arg5 = %381: tensor<1x128x12x20xbf16>, %arg6 = %30: tensor<64x128x3x3xbf16>, %arg7 = %29: tensor<64xbf16>) attributes { + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<64x128x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<64x128x3x3xbf16>, tensor<4xi32>) -> tensor<64x3x3x128xbf16> loc(#loc237) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc237) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_356", + PartOfOutputName = "Conv_356", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<64x3x3x128xbf16>, tensor<64xbf16>) -> tensor<1x12x20x64xbf16> loc(#loc237) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x64xbf16>, tensor<4xi32>) -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %467 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + %383 = xten_nn.subgraph (%arg5 = %382: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_357", OutputName = "Tanh_357"} : (tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + %384 = xten_nn.subgraph (%arg5 = %379: tensor<1x64x12x20xbf16>, %arg6 = %383: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_361", + OutputName = "Mul_361", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc239) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc239) + } -> tensor<1x64x12x20xbf16> loc(#loc239) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc239) + } -> tensor<1x64x12x20xbf16> loc(#loc239) + %385 = xten_nn.subgraph (%arg5 = %28: tensor<1x64x12x20xbf16>, %arg6 = %379: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_359", OutputName = "Sub_359"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + %386 = xten_nn.subgraph (%arg5 = %385: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_360", + OutputName = "Mul_360", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + %387 = xten_nn.subgraph (%arg5 = %386: tensor<1x64x12x20xbf16>, %arg6 = %384: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_362", OutputName = "Add_362"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + %388 = xten_nn.subgraph (%arg5 = %373: tensor<1x64x12x20xbf16>, %arg6 = %387: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_363", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_363", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_363", + OutputName = "Concat_363", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc242) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc242) + } -> tensor<1x128x12x20xbf16> loc(#loc242) + %389 = xten_nn.subgraph (%arg5 = %388: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Resize_365", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Resize_365", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 128 : ui32, + config.dim_2 = 12 : ui32, + config.dim_3 = 20 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 24 : ui32, + config.output_W = 40 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_365", + OutputName = "Resize_365", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x24x40xbf16> loc(#loc243) + xten_nn.output %461 : tensor<1x128x24x40xbf16> loc(#loc243) + } -> tensor<1x128x24x40xbf16> loc(#loc243) + %390 = xten_nn.subgraph (%arg5 = %389: tensor<1x128x24x40xbf16>) attributes { + LayerName = "Slice_371", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + OutputName = "Slice_371", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 128 : ui32, + config.dim_h = 24 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 23 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_371", + OutputName = "Slice_371", + size = array, + start = array} : (tensor<1x128x24x40xbf16>) -> tensor<1x128x23x40xbf16> loc(#loc244) + xten_nn.output %461 : tensor<1x128x23x40xbf16> loc(#loc244) + } -> tensor<1x128x23x40xbf16> loc(#loc244) + %391 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>) attributes { + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc12) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc12) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_346", + PartOfOutputName = "AveragePool_346", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>) -> tensor<1x90x160x3xbf16> loc(#loc12) + %466 = tosa.transpose %465, %462 : (tensor<1x90x160x3xbf16>, tensor<4xi32>) -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %466 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %461 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + %392 = xten_nn.subgraph (%arg5 = %391: tensor<1x3x90x160xbf16>) attributes { + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x90x160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc245) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x3xbf16> loc(#loc245) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_347", + PartOfOutputName = "AveragePool_347", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x90x160x3xbf16>) -> tensor<1x45x80x3xbf16> loc(#loc245) + %466 = tosa.transpose %465, %462 : (tensor<1x45x80x3xbf16>, tensor<4xi32>) -> tensor<1x3x45x80xbf16> loc(#loc245) + xten_nn.output %466 : tensor<1x3x45x80xbf16> loc(#loc245) + } -> tensor<1x3x45x80xbf16> loc(#loc245) + xten_nn.output %461 : tensor<1x3x45x80xbf16> loc(#loc245) + } -> tensor<1x3x45x80xbf16> loc(#loc245) + %393 = xten_nn.subgraph (%arg5 = %392: tensor<1x3x45x80xbf16>) attributes { + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x45x80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 1], [0, 0]], + HWPaddingNotCounted = [[0, 1], [0, 0]], + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc246) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x3xbf16> loc(#loc246) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_348", + PartOfOutputName = "AveragePool_348", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x45x80x3xbf16>) -> tensor<1x23x40x3xbf16> loc(#loc246) + %466 = tosa.transpose %465, %462 : (tensor<1x23x40x3xbf16>, tensor<4xi32>) -> tensor<1x3x23x40xbf16> loc(#loc246) + xten_nn.output %466 : tensor<1x3x23x40xbf16> loc(#loc246) + } -> tensor<1x3x23x40xbf16> loc(#loc246) + xten_nn.output %461 : tensor<1x3x23x40xbf16> loc(#loc246) + } -> tensor<1x3x23x40xbf16> loc(#loc246) + %394 = xten_nn.subgraph (%arg5 = %390: tensor<1x128x23x40xbf16>, %arg6 = %217: tensor<1x40x23x40xbf16>, %arg7 = %393: tensor<1x3x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_372", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_372", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_372", + OutputName = "Concat_372", + axis = 1 : i32} : (tensor<1x128x23x40xbf16>, tensor<1x40x23x40xbf16>, tensor<1x3x23x40xbf16>) -> tensor<1x171x23x40xbf16> loc(#loc247) + xten_nn.output %461 : tensor<1x171x23x40xbf16> loc(#loc247) + } -> tensor<1x171x23x40xbf16> loc(#loc247) + %395 = xten_nn.subgraph (%arg5 = %394: tensor<1x171x23x40xbf16>, %arg6 = %27: tensor<80x171x3x3xbf16>, %arg7 = %26: tensor<80xbf16>) attributes { + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x171x23x40xbf16>, %arg9 = %arg6: tensor<80x171x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x171x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x171xbf16> loc(#loc349) + %465 = tosa.transpose %arg8, %463 : (tensor<1x171x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x171xbf16> loc(#loc349) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_373", + PartOfOutputName = "Conv_373", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x171xbf16>, tensor<80x3x3x171xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc248) + %467 = tosa.clamp %466 { + LayerName = "Relu_374", + OutputName = "Relu_374", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc249) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc349) + xten_nn.output %468 : tensor<1x80x23x40xbf16> loc(#loc249) + } -> tensor<1x80x23x40xbf16> loc(#loc349) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc349) + } -> tensor<1x80x23x40xbf16> loc(#loc349) + %396 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %397 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %398 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_376", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_376", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_376", + OutputName = "Concat_376", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc251) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc251) + } -> tensor<1x80x23x40xbf16> loc(#loc251) + %399 = xten_nn.subgraph (%arg5 = %398: tensor<1x80x23x40xbf16>, %arg6 = %25: tensor<80x80x3x3xbf16>, %arg7 = %24: tensor<80xbf16>) attributes { + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<80x80x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x80x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x80xbf16> loc(#loc252) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_377", + PartOfOutputName = "Conv_377", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<80x3x3x80xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %467 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + %400 = xten_nn.subgraph (%arg5 = %399: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_378", OutputName = "Sigmoid_378"} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %462 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + %401 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %402 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %403 = xten_nn.subgraph (%arg5 = %401: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_380", + OutputName = "Mul_380", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + %404 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %403: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_381", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_381", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_381", + OutputName = "Concat_381", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc256) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc256) + } -> tensor<1x80x23x40xbf16> loc(#loc256) + %405 = xten_nn.subgraph (%arg5 = %404: tensor<1x80x23x40xbf16>, %arg6 = %23: tensor<40x80x3x3xbf16>, %arg7 = %22: tensor<40xbf16>) attributes { + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<40x80x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x80x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x80xbf16> loc(#loc257) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc257) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_382", + PartOfOutputName = "Conv_382", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<40x3x3x80xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc257) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc257) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc257) + } -> tensor<1x40x23x40xbf16> loc(#loc257) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc257) + } -> tensor<1x40x23x40xbf16> loc(#loc257) + %406 = xten_nn.subgraph (%arg5 = %405: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_383", OutputName = "Tanh_383"} : (tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + %407 = xten_nn.subgraph (%arg5 = %402: tensor<1x40x23x40xbf16>, %arg6 = %406: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_387", + OutputName = "Mul_387", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + %408 = xten_nn.subgraph (%arg5 = %21: tensor<1x40x23x40xbf16>, %arg6 = %402: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_385", OutputName = "Sub_385"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + %409 = xten_nn.subgraph (%arg5 = %408: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_386", + OutputName = "Mul_386", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + %410 = xten_nn.subgraph (%arg5 = %409: tensor<1x40x23x40xbf16>, %arg6 = %407: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_388", OutputName = "Add_388"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + %411 = xten_nn.subgraph (%arg5 = %396: tensor<1x40x23x40xbf16>, %arg6 = %410: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_389", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_389", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_389", + OutputName = "Concat_389", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc262) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc262) + } -> tensor<1x80x23x40xbf16> loc(#loc262) + %412 = xten_nn.subgraph (%arg5 = %411: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Resize_391", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Resize_391", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 80 : ui32, + config.dim_2 = 23 : ui32, + config.dim_3 = 40 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 46 : ui32, + config.output_W = 80 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_391", + OutputName = "Resize_391", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x46x80xbf16> loc(#loc263) + xten_nn.output %461 : tensor<1x80x46x80xbf16> loc(#loc263) + } -> tensor<1x80x46x80xbf16> loc(#loc263) + %413 = xten_nn.subgraph (%arg5 = %412: tensor<1x80x46x80xbf16>) attributes { + LayerName = "Slice_397", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + OutputName = "Slice_397", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 80 : ui32, + config.dim_h = 46 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 45 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_397", + OutputName = "Slice_397", + size = array, + start = array} : (tensor<1x80x46x80xbf16>) -> tensor<1x80x45x80xbf16> loc(#loc264) + xten_nn.output %461 : tensor<1x80x45x80xbf16> loc(#loc264) + } -> tensor<1x80x45x80xbf16> loc(#loc264) + %414 = xten_nn.subgraph (%arg5 = %413: tensor<1x80x45x80xbf16>, %arg6 = %181: tensor<1x24x45x80xbf16>, %arg7 = %392: tensor<1x3x45x80xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_398", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_398", + OutputName = "Concat_398", + axis = 1 : i32} : (tensor<1x80x45x80xbf16>, tensor<1x24x45x80xbf16>, tensor<1x3x45x80xbf16>) -> tensor<1x107x45x80xbf16> loc(#loc265) + xten_nn.output %461 : tensor<1x107x45x80xbf16> loc(#loc265) + } -> tensor<1x107x45x80xbf16> loc(#loc265) + %415 = xten_nn.subgraph (%arg5 = %414: tensor<1x107x45x80xbf16>, %arg6 = %20: tensor<40x107x3x3xbf16>, %arg7 = %19: tensor<40xbf16>) attributes { + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x107x45x80xbf16>, %arg9 = %arg6: tensor<40x107x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x107x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x107xbf16> loc(#loc350) + %465 = tosa.transpose %arg8, %463 : (tensor<1x107x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x107xbf16> loc(#loc350) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_399", + PartOfOutputName = "Conv_399", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x107xbf16>, tensor<40x3x3x107xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc266) + %467 = tosa.clamp %466 { + LayerName = "Relu_400", + OutputName = "Relu_400", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc267) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc350) + xten_nn.output %468 : tensor<1x40x45x80xbf16> loc(#loc267) + } -> tensor<1x40x45x80xbf16> loc(#loc350) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc350) + } -> tensor<1x40x45x80xbf16> loc(#loc350) + %416 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %417 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %418 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_402", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_402", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_402", + OutputName = "Concat_402", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc269) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc269) + } -> tensor<1x40x45x80xbf16> loc(#loc269) + %419 = xten_nn.subgraph (%arg5 = %418: tensor<1x40x45x80xbf16>, %arg6 = %18: tensor<40x40x3x3xbf16>, %arg7 = %17: tensor<40xbf16>) attributes { + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<40x40x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x40x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x40xbf16> loc(#loc270) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_403", + PartOfOutputName = "Conv_403", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<40x3x3x40xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %467 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + %420 = xten_nn.subgraph (%arg5 = %419: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_404", OutputName = "Sigmoid_404"} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %462 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + %421 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %422 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %423 = xten_nn.subgraph (%arg5 = %421: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_406", + OutputName = "Mul_406", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + %424 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %423: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_407", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_407", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_407", + OutputName = "Concat_407", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc274) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc274) + } -> tensor<1x40x45x80xbf16> loc(#loc274) + %425 = xten_nn.subgraph (%arg5 = %424: tensor<1x40x45x80xbf16>, %arg6 = %16: tensor<20x40x3x3xbf16>, %arg7 = %15: tensor<20xbf16>) attributes { + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<20x40x3x3xbf16>, %arg10 = %arg7: tensor<20xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<20x40x3x3xbf16>, tensor<4xi32>) -> tensor<20x3x3x40xbf16> loc(#loc275) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc275) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_408", + PartOfOutputName = "Conv_408", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<20x3x3x40xbf16>, tensor<20xbf16>) -> tensor<1x45x80x20xbf16> loc(#loc275) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x20xbf16>, tensor<4xi32>) -> tensor<1x20x45x80xbf16> loc(#loc275) + xten_nn.output %467 : tensor<1x20x45x80xbf16> loc(#loc275) + } -> tensor<1x20x45x80xbf16> loc(#loc275) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc275) + } -> tensor<1x20x45x80xbf16> loc(#loc275) + %426 = xten_nn.subgraph (%arg5 = %425: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_409", OutputName = "Tanh_409"} : (tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + %427 = xten_nn.subgraph (%arg5 = %422: tensor<1x20x45x80xbf16>, %arg6 = %426: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_413", + OutputName = "Mul_413", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + %428 = xten_nn.subgraph (%arg5 = %14: tensor<1x20x45x80xbf16>, %arg6 = %422: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_411", OutputName = "Sub_411"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + %429 = xten_nn.subgraph (%arg5 = %428: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_412", + OutputName = "Mul_412", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + %430 = xten_nn.subgraph (%arg5 = %429: tensor<1x20x45x80xbf16>, %arg6 = %427: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_414", OutputName = "Add_414"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + %431 = xten_nn.subgraph (%arg5 = %416: tensor<1x20x45x80xbf16>, %arg6 = %430: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_415", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_415", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_415", + OutputName = "Concat_415", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc280) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc280) + } -> tensor<1x40x45x80xbf16> loc(#loc280) + %432 = xten_nn.subgraph (%arg5 = %431: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Resize_417", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Resize_417", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 40 : ui32, + config.dim_2 = 45 : ui32, + config.dim_3 = 80 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 90 : ui32, + config.output_W = 160 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_417", + OutputName = "Resize_417", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x90x160xbf16> loc(#loc281) + xten_nn.output %461 : tensor<1x40x90x160xbf16> loc(#loc281) + } -> tensor<1x40x90x160xbf16> loc(#loc281) + %433 = xten_nn.subgraph (%arg5 = %432: tensor<1x40x90x160xbf16>, %arg6 = %175: tensor<1x16x90x160xbf16>, %arg7 = %391: tensor<1x3x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_418", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_418", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_418", + OutputName = "Concat_418", + axis = 1 : i32} : (tensor<1x40x90x160xbf16>, tensor<1x16x90x160xbf16>, tensor<1x3x90x160xbf16>) -> tensor<1x59x90x160xbf16> loc(#loc282) + xten_nn.output %461 : tensor<1x59x90x160xbf16> loc(#loc282) + } -> tensor<1x59x90x160xbf16> loc(#loc282) + %434 = xten_nn.subgraph (%arg5 = %433: tensor<1x59x90x160xbf16>, %arg6 = %13: tensor<32x59x3x3xbf16>, %arg7 = %12: tensor<32xbf16>) attributes { + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x59x90x160xbf16>, %arg9 = %arg6: tensor<32x59x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x59x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x59xbf16> loc(#loc351) + %465 = tosa.transpose %arg8, %463 : (tensor<1x59x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x59xbf16> loc(#loc351) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_419", + PartOfOutputName = "Conv_419", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x59xbf16>, tensor<32x3x3x59xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc283) + %467 = tosa.clamp %466 { + LayerName = "Relu_420", + OutputName = "Relu_420", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc284) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc351) + xten_nn.output %468 : tensor<1x32x90x160xbf16> loc(#loc284) + } -> tensor<1x32x90x160xbf16> loc(#loc351) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc351) + } -> tensor<1x32x90x160xbf16> loc(#loc351) + %435 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %436 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %437 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_422", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_422", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_422", + OutputName = "Concat_422", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc286) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc286) + } -> tensor<1x32x90x160xbf16> loc(#loc286) + %438 = xten_nn.subgraph (%arg5 = %437: tensor<1x32x90x160xbf16>, %arg6 = %11: tensor<32x32x3x3xbf16>, %arg7 = %10: tensor<32xbf16>) attributes { + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<32x32x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x32x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x32xbf16> loc(#loc287) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_423", + PartOfOutputName = "Conv_423", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<32x3x3x32xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %467 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + %439 = xten_nn.subgraph (%arg5 = %438: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_424", OutputName = "Sigmoid_424"} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %462 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + %440 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %441 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %442 = xten_nn.subgraph (%arg5 = %440: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_426", + OutputName = "Mul_426", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + %443 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %442: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_427", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_427", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_427", + OutputName = "Concat_427", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc291) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc291) + } -> tensor<1x32x90x160xbf16> loc(#loc291) + %444 = xten_nn.subgraph (%arg5 = %443: tensor<1x32x90x160xbf16>, %arg6 = %9: tensor<16x32x3x3xbf16>, %arg7 = %8: tensor<16xbf16>) attributes { + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<16x32x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x32x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x32xbf16> loc(#loc292) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc292) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_428", + PartOfOutputName = "Conv_428", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<16x3x3x32xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc292) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc292) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc292) + } -> tensor<1x16x90x160xbf16> loc(#loc292) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc292) + } -> tensor<1x16x90x160xbf16> loc(#loc292) + %445 = xten_nn.subgraph (%arg5 = %444: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_429", OutputName = "Tanh_429"} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + %446 = xten_nn.subgraph (%arg5 = %441: tensor<1x16x90x160xbf16>, %arg6 = %445: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_433", + OutputName = "Mul_433", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + %447 = xten_nn.subgraph (%arg5 = %7: tensor<1x16x90x160xbf16>, %arg6 = %441: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_431", OutputName = "Sub_431"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + %448 = xten_nn.subgraph (%arg5 = %447: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_432", + OutputName = "Mul_432", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + %449 = xten_nn.subgraph (%arg5 = %448: tensor<1x16x90x160xbf16>, %arg6 = %446: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_434", OutputName = "Add_434"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + %450 = xten_nn.subgraph (%arg5 = %435: tensor<1x16x90x160xbf16>, %arg6 = %449: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_435", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_435", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_435", + OutputName = "Concat_435", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc297) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc297) + } -> tensor<1x32x90x160xbf16> loc(#loc297) + %451 = xten_nn.subgraph (%arg5 = %450: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Resize_437", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Resize_437", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 32 : ui32, + config.dim_2 = 90 : ui32, + config.dim_3 = 160 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 180 : ui32, + config.output_W = 320 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_437", + OutputName = "Resize_437", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x180x320xbf16> loc(#loc298) + xten_nn.output %461 : tensor<1x32x180x320xbf16> loc(#loc298) + } -> tensor<1x32x180x320xbf16> loc(#loc298) + %452 = xten_nn.subgraph (%arg5 = %451: tensor<1x32x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_438", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Concat_438", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_438", + OutputName = "Concat_438", + axis = 1 : i32} : (tensor<1x32x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x35x180x320xbf16> loc(#loc299) + xten_nn.output %461 : tensor<1x35x180x320xbf16> loc(#loc299) + } -> tensor<1x35x180x320xbf16> loc(#loc299) + %453 = xten_nn.subgraph (%arg5 = %452: tensor<1x35x180x320xbf16>, %arg6 = %6: tensor<16x35x3x3xbf16>, %arg7 = %5: tensor<16xbf16>) attributes { + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x35x180x320xbf16>, %arg9 = %arg6: tensor<16x35x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x35x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x35xbf16> loc(#loc352) + %465 = tosa.transpose %arg8, %463 : (tensor<1x35x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x35xbf16> loc(#loc352) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_439", + PartOfOutputName = "Conv_439", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x35xbf16>, tensor<16x3x3x35xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc300) + %467 = tosa.clamp %466 { + LayerName = "Relu_440", + OutputName = "Relu_440", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc301) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc352) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc301) + } -> tensor<1x16x180x320xbf16> loc(#loc352) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc352) + } -> tensor<1x16x180x320xbf16> loc(#loc352) + %454 = xten_nn.subgraph (%arg5 = %453: tensor<1x16x180x320xbf16>, %arg6 = %4: tensor<16x16x3x3xbf16>, %arg7 = %3: tensor<16xbf16>) attributes { + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<16x16x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x16x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x16xbf16> loc(#loc353) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc353) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_441", + PartOfOutputName = "Conv_441", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<16x3x3x16xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc302) + %467 = tosa.clamp %466 { + LayerName = "Relu_442", + OutputName = "Relu_442", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc303) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc353) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc303) + } -> tensor<1x16x180x320xbf16> loc(#loc353) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc353) + } -> tensor<1x16x180x320xbf16> loc(#loc353) + %455 = xten_nn.subgraph (%arg5 = %454: tensor<1x16x180x320xbf16>, %arg6 = %2: tensor<4x16x1x1xbf16>, %arg7 = %1: tensor<4xbf16>) attributes { + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<4x16x1x1xbf16>, %arg10 = %arg7: tensor<4xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc304) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<4x16x1x1xbf16>) -> tensor<4x1x1x16xbf16> loc(#loc304) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc304) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_443", + PartOfOutputName = "Conv_443", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<4x1x1x16xbf16>, tensor<4xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc304) + %467 = tosa.transpose %466, %462 : (tensor<1x180x320x4xbf16>, tensor<4xi32>) -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %467 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %461 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + %456 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc305) + } -> tensor<1x3x180x320xbf16> loc(#loc305) + %457 = xten_nn.subgraph (%arg5 = %456: tensor<1x3x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_445", OutputName = "Add_445"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + %458 = xten_nn.subgraph (%arg5 = %457: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_446", + OutputName = "Clip_446", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + %459 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 4 : ui32, + config.start = 3 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc305) + } -> tensor<1x1x180x320xbf16> loc(#loc305) + %460 = xten_nn.subgraph (%arg5 = %459: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_447", + OutputName = "Clip_447", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %462 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + return %449, %430, %410, %387, %458, %460 : tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16> loc(#loc308) + } loc(#loc308) +} loc(#loc) +#loc1 = loc("Div_2") +#loc2 = loc("Sub_431") +#loc3 = loc("Sub_411") +#loc4 = loc("Sub_385") +#loc5 = loc("Sub_359") +#loc6 = loc("Div_16") +#loc7 = loc("Sub_14") +#loc8 = loc("Initializer_398") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("Conv_17") +#loc14 = loc("Add_19") +#loc15 = loc("Clip_22") +#loc16 = loc("Div_24") +#loc17 = loc("Mul_25") +#loc18 = loc("Conv_26") +#loc19 = loc("Relu_27") +#loc20 = loc("Conv_28") +#loc21 = loc("Add_29") +#loc22 = loc("Conv_30") +#loc23 = loc("Relu_31") +#loc24 = loc("Conv_32") +#loc25 = loc("Relu_33") +#loc26 = loc("Conv_34") +#loc27 = loc("Conv_35") +#loc28 = loc("Relu_36") +#loc29 = loc("Conv_37") +#loc30 = loc("Relu_38") +#loc31 = loc("Conv_39") +#loc32 = loc("Add_40") +#loc33 = loc("Conv_41") +#loc34 = loc("Relu_42") +#loc35 = loc("Conv_43") +#loc36 = loc("Relu_44") +#loc37 = loc("GlobalAveragePool_45") +#loc38 = loc("Conv_46") +#loc39 = loc("Relu_47") +#loc40 = loc("Conv_48") +#loc41 = loc("Add_50") +#loc42 = loc("Clip_53") +#loc43 = loc("Div_55") +#loc44 = loc("Mul_56") +#loc45 = loc("Conv_57") +#loc46 = loc("Conv_58") +#loc47 = loc("Relu_59") +#loc48 = loc("Conv_60") +#loc49 = loc("Relu_61") +#loc50 = loc("GlobalAveragePool_62") +#loc51 = loc("Conv_63") +#loc52 = loc("Relu_64") +#loc53 = loc("Conv_65") +#loc54 = loc("Add_67") +#loc55 = loc("Clip_70") +#loc56 = loc("Div_72") +#loc57 = loc("Mul_73") +#loc58 = loc("Conv_74") +#loc59 = loc("Add_75") +#loc60 = loc("Conv_76") +#loc61 = loc("Relu_77") +#loc62 = loc("Conv_78") +#loc63 = loc("Relu_79") +#loc64 = loc("GlobalAveragePool_80") +#loc65 = loc("Conv_81") +#loc66 = loc("Relu_82") +#loc67 = loc("Conv_83") +#loc68 = loc("Add_85") +#loc69 = loc("Clip_88") +#loc70 = loc("Div_90") +#loc71 = loc("Mul_91") +#loc72 = loc("Conv_92") +#loc73 = loc("Add_93") +#loc74 = loc("Conv_94") +#loc75 = loc("Add_96") +#loc76 = loc("Clip_99") +#loc77 = loc("Div_101") +#loc78 = loc("Mul_102") +#loc79 = loc("Conv_103") +#loc80 = loc("Add_105") +#loc81 = loc("Clip_108") +#loc82 = loc("Div_110") +#loc83 = loc("Mul_111") +#loc84 = loc("Conv_112") +#loc85 = loc("Conv_113") +#loc86 = loc("Add_115") +#loc87 = loc("Clip_118") +#loc88 = loc("Div_120") +#loc89 = loc("Mul_121") +#loc90 = loc("Conv_122") +#loc91 = loc("Add_124") +#loc92 = loc("Clip_127") +#loc93 = loc("Div_129") +#loc94 = loc("Mul_130") +#loc95 = loc("Conv_131") +#loc96 = loc("Add_132") +#loc97 = loc("Conv_133") +#loc98 = loc("Add_135") +#loc99 = loc("Clip_138") +#loc100 = loc("Div_140") +#loc101 = loc("Mul_141") +#loc102 = loc("Conv_142") +#loc103 = loc("Add_144") +#loc104 = loc("Clip_147") +#loc105 = loc("Div_149") +#loc106 = loc("Mul_150") +#loc107 = loc("Conv_151") +#loc108 = loc("Add_152") +#loc109 = loc("Conv_153") +#loc110 = loc("Add_155") +#loc111 = loc("Clip_158") +#loc112 = loc("Div_160") +#loc113 = loc("Mul_161") +#loc114 = loc("Conv_162") +#loc115 = loc("Add_164") +#loc116 = loc("Clip_167") +#loc117 = loc("Div_169") +#loc118 = loc("Mul_170") +#loc119 = loc("Conv_171") +#loc120 = loc("Add_172") +#loc121 = loc("Conv_173") +#loc122 = loc("Add_175") +#loc123 = loc("Clip_178") +#loc124 = loc("Div_180") +#loc125 = loc("Mul_181") +#loc126 = loc("Conv_182") +#loc127 = loc("Add_184") +#loc128 = loc("Clip_187") +#loc129 = loc("Div_189") +#loc130 = loc("Mul_190") +#loc131 = loc("GlobalAveragePool_191") +#loc132 = loc("Conv_192") +#loc133 = loc("Relu_193") +#loc134 = loc("Conv_194") +#loc135 = loc("Add_196") +#loc136 = loc("Clip_199") +#loc137 = loc("Div_201") +#loc138 = loc("Mul_202") +#loc139 = loc("Conv_203") +#loc140 = loc("Conv_204") +#loc141 = loc("Add_206") +#loc142 = loc("Clip_209") +#loc143 = loc("Div_211") +#loc144 = loc("Mul_212") +#loc145 = loc("Conv_213") +#loc146 = loc("Add_215") +#loc147 = loc("Clip_218") +#loc148 = loc("Div_220") +#loc149 = loc("Mul_221") +#loc150 = loc("GlobalAveragePool_222") +#loc151 = loc("Conv_223") +#loc152 = loc("Relu_224") +#loc153 = loc("Conv_225") +#loc154 = loc("Add_227") +#loc155 = loc("Clip_230") +#loc156 = loc("Div_232") +#loc157 = loc("Mul_233") +#loc158 = loc("Conv_234") +#loc159 = loc("Add_235") +#loc160 = loc("Conv_236") +#loc161 = loc("Add_238") +#loc162 = loc("Clip_241") +#loc163 = loc("Div_243") +#loc164 = loc("Mul_244") +#loc165 = loc("Conv_245") +#loc166 = loc("Add_247") +#loc167 = loc("Clip_250") +#loc168 = loc("Div_252") +#loc169 = loc("Mul_253") +#loc170 = loc("GlobalAveragePool_254") +#loc171 = loc("Conv_255") +#loc172 = loc("Relu_256") +#loc173 = loc("Conv_257") +#loc174 = loc("Add_259") +#loc175 = loc("Clip_262") +#loc176 = loc("Div_264") +#loc177 = loc("Mul_265") +#loc178 = loc("Conv_266") +#loc179 = loc("Conv_267") +#loc180 = loc("Add_269") +#loc181 = loc("Clip_272") +#loc182 = loc("Div_274") +#loc183 = loc("Mul_275") +#loc184 = loc("Conv_276") +#loc185 = loc("Add_278") +#loc186 = loc("Clip_281") +#loc187 = loc("Div_283") +#loc188 = loc("Mul_284") +#loc189 = loc("GlobalAveragePool_285") +#loc190 = loc("Conv_286") +#loc191 = loc("Relu_287") +#loc192 = loc("Conv_288") +#loc193 = loc("Add_290") +#loc194 = loc("Clip_293") +#loc195 = loc("Div_295") +#loc196 = loc("Mul_296") +#loc197 = loc("Conv_297") +#loc198 = loc("Add_298") +#loc199 = loc("Conv_299") +#loc200 = loc("Add_301") +#loc201 = loc("Clip_304") +#loc202 = loc("Div_306") +#loc203 = loc("Mul_307") +#loc204 = loc("Conv_308") +#loc205 = loc("Add_310") +#loc206 = loc("Clip_313") +#loc207 = loc("Div_315") +#loc208 = loc("Mul_316") +#loc209 = loc("GlobalAveragePool_317") +#loc210 = loc("Conv_318") +#loc211 = loc("Relu_319") +#loc212 = loc("Conv_320") +#loc213 = loc("Add_322") +#loc214 = loc("Clip_325") +#loc215 = loc("Div_327") +#loc216 = loc("Mul_328") +#loc217 = loc("Conv_329") +#loc218 = loc("Add_330") +#loc219 = loc("Conv_331") +#loc220 = loc("Add_333") +#loc221 = loc("Clip_336") +#loc222 = loc("Div_338") +#loc223 = loc("Mul_339") +#loc224 = loc("GlobalAveragePool_342") +#loc225 = loc("Conv_343") +#loc226 = loc("Sigmoid_344") +#loc227 = loc("Mul_345") +#loc228 = loc("Conv_340") +#loc229 = loc("Relu_341") +#loc230 = loc("Split_349") +#loc231 = loc("Concat_350") +#loc232 = loc("Conv_351") +#loc233 = loc("Sigmoid_352") +#loc234 = loc("Split_353") +#loc235 = loc("Mul_354") +#loc236 = loc("Concat_355") +#loc237 = loc("Conv_356") +#loc238 = loc("Tanh_357") +#loc239 = loc("Mul_361") +#loc240 = loc("Mul_360") +#loc241 = loc("Add_362") +#loc242 = loc("Concat_363") +#loc243 = loc("Resize_365") +#loc244 = loc("Slice_371") +#loc245 = loc("AveragePool_347") +#loc246 = loc("AveragePool_348") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_380") +#loc256 = loc("Concat_381") +#loc257 = loc("Conv_382") +#loc258 = loc("Tanh_383") +#loc259 = loc("Mul_387") +#loc260 = loc("Mul_386") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_406") +#loc274 = loc("Concat_407") +#loc275 = loc("Conv_408") +#loc276 = loc("Tanh_409") +#loc277 = loc("Mul_413") +#loc278 = loc("Mul_412") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_426") +#loc291 = loc("Concat_427") +#loc292 = loc("Conv_428") +#loc293 = loc("Tanh_429") +#loc294 = loc("Mul_433") +#loc295 = loc("Mul_432") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_446") +#loc307 = loc("Clip_447") +#loc308 = loc(fused[#loc1, #loc2, #loc3, #loc4, #loc5, #loc6, #loc7, #loc8, #loc9, #loc10, #loc11, #loc12, #loc13, #loc14, #loc15, #loc16, #loc17, #loc18, #loc19, #loc20, #loc21, #loc22, #loc23, #loc24, #loc25, #loc26, #loc27, #loc28, #loc29, #loc30, #loc31, #loc32, #loc33, #loc34, #loc35, #loc36, #loc37, #loc38, #loc39, #loc40, #loc41, #loc42, #loc43, #loc44, #loc45, #loc46, #loc47, #loc48, #loc49, #loc50, #loc51, #loc52, #loc53, #loc54, #loc55, #loc56, #loc57, #loc58, #loc59, #loc60, #loc61, #loc62, #loc63, #loc64, #loc65, #loc66, #loc67, #loc68, #loc69, #loc70, #loc71, #loc72, #loc73, #loc74, #loc75, #loc76, #loc77, #loc78, #loc79, #loc80, #loc81, #loc82, #loc83, #loc84, #loc85, #loc86, #loc87, #loc88, #loc89, #loc90, #loc91, #loc92, #loc93, #loc94, #loc95, #loc96, #loc97, #loc98, #loc99, #loc100, #loc101, #loc102, #loc103, #loc104, #loc105, #loc106, #loc107, #loc108, #loc109, #loc110, #loc111, #loc112, #loc113, #loc114, #loc115, #loc116, #loc117, #loc118, #loc119, #loc120, #loc121, #loc122, #loc123, #loc124, #loc125, #loc126, #loc127, #loc128, #loc129, #loc130, #loc131, #loc132, #loc133, #loc134, #loc135, #loc136, #loc137, #loc138, #loc139, #loc140, #loc141, #loc142, #loc143, #loc144, #loc145, #loc146, #loc147, #loc148, #loc149, #loc150, #loc151, #loc152, #loc153, #loc154, #loc155, #loc156, #loc157, #loc158, #loc159, #loc160, #loc161, #loc162, #loc163, #loc164, #loc165, #loc166, #loc167, #loc168, #loc169, #loc170, #loc171, #loc172, #loc173, #loc174, #loc175, #loc176, #loc177, #loc178, #loc179, #loc180, #loc181, #loc182, #loc183, #loc184, #loc185, #loc186, #loc187, #loc188, #loc189, #loc190, #loc191, #loc192, #loc193, #loc194, #loc195, #loc196, #loc197, #loc198, #loc199, #loc200, #loc201, #loc202, #loc203, #loc204, #loc205, #loc206, #loc207, #loc208, #loc209, #loc210, #loc211, #loc212, #loc213, #loc214, #loc215, #loc216, #loc217, #loc218, #loc219, #loc220, #loc221, #loc222, #loc223, #loc224, #loc225, #loc226, #loc227, #loc228, #loc229, #loc230, #loc231, #loc232, #loc233, #loc234, #loc235, #loc236, #loc237, #loc238, #loc239, #loc240, #loc241, #loc242, #loc243, #loc244, #loc245, #loc246, #loc247, #loc248, #loc249, #loc250, #loc251, #loc252, #loc253, #loc254, #loc255, #loc256, #loc257, #loc258, #loc259, #loc260, #loc261, #loc262, #loc263, #loc264, #loc265, #loc266, #loc267, #loc268, #loc269, #loc270, #loc271, #loc272, #loc273, #loc274, #loc275, #loc276, #loc277, #loc278, #loc279, #loc280, #loc281, #loc282, #loc283, #loc284, #loc285, #loc286, #loc287, #loc288, #loc289, #loc290, #loc291, #loc292, #loc293, #loc294, #loc295, #loc296, #loc297, #loc298, #loc299, #loc300, #loc301, #loc302, #loc303, #loc304, #loc305, #loc306, #loc307]) +#loc309 = loc(fused[#loc7, #loc8]) +#loc310 = loc(fused[#loc11, #loc9, #loc12]) +#loc311 = loc(fused[#loc9, #loc12, #loc11]) +#loc312 = loc(fused[#loc18, #loc19]) +#loc313 = loc(fused[#loc20, #loc21]) +#loc314 = loc(fused[#loc22, #loc23]) +#loc315 = loc(fused[#loc24, #loc25]) +#loc316 = loc(fused[#loc27, #loc28]) +#loc317 = loc(fused[#loc29, #loc30]) +#loc318 = loc(fused[#loc31, #loc32]) +#loc319 = loc(fused[#loc33, #loc34]) +#loc320 = loc(fused[#loc35, #loc36]) +#loc321 = loc(fused[#loc38, #loc39]) +#loc322 = loc(fused[#loc46, #loc47]) +#loc323 = loc(fused[#loc48, #loc49]) +#loc324 = loc(fused[#loc51, #loc52]) +#loc325 = loc(fused[#loc58, #loc59]) +#loc326 = loc(fused[#loc60, #loc61]) +#loc327 = loc(fused[#loc62, #loc63]) +#loc328 = loc(fused[#loc65, #loc66]) +#loc329 = loc(fused[#loc72, #loc73]) +#loc330 = loc(fused[#loc95, #loc96]) +#loc331 = loc(fused[#loc107, #loc108]) +#loc332 = loc(fused[#loc119, #loc120]) +#loc333 = loc(fused[#loc130, #loc131]) +#loc334 = loc(fused[#loc132, #loc133]) +#loc335 = loc(fused[#loc149, #loc150]) +#loc336 = loc(fused[#loc151, #loc152]) +#loc337 = loc(fused[#loc158, #loc159]) +#loc338 = loc(fused[#loc169, #loc170]) +#loc339 = loc(fused[#loc171, #loc172]) +#loc340 = loc(fused[#loc188, #loc189]) +#loc341 = loc(fused[#loc190, #loc191]) +#loc342 = loc(fused[#loc197, #loc198]) +#loc343 = loc(fused[#loc208, #loc209]) +#loc344 = loc(fused[#loc210, #loc211]) +#loc345 = loc(fused[#loc217, #loc218]) +#loc346 = loc(fused[#loc223, #loc224]) +#loc347 = loc(fused[#loc228, #loc229, #loc227]) +#loc348 = loc(fused[#loc228, #loc229]) +#loc349 = loc(fused[#loc248, #loc249]) +#loc350 = loc(fused[#loc266, #loc267]) +#loc351 = loc(fused[#loc283, #loc284]) +#loc352 = loc(fused[#loc300, #loc301]) +#loc353 = loc(fused[#loc302, #loc303]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/par.subgraph.dse.mlir b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/par.subgraph.dse.mlir new file mode 100644 index 0000000000000000000000000000000000000000..0a096d7373fd3c238b2f91db26bc253f612c5b9f --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/par.subgraph.dse.mlir @@ -0,0 +1,32555 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func private @forward(%arg0: tensor<1x180x320x4xbf16> loc(unknown), %arg1: tensor<1x16x90x160xbf16> loc(unknown), %arg2: tensor<1x20x45x80xbf16> loc(unknown), %arg3: tensor<1x40x23x40xbf16> loc(unknown), %arg4: tensor<1x64x12x20xbf16> loc(unknown)) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) attributes { + max_heap_size = 2240 : ui32, + max_stack_size = 2368 : ui32, + stack_heap_start_address = 45696 : ui32, + total_stack_heap_region_size = 6912 : ui32} { + %0 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/biases"} -> tensor<4xbf16> loc(#loc) + %1 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/weights"} -> tensor<4x16x1x1xbf16> loc(#loc) + %2 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/biases"} -> tensor<16xbf16> loc(#loc) + %3 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/weights"} -> tensor<16x16x3x3xbf16> loc(#loc) + %4 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/biases"} -> tensor<16xbf16> loc(#loc) + %5 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/weights"} -> tensor<16x35x3x3xbf16> loc(#loc) + %6 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_431/Constant_0_0"} -> tensor<1x16x90x160xbf16> loc(#loc2) + %7 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/biases"} -> tensor<16xbf16> loc(#loc) + %8 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/weights"} -> tensor<16x32x3x3xbf16> loc(#loc) + %9 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/biases"} -> tensor<32xbf16> loc(#loc) + %10 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/weights"} -> tensor<32x32x3x3xbf16> loc(#loc) + %11 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/biases"} -> tensor<32xbf16> loc(#loc) + %12 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/weights"} -> tensor<32x59x3x3xbf16> loc(#loc) + %13 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_411/Constant_0_0"} -> tensor<1x20x45x80xbf16> loc(#loc3) + %14 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/biases"} -> tensor<20xbf16> loc(#loc) + %15 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/weights"} -> tensor<20x40x3x3xbf16> loc(#loc) + %16 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/biases"} -> tensor<40xbf16> loc(#loc) + %17 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/weights"} -> tensor<40x40x3x3xbf16> loc(#loc) + %18 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/biases"} -> tensor<40xbf16> loc(#loc) + %19 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/weights"} -> tensor<40x107x3x3xbf16> loc(#loc) + %20 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_385/Constant_0_0"} -> tensor<1x40x23x40xbf16> loc(#loc4) + %21 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/biases"} -> tensor<40xbf16> loc(#loc) + %22 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/weights"} -> tensor<40x80x3x3xbf16> loc(#loc) + %23 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/biases"} -> tensor<80xbf16> loc(#loc) + %24 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/weights"} -> tensor<80x80x3x3xbf16> loc(#loc) + %25 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/biases"} -> tensor<80xbf16> loc(#loc) + %26 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/weights"} -> tensor<80x171x3x3xbf16> loc(#loc) + %27 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_359/Constant_0_0"} -> tensor<1x64x12x20xbf16> loc(#loc5) + %28 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/biases"} -> tensor<64xbf16> loc(#loc) + %29 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/weights"} -> tensor<64x128x3x3xbf16> loc(#loc) + %30 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/biases"} -> tensor<128xbf16> loc(#loc) + %31 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/weights"} -> tensor<128x128x3x3xbf16> loc(#loc) + %32 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/biases"} -> tensor<128xbf16> loc(#loc) + %33 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %34 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/biases"} -> tensor<128xbf16> loc(#loc) + %35 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %36 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/biases"} -> tensor<960xbf16> loc(#loc) + %37 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %38 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/biases"} -> tensor<160xbf16> loc(#loc) + %39 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %40 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/biases"} -> tensor<960xbf16> loc(#loc) + %41 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %42 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/biases"} -> tensor<240xbf16> loc(#loc) + %43 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %44 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/biases"} -> tensor<960xbf16> loc(#loc) + %45 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %46 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/biases"} -> tensor<960xbf16> loc(#loc) + %47 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %48 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/biases"} -> tensor<160xbf16> loc(#loc) + %49 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %50 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/biases"} -> tensor<960xbf16> loc(#loc) + %51 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %52 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/biases"} -> tensor<240xbf16> loc(#loc) + %53 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %54 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/biases"} -> tensor<960xbf16> loc(#loc) + %55 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %56 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/biases"} -> tensor<960xbf16> loc(#loc) + %57 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %58 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/biases"} -> tensor<160xbf16> loc(#loc) + %59 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/weights"} -> tensor<160x672x1x1xbf16> loc(#loc) + %60 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/biases"} -> tensor<672xbf16> loc(#loc) + %61 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %62 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/biases"} -> tensor<168xbf16> loc(#loc) + %63 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %64 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/biases"} -> tensor<672xbf16> loc(#loc) + %65 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/weights"} -> tensor<672x1x9x9xbf16> loc(#loc) + %66 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/biases"} -> tensor<672xbf16> loc(#loc) + %67 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %68 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/biases"} -> tensor<112xbf16> loc(#loc) + %69 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/weights"} -> tensor<112x672x1x1xbf16> loc(#loc) + %70 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/biases"} -> tensor<672xbf16> loc(#loc) + %71 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %72 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/biases"} -> tensor<168xbf16> loc(#loc) + %73 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %74 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/biases"} -> tensor<672xbf16> loc(#loc) + %75 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/weights"} -> tensor<672x1x3x3xbf16> loc(#loc) + %76 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/biases"} -> tensor<672xbf16> loc(#loc) + %77 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %78 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/biases"} -> tensor<112xbf16> loc(#loc) + %79 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/weights"} -> tensor<112x480x1x1xbf16> loc(#loc) + %80 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/biases"} -> tensor<480xbf16> loc(#loc) + %81 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/weights"} -> tensor<480x120x1x1xbf16> loc(#loc) + %82 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/biases"} -> tensor<120xbf16> loc(#loc) + %83 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/weights"} -> tensor<120x480x1x1xbf16> loc(#loc) + %84 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/biases"} -> tensor<480xbf16> loc(#loc) + %85 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/weights"} -> tensor<480x1x3x3xbf16> loc(#loc) + %86 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/biases"} -> tensor<480xbf16> loc(#loc) + %87 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/weights"} -> tensor<480x80x1x1xbf16> loc(#loc) + %88 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/biases"} -> tensor<80xbf16> loc(#loc) + %89 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %90 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/biases"} -> tensor<184xbf16> loc(#loc) + %91 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %92 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/biases"} -> tensor<184xbf16> loc(#loc) + %93 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %94 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/biases"} -> tensor<80xbf16> loc(#loc) + %95 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %96 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/biases"} -> tensor<184xbf16> loc(#loc) + %97 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %98 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/biases"} -> tensor<184xbf16> loc(#loc) + %99 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %100 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/biases"} -> tensor<80xbf16> loc(#loc) + %101 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/weights"} -> tensor<80x200x1x1xbf16> loc(#loc) + %102 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/biases"} -> tensor<200xbf16> loc(#loc) + %103 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/weights"} -> tensor<200x1x3x3xbf16> loc(#loc) + %104 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/biases"} -> tensor<200xbf16> loc(#loc) + %105 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/weights"} -> tensor<200x80x1x1xbf16> loc(#loc) + %106 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/biases"} -> tensor<80xbf16> loc(#loc) + %107 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/weights"} -> tensor<80x240x1x1xbf16> loc(#loc) + %108 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/biases"} -> tensor<240xbf16> loc(#loc) + %109 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/weights"} -> tensor<240x1x3x3xbf16> loc(#loc) + %110 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/biases"} -> tensor<240xbf16> loc(#loc) + %111 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/weights"} -> tensor<240x40x1x1xbf16> loc(#loc) + %112 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/biases"} -> tensor<40xbf16> loc(#loc) + %113 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %114 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/biases"} -> tensor<120xbf16> loc(#loc) + %115 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %116 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/biases"} -> tensor<32xbf16> loc(#loc) + %117 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %118 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/biases"} -> tensor<120xbf16> loc(#loc) + %119 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %120 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/biases"} -> tensor<120xbf16> loc(#loc) + %121 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %122 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/biases"} -> tensor<40xbf16> loc(#loc) + %123 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %124 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/biases"} -> tensor<120xbf16> loc(#loc) + %125 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %126 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/biases"} -> tensor<32xbf16> loc(#loc) + %127 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %128 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/biases"} -> tensor<120xbf16> loc(#loc) + %129 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %130 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/biases"} -> tensor<120xbf16> loc(#loc) + %131 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %132 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/biases"} -> tensor<40xbf16> loc(#loc) + %133 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/weights"} -> tensor<40x72x1x1xbf16> loc(#loc) + %134 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/biases"} -> tensor<72xbf16> loc(#loc) + %135 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %136 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/biases"} -> tensor<24xbf16> loc(#loc) + %137 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %138 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/biases"} -> tensor<72xbf16> loc(#loc) + %139 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/weights"} -> tensor<72x1x5x5xbf16> loc(#loc) + %140 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/biases"} -> tensor<72xbf16> loc(#loc) + %141 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %142 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/biases"} -> tensor<24xbf16> loc(#loc) + %143 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %144 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/biases"} -> tensor<72xbf16> loc(#loc) + %145 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/weights"} -> tensor<72x1x3x3xbf16> loc(#loc) + %146 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/biases"} -> tensor<72xbf16> loc(#loc) + %147 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %148 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/biases"} -> tensor<24xbf16> loc(#loc) + %149 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/weights"} -> tensor<24x64x1x1xbf16> loc(#loc) + %150 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/biases"} -> tensor<64xbf16> loc(#loc) + %151 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/weights"} -> tensor<64x1x3x3xbf16> loc(#loc) + %152 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/biases"} -> tensor<64xbf16> loc(#loc) + %153 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/weights"} -> tensor<64x16x1x1xbf16> loc(#loc) + %154 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/biases"} -> tensor<16xbf16> loc(#loc) + %155 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/weights"} -> tensor<16x16x1x1xbf16> loc(#loc) + %156 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/biases"} -> tensor<16xbf16> loc(#loc) + %157 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/weights"} -> tensor<16x1x3x3xbf16> loc(#loc) + %158 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/biases"} -> tensor<16xbf16> loc(#loc) + %159 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/weights"} -> tensor<16x3x3x3xbf16> loc(#loc) + %160 = xten_nn.load_external_const {file = "constants.h5", key = "Div_16/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc6) + %161 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_14/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc309) + %162 = xten_nn.subgraph (%arg5 = %arg0: tensor<1x180x320x4xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 20, 4]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 320, 4]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 20, 4]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 320, 4]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 30800 : ui64, ofm = 15440 : ui64}, + computation_cycles = 2373 : ui64, + cycle_count = 30800 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 20 : ui32, + config.ifmsv_width = 4 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.906250e-03 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %465 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %464 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + %163 = xten_nn.subgraph (%arg5 = %162: tensor<1x180x320x4xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Slice_7", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Slice_7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "W", + config.dim_c = 184 : ui32, + config.dim_h = 320 : ui32, + config.dim_w = 4 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x180x320x3xbf16> loc(#loc9) + xten_nn.output %464 : tensor<1x180x320x3xbf16> loc(#loc9) + } -> tensor<1x180x320x3xbf16> loc(#loc9) + %164 = xten_nn.subgraph (%arg5 = %163: tensor<1x180x320x3xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Generated-#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "BufferPadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_0_padded = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_1_padded = 23 : ui32, + config.dim_2 = 3 : ui32, + config.dim_2_padded = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_padded = 8 : ui32, + config.dtype = "bfloat16" + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x180x320x3xbf16> loc(#loc10) + xten_nn.output %464 : tensor<1x180x320x3xbf16> loc(#loc10) + } -> tensor<1x180x320x3xbf16> loc(#loc10) + %165 = xten_nn.subgraph (%arg5 = %164: tensor<1x180x320x3xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#2", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Generated-#3", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_2 = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 10 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x3x180x320xbf16> loc(#loc311) + xten_nn.output %464 : tensor<1x3x180x320xbf16> loc(#loc311) + } -> tensor<1x3x180x320xbf16> loc(#loc310) + %166 = xten_nn.subgraph (%arg5 = %165: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#4", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Generated-#5", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 184 : ui32, + config.dim_0_unpadded = 180 : ui32, + config.dim_1 = 1 : ui32, + config.dim_1_unpadded = 1 : ui32, + config.dim_2 = 320 : ui32, + config.dim_2_unpadded = 320 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x3x180x320xbf16> loc(#loc10) + xten_nn.output %464 : tensor<1x3x180x320xbf16> loc(#loc10) + } -> tensor<1x3x180x320xbf16> loc(#loc10) + %167 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>, %arg6 = %161: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 5, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 5, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Initializer_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 3, 5, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 232200 : ui64, ofm = 117000 : ui64, wts = 232200 : ui64}, + computation_cycles = 34045 : ui64, + cycle_count = 2089800 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Initializer_398", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 5 : ui32, + config.ifmsv_width = 16 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 51 : ui64, + startup_cycles = 22 : ui64 + }} { + %465 = tensor.empty() : tensor<1x3x180x320xbf16> loc(#loc309) + xten_nn.output %465 : tensor<1x3x180x320xbf16> loc(#loc309) + } -> tensor<1x3x180x320xbf16> loc(#loc309) + xten_nn.output %464 : tensor<1x3x180x320xbf16> loc(#loc309) + } -> tensor<1x3x180x320xbf16> loc(#loc309) + %168 = xten_nn.subgraph (%arg5 = %167: tensor<1x3x180x320xbf16>, %arg6 = %160: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 5, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 5, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 3, 5, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 232200 : ui64, ofm = 117000 : ui64, wts = 232200 : ui64}, + computation_cycles = 30988 : ui64, + cycle_count = 2089800 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 5 : ui32, + config.ifmsv_width = 16 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 34 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %465 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %464 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + %169 = xten_nn.subgraph (%arg5 = %168: tensor<1x3x180x320xbf16>, %arg6 = %159: tensor<16x3x3x3xbf16>, %arg7 = %158: tensor<16xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 10, 66]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 61, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 3, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[16, 5, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[16, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 4, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 79500 : ui64, ofm = 38700 : ui64, wts = 330 : ui64}, + computation_cycles = 62525 : ui64, + cycle_count = 238500 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x3x180x320xbf16>, %arg9 = %arg6: tensor<16x3x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 8 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 10 : ui8, + config.ifm_sv_width.size = 66 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 8 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 2 : ui8, + config.stride_w = 2 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 1935 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc13) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc13) + } -> tensor<1x16x90x160xbf16> loc(#loc13) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc13) + } -> tensor<1x16x90x160xbf16> loc(#loc13) + %170 = xten_nn.subgraph (%arg5 = %169: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 23, 8]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 23, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 59080 : ui64, ofm = 29640 : ui64}, + computation_cycles = 5029 : ui64, + cycle_count = 59080 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 23 : ui32, + config.ifmsv_width = 8 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 103 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc14) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc14) + } -> tensor<1x16x90x160xbf16> loc(#loc14) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc14) + } -> tensor<1x16x90x160xbf16> loc(#loc14) + %171 = xten_nn.subgraph (%arg5 = %170: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 16, 45, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 12, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 16, 45, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61600 : ui64, ofm = 30880 : ui64}, + computation_cycles = 4639 : ui64, + cycle_count = 123200 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 12 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 139 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + %172 = xten_nn.subgraph (%arg5 = %171: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 23, 8]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 23, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 59080 : ui64, ofm = 29640 : ui64}, + computation_cycles = 5029 : ui64, + cycle_count = 59080 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 23 : ui32, + config.ifmsv_width = 8 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 103 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + %173 = xten_nn.subgraph (%arg5 = %169: tensor<1x16x90x160xbf16>, %arg6 = %172: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61740 : ui64, ifm2 = 61740 : ui64, ofm = 31020 : ui64}, + computation_cycles = 5698 : ui64, + cycle_count = 185220 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 8 : ui32, + config.ifmsv_width = 16 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 46 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + %174 = xten_nn.subgraph (%arg5 = %173: tensor<1x16x90x160xbf16>, %arg6 = %157: tensor<16x1x3x3xbf16>, %arg7 = %156: tensor<16xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 2]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 18]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[16, 0, 0, 1]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[16, 0, 0, 1]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 6, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 6, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 102800 : ui64, ofm = 31120 : ui64, wts = 74 : ui64}, + computation_cycles = 49653 : ui64, + cycle_count = 102800 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<16x1x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 8 : ui8, + config.ifm_sv_width.size = 18 : ui8, + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 1099 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc19) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc19) + } -> tensor<1x16x90x160xbf16> loc(#loc312) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc312) + } -> tensor<1x16x90x160xbf16> loc(#loc312) + %175 = xten_nn.subgraph (%arg5 = %174: tensor<1x16x90x160xbf16>, %arg6 = %155: tensor<16x16x1x1xbf16>, %arg7 = %154: tensor<16xbf16>, %arg8 = %173: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_28", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 48, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 16, 45, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 48, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[16, 48, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[16, 48, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 16, 45, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 16, 45, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 123480 : ui64, ifm2 = 62040 : ui64, ofm = 31320 : ui64, wts = 298 : ui64}, + computation_cycles = 76887 : ui64, + cycle_count = 246960 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x16x90x160xbf16>, %arg10 = %arg6: tensor<16x16x1x1xbf16>, %arg11 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_28", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_28", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 2 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 8 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 1095 : ui64, + startup_cycles = 182 : ui64 + }} { + %466 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc20) + xten_nn.output %466 : tensor<1x16x90x160xbf16> loc(#loc20) + } -> tensor<1x16x90x160xbf16> loc(#loc20) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x16x90x160xbf16>, %arg10 = %arg8: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_29", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 32 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 43 : ui64, + startup_cycles = 22 : ui64 + }} { + %466 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %466 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc313) + %176 = xten_nn.subgraph (%arg5 = %175: tensor<1x16x90x160xbf16>, %arg6 = %153: tensor<64x16x1x1xbf16>, %arg7 = %152: tensor<64xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 48, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 9, 8]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 18, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 48, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 48, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[0, 48, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 9, 8]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 18, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 115700 : ui64, ofm = 122100 : ui64, wts = 586 : ui64}, + computation_cycles = 124593 : ui64, + cycle_count = 622965 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<64x16x1x1xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 9 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 2348 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x64x90x160xbf16> loc(#loc23) + xten_nn.output %465 : tensor<1x64x90x160xbf16> loc(#loc23) + } -> tensor<1x64x90x160xbf16> loc(#loc314) + xten_nn.output %464 : tensor<1x64x90x160xbf16> loc(#loc314) + } -> tensor<1x64x90x160xbf16> loc(#loc314) + %177 = xten_nn.subgraph (%arg5 = %176: tensor<1x64x90x160xbf16>, %arg6 = %151: tensor<64x1x3x3xbf16>, %arg7 = %150: tensor<64xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 2]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 6, 26]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 17, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 2, 12]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 8, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 226632 : ui64, ofm = 16968 : ui64, wts = 148 : ui64}, + computation_cycles = 58677 : ui64, + cycle_count = 1359792 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x90x160xbf16>, %arg9 = %arg6: tensor<64x1x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 6 : ui8, + config.ifm_sv_width.size = 26 : ui8, + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 559 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x64x45x80xbf16> loc(#loc25) + xten_nn.output %465 : tensor<1x64x45x80xbf16> loc(#loc25) + } -> tensor<1x64x45x80xbf16> loc(#loc315) + xten_nn.output %464 : tensor<1x64x45x80xbf16> loc(#loc315) + } -> tensor<1x64x45x80xbf16> loc(#loc315) + %178 = xten_nn.subgraph (%arg5 = %177: tensor<1x64x45x80xbf16>) attributes { + LayerName = "Generated-#60", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Generated-#61", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 48 : ui32, + config.dim_0_unpadded = 45 : ui32, + config.dim_1 = 8 : ui32, + config.dim_1_unpadded = 8 : ui32, + config.dim_2 = 80 : ui32, + config.dim_2_unpadded = 80 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x64x45x80xbf16> loc(#loc10) + } -> tensor<1x64x45x80xbf16> loc(#loc10) + %179 = xten_nn.subgraph (%arg5 = %178: tensor<1x64x45x80xbf16>, %arg6 = %149: tensor<24x64x1x1xbf16>, %arg7 = %148: tensor<24xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 64, 4, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 64, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[40, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[40, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 4, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 40, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 30870 : ui64, ofm = 32790 : ui64, wts = 586 : ui64}, + computation_cycles = 24283 : ui64, + cycle_count = 32790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x45x80xbf16>, %arg9 = %arg6: tensor<24x64x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 4 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 1459 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x24x45x80xbf16> loc(#loc26) + xten_nn.output %465 : tensor<1x24x45x80xbf16> loc(#loc26) + } -> tensor<1x24x45x80xbf16> loc(#loc26) + xten_nn.output %464 : tensor<1x24x45x80xbf16> loc(#loc26) + } -> tensor<1x24x45x80xbf16> loc(#loc26) + %180 = xten_nn.subgraph (%arg5 = %179: tensor<1x24x45x80xbf16>, %arg6 = %147: tensor<72x24x1x1xbf16>, %arg7 = %146: tensor<72xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 40, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 40, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 40, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[24, 24, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[24, 40, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[24, 40, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 3, 16]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 37044 : ui64, ofm = 32436 : ui64, wts = 874 : ui64}, + computation_cycles = 32981 : ui64, + cycle_count = 37044 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 2 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 24 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 1675 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x72x45x80xbf16> loc(#loc28) + xten_nn.output %465 : tensor<1x72x45x80xbf16> loc(#loc28) + } -> tensor<1x72x45x80xbf16> loc(#loc316) + xten_nn.output %464 : tensor<1x72x45x80xbf16> loc(#loc316) + } -> tensor<1x72x45x80xbf16> loc(#loc316) + %181 = xten_nn.subgraph (%arg5 = %180: tensor<1x72x45x80xbf16>, %arg6 = %145: tensor<72x1x3x3xbf16>, %arg7 = %144: tensor<72xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 2]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 18]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 72, 25, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[24, 0, 0, 1]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[24, 0, 0, 1]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 6, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 72, 23, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 77100 : ui64, ofm = 23340 : ui64, wts = 222 : ui64}, + computation_cycles = 37293 : ui64, + cycle_count = 154200 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x3x3xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 8 : ui8, + config.ifm_sv_width.size = 18 : ui8, + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 1099 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x72x45x80xbf16> loc(#loc30) + xten_nn.output %465 : tensor<1x72x45x80xbf16> loc(#loc30) + } -> tensor<1x72x45x80xbf16> loc(#loc317) + xten_nn.output %464 : tensor<1x72x45x80xbf16> loc(#loc317) + } -> tensor<1x72x45x80xbf16> loc(#loc317) + %182 = xten_nn.subgraph (%arg5 = %181: tensor<1x72x45x80xbf16>) attributes { + LayerName = "Generated-#62", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Generated-#63", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 46 : ui32, + config.dim_0_unpadded = 45 : ui32, + config.dim_1 = 9 : ui32, + config.dim_1_unpadded = 9 : ui32, + config.dim_2 = 80 : ui32, + config.dim_2_unpadded = 80 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x72x45x80xbf16> loc(#loc10) + } -> tensor<1x72x45x80xbf16> loc(#loc10) + %183 = xten_nn.subgraph (%arg5 = %182: tensor<1x72x45x80xbf16>, %arg6 = %143: tensor<24x72x1x1xbf16>, %arg7 = %142: tensor<24xbf16>, %arg8 = %179: tensor<1x24x45x80xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_39", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 72, 4, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 72, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[40, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[40, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 4, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 40, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 4, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 40, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 34710 : ui64, ifm2 = 30870 : ui64, ofm = 15510 : ui64, wts = 650 : ui64}, + computation_cycles = 27290 : ui64, + cycle_count = 34710 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x72x45x80xbf16>, %arg10 = %arg6: tensor<24x72x1x1xbf16>, %arg11 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_39", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_39", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 72 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 4 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 1583 : ui64, + startup_cycles = 160 : ui64 + }} { + %466 = tensor.empty() : tensor<1x24x45x80xbf16> loc(#loc31) + xten_nn.output %466 : tensor<1x24x45x80xbf16> loc(#loc31) + } -> tensor<1x24x45x80xbf16> loc(#loc31) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x24x45x80xbf16>, %arg10 = %arg8: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Add_40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 4 : ui32, + config.ifmsv_width = 16 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 75 : ui64, + startup_cycles = 22 : ui64 + }} { + %466 = tensor.empty() : tensor<1x24x45x80xbf16> loc(#loc32) + xten_nn.output %466 : tensor<1x24x45x80xbf16> loc(#loc32) + } -> tensor<1x24x45x80xbf16> loc(#loc32) + xten_nn.output %465 : tensor<1x24x45x80xbf16> loc(#loc32) + } -> tensor<1x24x45x80xbf16> loc(#loc318) + %184 = xten_nn.subgraph (%arg5 = %183: tensor<1x24x45x80xbf16>, %arg6 = %141: tensor<72x24x1x1xbf16>, %arg7 = %140: tensor<72xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 40, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 40, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 40, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[24, 24, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[24, 40, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[24, 40, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 3, 16]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 37044 : ui64, ofm = 32436 : ui64, wts = 874 : ui64}, + computation_cycles = 32981 : ui64, + cycle_count = 37044 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 2 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 24 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 1675 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x72x45x80xbf16> loc(#loc34) + xten_nn.output %465 : tensor<1x72x45x80xbf16> loc(#loc34) + } -> tensor<1x72x45x80xbf16> loc(#loc319) + xten_nn.output %464 : tensor<1x72x45x80xbf16> loc(#loc319) + } -> tensor<1x72x45x80xbf16> loc(#loc319) + %185 = xten_nn.subgraph (%arg5 = %184: tensor<1x72x45x80xbf16>, %arg6 = %139: tensor<72x1x5x5xbf16>, %arg7 = %138: tensor<72xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 3, 16]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 5, 5]> : vector<4xindex>, + l2_extend_end = dense<[24, 0, 0, 3]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex>, + l3_extend_end = dense<[24, 0, 0, 3]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 2, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 115650 : ui64, ofm = 6210 : ui64, wts = 558 : ui64}, + computation_cycles = 35313 : ui64, + cycle_count = 115650 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x5x5xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 1]], + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 8 : ui8, + config.ifm_sv_width.size = 20 : ui8, + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 2 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 643 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x72x23x40xbf16> loc(#loc36) + xten_nn.output %465 : tensor<1x72x23x40xbf16> loc(#loc36) + } -> tensor<1x72x23x40xbf16> loc(#loc320) + xten_nn.output %464 : tensor<1x72x23x40xbf16> loc(#loc320) + } -> tensor<1x72x23x40xbf16> loc(#loc320) + %186 = xten_nn.subgraph (%arg5 = %185: tensor<1x72x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#6", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 9 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x72x1x920xbf16> loc(#loc37) + xten_nn.output %464 : tensor<1x72x1x920xbf16> loc(#loc37) + } -> tensor<1x72x1x920xbf16> loc(#loc37) + %187 = xten_nn.subgraph (%arg5 = %186: tensor<1x72x1x920xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#8", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 64]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#9", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61590 : ui64, ofm = 3990 : ui64}, + computation_cycles = 66253 : ui64, + cycle_count = 66253 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x920xbf16>) attributes { + LayerName = "Generated-#8", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#9", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 72 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W", + config.sv_channel = 32 : ui32, + config.sv_height = 1 : ui32, + config.sv_width = 64 : ui32 + }, + estimation = { + compute_cycles = 2337 : ui64, + startup_cycles = 190 : ui64 + }} { + %465 = tensor.empty() : tensor<1x72x1x1xbf16> loc(#loc37) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc37) + } -> tensor<1x72x1x1xbf16> loc(#loc37) + xten_nn.output %464 : tensor<1x72x1x1xbf16> loc(#loc37) + } -> tensor<1x72x1x1xbf16> loc(#loc37) + %188 = xten_nn.subgraph (%arg5 = %187: tensor<1x72x1x1xbf16>, %arg6 = %137: tensor<24x72x1x1xbf16>, %arg7 = %136: tensor<24xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 72, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[40, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[40, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 40, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 586 : ui64, ofm = 778 : ui64, wts = 650 : ui64}, + computation_cycles = 1216 : ui64, + cycle_count = 1216 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x1x1xbf16>, %arg9 = %arg6: tensor<24x72x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 72 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 736 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x24x1x1xbf16> loc(#loc39) + xten_nn.output %465 : tensor<1x24x1x1xbf16> loc(#loc39) + } -> tensor<1x24x1x1xbf16> loc(#loc321) + xten_nn.output %464 : tensor<1x24x1x1xbf16> loc(#loc321) + } -> tensor<1x24x1x1xbf16> loc(#loc321) + %189 = xten_nn.subgraph (%arg5 = %188: tensor<1x24x1x1xbf16>, %arg6 = %135: tensor<72x24x1x1xbf16>, %arg7 = %134: tensor<72xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 40, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 40, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 40, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 24, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[56, 40, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[56, 40, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 266 : ui64, ofm = 650 : ui64, wts = 1162 : ui64}, + computation_cycles = 1233 : ui64, + cycle_count = 1233 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x1x1xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 753 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x72x1x1xbf16> loc(#loc40) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc40) + } -> tensor<1x72x1x1xbf16> loc(#loc40) + xten_nn.output %464 : tensor<1x72x1x1xbf16> loc(#loc40) + } -> tensor<1x72x1x1xbf16> loc(#loc40) + %190 = xten_nn.subgraph (%arg5 = %189: tensor<1x72x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 15, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 15, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 63, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 778 : ui64, ofm = 394 : ui64}, + computation_cycles = 401 : ui64, + cycle_count = 778 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 16 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x72x1x1xbf16> loc(#loc41) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc41) + } -> tensor<1x72x1x1xbf16> loc(#loc41) + xten_nn.output %464 : tensor<1x72x1x1xbf16> loc(#loc41) + } -> tensor<1x72x1x1xbf16> loc(#loc41) + %191 = xten_nn.subgraph (%arg5 = %190: tensor<1x72x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 7, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 63, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 7, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 31, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 522 : ui64, ofm = 266 : ui64}, + computation_cycles = 395 : ui64, + cycle_count = 522 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 8 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %464 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + %192 = xten_nn.subgraph (%arg5 = %191: tensor<1x72x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 15, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 31, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 15, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 63, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1034 : ui64, ofm = 522 : ui64}, + computation_cycles = 409 : ui64, + cycle_count = 1034 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 16 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 43 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %464 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + %193 = xten_nn.subgraph (%arg5 = %192: tensor<1x72x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#10", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#11", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 72 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %464 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + %194 = xten_nn.subgraph (%arg5 = %193: tensor<1x72x23x40xbf16>, %arg6 = %185: tensor<1x72x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 11580 : ui64, ifm2 = 11580 : ui64, ofm = 5820 : ui64}, + computation_cycles = 1294 : ui64, + cycle_count = 11580 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x72x23x40xbf16>, %arg8 = %arg6: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %465 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %464 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + %195 = xten_nn.subgraph (%arg5 = %194: tensor<1x72x23x40xbf16>, %arg6 = %133: tensor<40x72x1x1xbf16>, %arg7 = %132: tensor<40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 72, 6, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 72, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[24, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[24, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 8690 : ui64, ofm = 8370 : ui64, wts = 650 : ui64}, + computation_cycles = 9983 : ui64, + cycle_count = 9983 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x23x40xbf16>, %arg9 = %arg6: tensor<40x72x1x1xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 72 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 6 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 1791 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc45) + xten_nn.output %465 : tensor<1x40x23x40xbf16> loc(#loc45) + } -> tensor<1x40x23x40xbf16> loc(#loc45) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc45) + } -> tensor<1x40x23x40xbf16> loc(#loc45) + %196 = xten_nn.subgraph (%arg5 = %195: tensor<1x40x23x40xbf16>, %arg6 = %131: tensor<120x40x1x1xbf16>, %arg7 = %130: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 40, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 40, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[8, 24, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[8, 24, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 24]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 12348 : ui64, ofm = 13884 : ui64, wts = 1162 : ui64}, + computation_cycles = 12977 : ui64, + cycle_count = 13884 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 2 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 1965 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x23x40xbf16> loc(#loc47) + xten_nn.output %465 : tensor<1x120x23x40xbf16> loc(#loc47) + } -> tensor<1x120x23x40xbf16> loc(#loc322) + xten_nn.output %464 : tensor<1x120x23x40xbf16> loc(#loc322) + } -> tensor<1x120x23x40xbf16> loc(#loc322) + %197 = xten_nn.subgraph (%arg5 = %196: tensor<1x120x23x40xbf16>, %arg6 = %129: tensor<120x1x5x5xbf16>, %arg7 = %128: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 10, 12]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 24]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 5, 5]> : vector<4xindex>, + l2_extend_end = dense<[8, 0, 0, 3]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex>, + l3_extend_end = dense<[8, 0, 0, 3]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 6, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 38600 : ui64, ofm = 7880 : ui64, wts = 744 : ui64}, + computation_cycles = 22053 : ui64, + cycle_count = 38600 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 10 : ui8, + config.ifm_sv_width.size = 12 : ui8, + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 955 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x23x40xbf16> loc(#loc49) + xten_nn.output %465 : tensor<1x120x23x40xbf16> loc(#loc49) + } -> tensor<1x120x23x40xbf16> loc(#loc323) + xten_nn.output %464 : tensor<1x120x23x40xbf16> loc(#loc323) + } -> tensor<1x120x23x40xbf16> loc(#loc323) + %198 = xten_nn.subgraph (%arg5 = %197: tensor<1x120x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#12", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#13", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x120x1x920xbf16> loc(#loc50) + xten_nn.output %464 : tensor<1x120x1x920xbf16> loc(#loc50) + } -> tensor<1x120x1x920xbf16> loc(#loc50) + %199 = xten_nn.subgraph (%arg5 = %198: tensor<1x120x1x920xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 64]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#15", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61590 : ui64, ofm = 3990 : ui64}, + computation_cycles = 66253 : ui64, + cycle_count = 66253 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#15", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W", + config.sv_channel = 32 : ui32, + config.sv_height = 1 : ui32, + config.sv_width = 64 : ui32 + }, + estimation = { + compute_cycles = 2337 : ui64, + startup_cycles = 190 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc50) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc50) + } -> tensor<1x120x1x1xbf16> loc(#loc50) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc50) + } -> tensor<1x120x1x1xbf16> loc(#loc50) + %200 = xten_nn.subgraph (%arg5 = %199: tensor<1x120x1x1xbf16>, %arg6 = %127: tensor<32x120x1x1xbf16>, %arg7 = %126: tensor<32xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 120, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[32, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[32, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 970 : ui64, ofm = 778 : ui64, wts = 1034 : ui64}, + computation_cycles = 1498 : ui64, + cycle_count = 1498 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 120 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 1018 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x32x1x1xbf16> loc(#loc52) + xten_nn.output %465 : tensor<1x32x1x1xbf16> loc(#loc52) + } -> tensor<1x32x1x1xbf16> loc(#loc324) + xten_nn.output %464 : tensor<1x32x1x1xbf16> loc(#loc324) + } -> tensor<1x32x1x1xbf16> loc(#loc324) + %201 = xten_nn.subgraph (%arg5 = %200: tensor<1x32x1x1xbf16>, %arg6 = %125: tensor<120x32x1x1xbf16>, %arg7 = %124: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 32, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[8, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[8, 32, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 266 : ui64, ofm = 650 : ui64, wts = 1162 : ui64}, + computation_cycles = 1233 : ui64, + cycle_count = 1233 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 753 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc53) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc53) + } -> tensor<1x120x1x1xbf16> loc(#loc53) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc53) + } -> tensor<1x120x1x1xbf16> loc(#loc53) + %202 = xten_nn.subgraph (%arg5 = %201: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 11, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 11, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 47, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 778 : ui64, ofm = 394 : ui64}, + computation_cycles = 401 : ui64, + cycle_count = 778 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 12 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc54) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc54) + } -> tensor<1x120x1x1xbf16> loc(#loc54) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc54) + } -> tensor<1x120x1x1xbf16> loc(#loc54) + %203 = xten_nn.subgraph (%arg5 = %202: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 7, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 47, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 7, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 31, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 522 : ui64, ofm = 266 : ui64}, + computation_cycles = 395 : ui64, + cycle_count = 522 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 8 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + %204 = xten_nn.subgraph (%arg5 = %203: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 15, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 31, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 15, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 63, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1034 : ui64, ofm = 522 : ui64}, + computation_cycles = 409 : ui64, + cycle_count = 1034 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 16 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 43 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + %205 = xten_nn.subgraph (%arg5 = %204: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#16", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#17", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %464 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + %206 = xten_nn.subgraph (%arg5 = %205: tensor<1x120x23x40xbf16>, %arg6 = %197: tensor<1x120x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_73", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15440 : ui64, ifm2 = 15440 : ui64, ofm = 7760 : ui64}, + computation_cycles = 1656 : ui64, + cycle_count = 15440 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %465 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %464 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + %207 = xten_nn.subgraph (%arg5 = %206: tensor<1x120x23x40xbf16>, %arg6 = %123: tensor<40x120x1x1xbf16>, %arg7 = %122: tensor<40xbf16>, %arg8 = %195: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_74", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 64, 1, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 64, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[24, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[24, 8, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 1, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 1, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15480 : ui64, ifm2 = 7740 : ui64, ofm = 3900 : ui64, wts = 1172 : ui64}, + computation_cycles = 16811 : ui64, + cycle_count = 16811 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_74", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_74", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 40 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 40 : ui8 + }, + estimation = { + compute_cycles = 2416 : ui64, + startup_cycles = 160 : ui64 + }} { + %466 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc58) + xten_nn.output %466 : tensor<1x40x23x40xbf16> loc(#loc58) + } -> tensor<1x40x23x40xbf16> loc(#loc58) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_75", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 40 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 51 : ui64, + startup_cycles = 22 : ui64 + }} { + %466 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc59) + xten_nn.output %466 : tensor<1x40x23x40xbf16> loc(#loc59) + } -> tensor<1x40x23x40xbf16> loc(#loc59) + xten_nn.output %465 : tensor<1x40x23x40xbf16> loc(#loc59) + } -> tensor<1x40x23x40xbf16> loc(#loc325) + %208 = xten_nn.subgraph (%arg5 = %207: tensor<1x40x23x40xbf16>, %arg6 = %121: tensor<120x40x1x1xbf16>, %arg7 = %120: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 40, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 40, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[8, 24, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[8, 24, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 24]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 12348 : ui64, ofm = 13884 : ui64, wts = 1162 : ui64}, + computation_cycles = 12977 : ui64, + cycle_count = 13884 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 2 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 1965 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x23x40xbf16> loc(#loc61) + xten_nn.output %465 : tensor<1x120x23x40xbf16> loc(#loc61) + } -> tensor<1x120x23x40xbf16> loc(#loc326) + xten_nn.output %464 : tensor<1x120x23x40xbf16> loc(#loc326) + } -> tensor<1x120x23x40xbf16> loc(#loc326) + %209 = xten_nn.subgraph (%arg5 = %208: tensor<1x120x23x40xbf16>, %arg6 = %119: tensor<120x1x5x5xbf16>, %arg7 = %118: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 10, 12]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 24]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 5, 5]> : vector<4xindex>, + l2_extend_end = dense<[8, 0, 0, 3]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex>, + l3_extend_end = dense<[8, 0, 0, 3]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 6, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 38600 : ui64, ofm = 7880 : ui64, wts = 744 : ui64}, + computation_cycles = 22053 : ui64, + cycle_count = 38600 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 10 : ui8, + config.ifm_sv_width.size = 12 : ui8, + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 955 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x23x40xbf16> loc(#loc63) + xten_nn.output %465 : tensor<1x120x23x40xbf16> loc(#loc63) + } -> tensor<1x120x23x40xbf16> loc(#loc327) + xten_nn.output %464 : tensor<1x120x23x40xbf16> loc(#loc327) + } -> tensor<1x120x23x40xbf16> loc(#loc327) + %210 = xten_nn.subgraph (%arg5 = %209: tensor<1x120x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#18", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#19", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x120x1x920xbf16> loc(#loc64) + xten_nn.output %464 : tensor<1x120x1x920xbf16> loc(#loc64) + } -> tensor<1x120x1x920xbf16> loc(#loc64) + %211 = xten_nn.subgraph (%arg5 = %210: tensor<1x120x1x920xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#20", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 64]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#21", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61590 : ui64, ofm = 3990 : ui64}, + computation_cycles = 66253 : ui64, + cycle_count = 66253 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#20", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#21", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W", + config.sv_channel = 32 : ui32, + config.sv_height = 1 : ui32, + config.sv_width = 64 : ui32 + }, + estimation = { + compute_cycles = 2337 : ui64, + startup_cycles = 190 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc64) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc64) + } -> tensor<1x120x1x1xbf16> loc(#loc64) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc64) + } -> tensor<1x120x1x1xbf16> loc(#loc64) + %212 = xten_nn.subgraph (%arg5 = %211: tensor<1x120x1x1xbf16>, %arg6 = %117: tensor<32x120x1x1xbf16>, %arg7 = %116: tensor<32xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 120, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[32, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[32, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 970 : ui64, ofm = 778 : ui64, wts = 1034 : ui64}, + computation_cycles = 1498 : ui64, + cycle_count = 1498 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 120 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 1018 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x32x1x1xbf16> loc(#loc66) + xten_nn.output %465 : tensor<1x32x1x1xbf16> loc(#loc66) + } -> tensor<1x32x1x1xbf16> loc(#loc328) + xten_nn.output %464 : tensor<1x32x1x1xbf16> loc(#loc328) + } -> tensor<1x32x1x1xbf16> loc(#loc328) + %213 = xten_nn.subgraph (%arg5 = %212: tensor<1x32x1x1xbf16>, %arg6 = %115: tensor<120x32x1x1xbf16>, %arg7 = %114: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 32, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[8, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[8, 32, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 266 : ui64, ofm = 650 : ui64, wts = 1162 : ui64}, + computation_cycles = 1233 : ui64, + cycle_count = 1233 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 753 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc67) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc67) + } -> tensor<1x120x1x1xbf16> loc(#loc67) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc67) + } -> tensor<1x120x1x1xbf16> loc(#loc67) + %214 = xten_nn.subgraph (%arg5 = %213: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 11, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 11, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 47, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 778 : ui64, ofm = 394 : ui64}, + computation_cycles = 401 : ui64, + cycle_count = 778 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 12 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc68) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc68) + } -> tensor<1x120x1x1xbf16> loc(#loc68) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc68) + } -> tensor<1x120x1x1xbf16> loc(#loc68) + %215 = xten_nn.subgraph (%arg5 = %214: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 7, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 47, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 7, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 31, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 522 : ui64, ofm = 266 : ui64}, + computation_cycles = 395 : ui64, + cycle_count = 522 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 8 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + %216 = xten_nn.subgraph (%arg5 = %215: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 15, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 31, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 15, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 63, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1034 : ui64, ofm = 522 : ui64}, + computation_cycles = 409 : ui64, + cycle_count = 1034 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 16 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 43 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + %217 = xten_nn.subgraph (%arg5 = %216: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#22", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#23", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %464 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + %218 = xten_nn.subgraph (%arg5 = %217: tensor<1x120x23x40xbf16>, %arg6 = %209: tensor<1x120x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_91", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15440 : ui64, ifm2 = 15440 : ui64, ofm = 7760 : ui64}, + computation_cycles = 1656 : ui64, + cycle_count = 15440 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %465 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %464 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + %219 = xten_nn.subgraph (%arg5 = %218: tensor<1x120x23x40xbf16>, %arg6 = %113: tensor<40x120x1x1xbf16>, %arg7 = %112: tensor<40xbf16>, %arg8 = %207: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_92", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 64, 1, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 64, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[24, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[24, 8, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 1, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 1, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15480 : ui64, ifm2 = 7740 : ui64, ofm = 3900 : ui64, wts = 1172 : ui64}, + computation_cycles = 16811 : ui64, + cycle_count = 16811 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_92", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_92", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 40 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 40 : ui8 + }, + estimation = { + compute_cycles = 2416 : ui64, + startup_cycles = 160 : ui64 + }} { + %466 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc72) + xten_nn.output %466 : tensor<1x40x23x40xbf16> loc(#loc72) + } -> tensor<1x40x23x40xbf16> loc(#loc72) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_93", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 40 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 51 : ui64, + startup_cycles = 22 : ui64 + }} { + %466 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc73) + xten_nn.output %466 : tensor<1x40x23x40xbf16> loc(#loc73) + } -> tensor<1x40x23x40xbf16> loc(#loc73) + xten_nn.output %465 : tensor<1x40x23x40xbf16> loc(#loc73) + } -> tensor<1x40x23x40xbf16> loc(#loc329) + %220 = xten_nn.subgraph (%arg5 = %219: tensor<1x40x23x40xbf16>, %arg6 = %111: tensor<240x40x1x1xbf16>, %arg7 = %110: tensor<240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 40, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 40, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[16, 24, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[16, 24, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 2, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 1, 24]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 12348 : ui64, ofm = 27768 : ui64, wts = 2324 : ui64}, + computation_cycles = 25589 : ui64, + cycle_count = 27768 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<240x40x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 2 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 1965 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x23x40xbf16> loc(#loc74) + xten_nn.output %465 : tensor<1x240x23x40xbf16> loc(#loc74) + } -> tensor<1x240x23x40xbf16> loc(#loc74) + xten_nn.output %464 : tensor<1x240x23x40xbf16> loc(#loc74) + } -> tensor<1x240x23x40xbf16> loc(#loc74) + %221 = xten_nn.subgraph (%arg5 = %220: tensor<1x240x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 1, 24]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 30800 : ui64, ofm = 15440 : ui64}, + computation_cycles = 2373 : ui64, + cycle_count = 30800 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x23x40xbf16> loc(#loc75) + xten_nn.output %465 : tensor<1x240x23x40xbf16> loc(#loc75) + } -> tensor<1x240x23x40xbf16> loc(#loc75) + xten_nn.output %464 : tensor<1x240x23x40xbf16> loc(#loc75) + } -> tensor<1x240x23x40xbf16> loc(#loc75) + %222 = xten_nn.subgraph (%arg5 = %221: tensor<1x240x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 30800 : ui64, ofm = 15440 : ui64}, + computation_cycles = 2431 : ui64, + cycle_count = 30800 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 139 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %465 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %464 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + %223 = xten_nn.subgraph (%arg5 = %222: tensor<1x240x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 30800 : ui64, ofm = 15440 : ui64}, + computation_cycles = 2373 : ui64, + cycle_count = 30800 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %465 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %464 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + %224 = xten_nn.subgraph (%arg5 = %220: tensor<1x240x23x40xbf16>, %arg6 = %223: tensor<1x240x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 240, 12, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 240, 12, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 240, 12, 40]> : vector<4xindex>, + l3_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 30880 : ui64, ifm2 = 30880 : ui64, ofm = 15520 : ui64}, + computation_cycles = 3104 : ui64, + cycle_count = 61760 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x23x40xbf16>, %arg8 = %arg6: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %465 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %464 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + %225 = xten_nn.subgraph (%arg5 = %224: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Generated-#64", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#65", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 24 : ui32, + config.dim_0_unpadded = 23 : ui32, + config.dim_1 = 30 : ui32, + config.dim_1_unpadded = 30 : ui32, + config.dim_2 = 40 : ui32, + config.dim_2_unpadded = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x240x23x40xbf16> loc(#loc10) + } -> tensor<1x240x23x40xbf16> loc(#loc10) + %226 = xten_nn.subgraph (%arg5 = %225: tensor<1x240x23x40xbf16>, %arg6 = %109: tensor<240x1x3x3xbf16>, %arg7 = %108: tensor<240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 2]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 10, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[16, 0, 0, 1]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[16, 0, 0, 1]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 4, 4]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 4, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 77200 : ui64, ofm = 5520 : ui64, wts = 592 : ui64}, + computation_cycles = 20853 : ui64, + cycle_count = 77200 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x23x40xbf16>, %arg9 = %arg6: tensor<240x1x3x3xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 0]], + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 10 : ui8, + config.ifm_sv_width.size = 10 : ui8, + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 379 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x12x20xbf16> loc(#loc79) + xten_nn.output %465 : tensor<1x240x12x20xbf16> loc(#loc79) + } -> tensor<1x240x12x20xbf16> loc(#loc79) + xten_nn.output %464 : tensor<1x240x12x20xbf16> loc(#loc79) + } -> tensor<1x240x12x20xbf16> loc(#loc79) + %227 = xten_nn.subgraph (%arg5 = %226: tensor<1x240x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 4, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7700 : ui64, ofm = 3860 : ui64}, + computation_cycles = 765 : ui64, + cycle_count = 7700 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x12x20xbf16> loc(#loc80) + xten_nn.output %465 : tensor<1x240x12x20xbf16> loc(#loc80) + } -> tensor<1x240x12x20xbf16> loc(#loc80) + xten_nn.output %464 : tensor<1x240x12x20xbf16> loc(#loc80) + } -> tensor<1x240x12x20xbf16> loc(#loc80) + %228 = xten_nn.subgraph (%arg5 = %227: tensor<1x240x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7700 : ui64, ofm = 3860 : ui64}, + computation_cycles = 775 : ui64, + cycle_count = 7700 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 139 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %465 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %464 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + %229 = xten_nn.subgraph (%arg5 = %228: tensor<1x240x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7700 : ui64, ofm = 3860 : ui64}, + computation_cycles = 765 : ui64, + cycle_count = 7700 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %465 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %464 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + %230 = xten_nn.subgraph (%arg5 = %226: tensor<1x240x12x20xbf16>, %arg6 = %229: tensor<1x240x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 4, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7720 : ui64, ifm2 = 7720 : ui64, ofm = 3880 : ui64}, + computation_cycles = 932 : ui64, + cycle_count = 7720 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x12x20xbf16>, %arg8 = %arg6: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %465 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %464 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + %231 = xten_nn.subgraph (%arg5 = %230: tensor<1x240x12x20xbf16>, %arg6 = %107: tensor<80x240x1x1xbf16>, %arg7 = %106: tensor<80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[24, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 11610 : ui64, ofm = 3102 : ui64, wts = 3198 : ui64}, + computation_cycles = 12272 : ui64, + cycle_count = 12272 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x12x20xbf16>, %arg9 = %arg6: tensor<80x240x1x1xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 3 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 24 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 3558 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x80x12x20xbf16> loc(#loc84) + xten_nn.output %465 : tensor<1x80x12x20xbf16> loc(#loc84) + } -> tensor<1x80x12x20xbf16> loc(#loc84) + xten_nn.output %464 : tensor<1x80x12x20xbf16> loc(#loc84) + } -> tensor<1x80x12x20xbf16> loc(#loc84) + %232 = xten_nn.subgraph (%arg5 = %231: tensor<1x80x12x20xbf16>, %arg6 = %105: tensor<200x80x1x1xbf16>, %arg7 = %104: tensor<200xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[56, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[56, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 3870 : ui64, ofm = 7740 : ui64, wts = 2836 : ui64}, + computation_cycles = 9533 : ui64, + cycle_count = 9533 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<200x80x1x1xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 1391 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x200x12x20xbf16> loc(#loc85) + xten_nn.output %465 : tensor<1x200x12x20xbf16> loc(#loc85) + } -> tensor<1x200x12x20xbf16> loc(#loc85) + xten_nn.output %464 : tensor<1x200x12x20xbf16> loc(#loc85) + } -> tensor<1x200x12x20xbf16> loc(#loc85) + %233 = xten_nn.subgraph (%arg5 = %232: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7700 : ui64, ofm = 3860 : ui64}, + computation_cycles = 765 : ui64, + cycle_count = 7700 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x200x12x20xbf16> loc(#loc86) + xten_nn.output %465 : tensor<1x200x12x20xbf16> loc(#loc86) + } -> tensor<1x200x12x20xbf16> loc(#loc86) + xten_nn.output %464 : tensor<1x200x12x20xbf16> loc(#loc86) + } -> tensor<1x200x12x20xbf16> loc(#loc86) + %234 = xten_nn.subgraph (%arg5 = %233: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7700 : ui64, ofm = 3860 : ui64}, + computation_cycles = 775 : ui64, + cycle_count = 7700 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 139 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %465 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %464 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + %235 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7700 : ui64, ofm = 3860 : ui64}, + computation_cycles = 765 : ui64, + cycle_count = 7700 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %465 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %464 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + %236 = xten_nn.subgraph (%arg5 = %232: tensor<1x200x12x20xbf16>, %arg6 = %235: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7720 : ui64, ifm2 = 7720 : ui64, ofm = 3880 : ui64}, + computation_cycles = 932 : ui64, + cycle_count = 7720 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %465 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %464 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + %237 = xten_nn.subgraph (%arg5 = %236: tensor<1x200x12x20xbf16>, %arg6 = %103: tensor<200x1x3x3xbf16>, %arg7 = %102: tensor<200xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 8]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[24, 0, 0, 1]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[24, 0, 0, 1]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 4]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 4, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 18886 : ui64, ofm = 5446 : ui64, wts = 518 : ui64}, + computation_cycles = 8865 : ui64, + cycle_count = 18886 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x200x12x20xbf16>, %arg9 = %arg6: tensor<200x1x3x3xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 6 : ui8, + config.ifm_sv_width.size = 26 : ui8, + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 1099 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %465 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %464 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + %238 = xten_nn.subgraph (%arg5 = %237: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7700 : ui64, ofm = 3860 : ui64}, + computation_cycles = 765 : ui64, + cycle_count = 7700 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %465 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %464 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + %239 = xten_nn.subgraph (%arg5 = %238: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7700 : ui64, ofm = 3860 : ui64}, + computation_cycles = 775 : ui64, + cycle_count = 7700 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 139 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %465 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %464 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + %240 = xten_nn.subgraph (%arg5 = %239: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7700 : ui64, ofm = 3860 : ui64}, + computation_cycles = 765 : ui64, + cycle_count = 7700 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %465 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %464 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + %241 = xten_nn.subgraph (%arg5 = %237: tensor<1x200x12x20xbf16>, %arg6 = %240: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7720 : ui64, ifm2 = 7720 : ui64, ofm = 3880 : ui64}, + computation_cycles = 932 : ui64, + cycle_count = 7720 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %465 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %464 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + %242 = xten_nn.subgraph (%arg5 = %241: tensor<1x200x12x20xbf16>, %arg6 = %101: tensor<80x200x1x1xbf16>, %arg7 = %100: tensor<80xbf16>, %arg8 = %231: tensor<1x80x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_131", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 104, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 56, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[24, 104, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[16, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[16, 8, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 10044 : ui64, ifm2 = 4638 : ui64, ofm = 2334 : ui64, wts = 2708 : ui64}, + computation_cycles = 9960 : ui64, + cycle_count = 10044 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x200x12x20xbf16>, %arg10 = %arg6: tensor<80x200x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_131", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_131", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 104 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 24 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 2858 : ui64, + startup_cycles = 182 : ui64 + }} { + %466 = tensor.empty() : tensor<1x80x12x20xbf16> loc(#loc95) + xten_nn.output %466 : tensor<1x80x12x20xbf16> loc(#loc95) + } -> tensor<1x80x12x20xbf16> loc(#loc95) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_132", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 32 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 59 : ui64, + startup_cycles = 22 : ui64 + }} { + %466 = tensor.empty() : tensor<1x80x12x20xbf16> loc(#loc96) + xten_nn.output %466 : tensor<1x80x12x20xbf16> loc(#loc96) + } -> tensor<1x80x12x20xbf16> loc(#loc96) + xten_nn.output %465 : tensor<1x80x12x20xbf16> loc(#loc96) + } -> tensor<1x80x12x20xbf16> loc(#loc330) + %243 = xten_nn.subgraph (%arg5 = %242: tensor<1x80x12x20xbf16>, %arg6 = %99: tensor<184x80x1x1xbf16>, %arg7 = %98: tensor<184xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 3, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[48, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[8, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[8, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 48, 3, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 2910 : ui64, ofm = 7326 : ui64, wts = 2122 : ui64}, + computation_cycles = 7180 : ui64, + cycle_count = 7326 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 3 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 48 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 2142 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc97) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc97) + } -> tensor<1x184x12x20xbf16> loc(#loc97) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc97) + } -> tensor<1x184x12x20xbf16> loc(#loc97) + %244 = xten_nn.subgraph (%arg5 = %243: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 853 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc98) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc98) + } -> tensor<1x184x12x20xbf16> loc(#loc98) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc98) + } -> tensor<1x184x12x20xbf16> loc(#loc98) + %245 = xten_nn.subgraph (%arg5 = %244: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 871 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 79 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + %246 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 853 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + %247 = xten_nn.subgraph (%arg5 = %243: tensor<1x184x12x20xbf16>, %arg6 = %246: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ifm2 = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 751 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + %248 = xten_nn.subgraph (%arg5 = %247: tensor<1x184x12x20xbf16>, %arg6 = %97: tensor<184x1x3x3xbf16>, %arg7 = %96: tensor<184xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 8]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[8, 0, 0, 1]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[8, 0, 0, 1]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 4]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 4, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 16188 : ui64, ofm = 4668 : ui64, wts = 444 : ui64}, + computation_cycles = 7629 : ui64, + cycle_count = 16188 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 6 : ui8, + config.ifm_sv_width.size = 26 : ui8, + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 1099 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + %249 = xten_nn.subgraph (%arg5 = %248: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 853 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + %250 = xten_nn.subgraph (%arg5 = %249: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 871 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 79 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + %251 = xten_nn.subgraph (%arg5 = %250: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 853 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + %252 = xten_nn.subgraph (%arg5 = %248: tensor<1x184x12x20xbf16>, %arg6 = %251: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ifm2 = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 751 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + %253 = xten_nn.subgraph (%arg5 = %252: tensor<1x184x12x20xbf16>, %arg6 = %95: tensor<80x184x1x1xbf16>, %arg7 = %94: tensor<80xbf16>, %arg8 = %242: tensor<1x80x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_151", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 96, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[24, 96, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[16, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[16, 8, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 9276 : ui64, ifm2 = 4638 : ui64, ofm = 2334 : ui64, wts = 2516 : ui64}, + computation_cycles = 9474 : ui64, + cycle_count = 9474 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_151", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_151", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 96 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 24 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 2696 : ui64, + startup_cycles = 182 : ui64 + }} { + %466 = tensor.empty() : tensor<1x80x12x20xbf16> loc(#loc107) + xten_nn.output %466 : tensor<1x80x12x20xbf16> loc(#loc107) + } -> tensor<1x80x12x20xbf16> loc(#loc107) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_152", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 32 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 59 : ui64, + startup_cycles = 22 : ui64 + }} { + %466 = tensor.empty() : tensor<1x80x12x20xbf16> loc(#loc108) + xten_nn.output %466 : tensor<1x80x12x20xbf16> loc(#loc108) + } -> tensor<1x80x12x20xbf16> loc(#loc108) + xten_nn.output %465 : tensor<1x80x12x20xbf16> loc(#loc108) + } -> tensor<1x80x12x20xbf16> loc(#loc331) + %254 = xten_nn.subgraph (%arg5 = %253: tensor<1x80x12x20xbf16>, %arg6 = %93: tensor<184x80x1x1xbf16>, %arg7 = %92: tensor<184xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 3, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[48, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[8, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[8, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 48, 3, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 2910 : ui64, ofm = 7326 : ui64, wts = 2122 : ui64}, + computation_cycles = 7180 : ui64, + cycle_count = 7326 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 3 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 48 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 2142 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + %255 = xten_nn.subgraph (%arg5 = %254: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 853 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + %256 = xten_nn.subgraph (%arg5 = %255: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 871 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 79 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + %257 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 853 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + %258 = xten_nn.subgraph (%arg5 = %254: tensor<1x184x12x20xbf16>, %arg6 = %257: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ifm2 = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 751 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + %259 = xten_nn.subgraph (%arg5 = %258: tensor<1x184x12x20xbf16>, %arg6 = %91: tensor<184x1x3x3xbf16>, %arg7 = %90: tensor<184xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 8]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[8, 0, 0, 1]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[8, 0, 0, 1]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 4]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 4, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 16188 : ui64, ofm = 4668 : ui64, wts = 444 : ui64}, + computation_cycles = 7629 : ui64, + cycle_count = 16188 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 6 : ui8, + config.ifm_sv_width.size = 26 : ui8, + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 1099 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + %260 = xten_nn.subgraph (%arg5 = %259: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 853 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + %261 = xten_nn.subgraph (%arg5 = %260: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 871 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 79 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + %262 = xten_nn.subgraph (%arg5 = %261: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 853 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + %263 = xten_nn.subgraph (%arg5 = %259: tensor<1x184x12x20xbf16>, %arg6 = %262: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 5790 : ui64, ifm2 = 5790 : ui64, ofm = 2910 : ui64}, + computation_cycles = 751 : ui64, + cycle_count = 5790 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %465 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %464 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + %264 = xten_nn.subgraph (%arg5 = %263: tensor<1x184x12x20xbf16>, %arg6 = %89: tensor<80x184x1x1xbf16>, %arg7 = %88: tensor<80xbf16>, %arg8 = %253: tensor<1x80x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_171", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 96, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[24, 96, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[16, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[16, 8, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 9276 : ui64, ifm2 = 4638 : ui64, ofm = 2334 : ui64, wts = 2516 : ui64}, + computation_cycles = 9474 : ui64, + cycle_count = 9474 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_171", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_171", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 96 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 24 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 2696 : ui64, + startup_cycles = 182 : ui64 + }} { + %466 = tensor.empty() : tensor<1x80x12x20xbf16> loc(#loc119) + xten_nn.output %466 : tensor<1x80x12x20xbf16> loc(#loc119) + } -> tensor<1x80x12x20xbf16> loc(#loc119) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_172", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 32 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 59 : ui64, + startup_cycles = 22 : ui64 + }} { + %466 = tensor.empty() : tensor<1x80x12x20xbf16> loc(#loc120) + xten_nn.output %466 : tensor<1x80x12x20xbf16> loc(#loc120) + } -> tensor<1x80x12x20xbf16> loc(#loc120) + xten_nn.output %465 : tensor<1x80x12x20xbf16> loc(#loc120) + } -> tensor<1x80x12x20xbf16> loc(#loc332) + %265 = xten_nn.subgraph (%arg5 = %264: tensor<1x80x12x20xbf16>, %arg6 = %87: tensor<480x80x1x1xbf16>, %arg7 = %86: tensor<480xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[40, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 40, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 3870 : ui64, ofm = 13914 : ui64, wts = 5310 : ui64}, + computation_cycles = 15962 : ui64, + cycle_count = 15962 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<480x80x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 40 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 1596 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc121) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc121) + } -> tensor<1x480x12x20xbf16> loc(#loc121) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc121) + } -> tensor<1x480x12x20xbf16> loc(#loc121) + %266 = xten_nn.subgraph (%arg5 = %265: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15400 : ui64, ofm = 7720 : ui64}, + computation_cycles = 1301 : ui64, + cycle_count = 15400 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc122) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc122) + } -> tensor<1x480x12x20xbf16> loc(#loc122) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc122) + } -> tensor<1x480x12x20xbf16> loc(#loc122) + %267 = xten_nn.subgraph (%arg5 = %266: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15400 : ui64, ofm = 7720 : ui64}, + computation_cycles = 1327 : ui64, + cycle_count = 15400 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 139 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + %268 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15400 : ui64, ofm = 7720 : ui64}, + computation_cycles = 1301 : ui64, + cycle_count = 15400 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + %269 = xten_nn.subgraph (%arg5 = %265: tensor<1x480x12x20xbf16>, %arg6 = %268: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15440 : ui64, ifm2 = 15440 : ui64, ofm = 7760 : ui64}, + computation_cycles = 1656 : ui64, + cycle_count = 15440 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + %270 = xten_nn.subgraph (%arg5 = %269: tensor<1x480x12x20xbf16>, %arg6 = %85: tensor<480x1x3x3xbf16>, %arg7 = %84: tensor<480xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 8]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 4]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 4, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 40470 : ui64, ofm = 11670 : ui64, wts = 1110 : ui64}, + computation_cycles = 18753 : ui64, + cycle_count = 40470 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<480x1x3x3xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 6 : ui8, + config.ifm_sv_width.size = 26 : ui8, + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 1099 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + %271 = xten_nn.subgraph (%arg5 = %270: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15400 : ui64, ofm = 7720 : ui64}, + computation_cycles = 1301 : ui64, + cycle_count = 15400 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + %272 = xten_nn.subgraph (%arg5 = %271: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15400 : ui64, ofm = 7720 : ui64}, + computation_cycles = 1327 : ui64, + cycle_count = 15400 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 139 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + %273 = xten_nn.subgraph (%arg5 = %272: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15400 : ui64, ofm = 7720 : ui64}, + computation_cycles = 1301 : ui64, + cycle_count = 15400 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + %274 = xten_nn.subgraph (%arg5 = %270: tensor<1x480x12x20xbf16>, %arg6 = %273: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_190", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15440 : ui64, ifm2 = 15440 : ui64, ofm = 7760 : ui64}, + computation_cycles = 1656 : ui64, + cycle_count = 15440 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + %275 = xten_nn.subgraph (%arg5 = %274: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#24", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#25", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 60 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x480x1x240xbf16> loc(#loc333) + xten_nn.output %464 : tensor<1x480x1x240xbf16> loc(#loc333) + } -> tensor<1x480x1x240xbf16> loc(#loc333) + %276 = xten_nn.subgraph (%arg5 = %275: tensor<1x480x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 40, 1, 48]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 40, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 57750 : ui64, ofm = 4950 : ui64}, + computation_cycles = 45313 : ui64, + cycle_count = 57750 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x240xbf16>) attributes { + LayerName = "Generated-#26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#27", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 480 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W", + config.sv_channel = 40 : ui32, + config.sv_height = 1 : ui32, + config.sv_width = 48 : ui32 + }, + estimation = { + compute_cycles = 2311 : ui64, + startup_cycles = 190 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x1x1xbf16> loc(#loc131) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc131) + } -> tensor<1x480x1x1xbf16> loc(#loc131) + xten_nn.output %464 : tensor<1x480x1x1xbf16> loc(#loc131) + } -> tensor<1x480x1x1xbf16> loc(#loc131) + %277 = xten_nn.subgraph (%arg5 = %276: tensor<1x480x1x1xbf16>, %arg6 = %83: tensor<120x480x1x1xbf16>, %arg7 = %82: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[8, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[8, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1960 : ui64, ofm = 650 : ui64, wts = 8232 : ui64}, + computation_cycles = 5443 : ui64, + cycle_count = 8232 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x1x1xbf16>, %arg9 = %arg6: tensor<120x480x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 120 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 4 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 4552 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x120x1x1xbf16> loc(#loc133) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc133) + } -> tensor<1x120x1x1xbf16> loc(#loc334) + xten_nn.output %464 : tensor<1x120x1x1xbf16> loc(#loc334) + } -> tensor<1x120x1x1xbf16> loc(#loc334) + %278 = xten_nn.subgraph (%arg5 = %277: tensor<1x120x1x1xbf16>, %arg6 = %81: tensor<480x120x1x1xbf16>, %arg7 = %80: tensor<480xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[32, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[32, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 490 : ui64, ofm = 2600 : ui64, wts = 8232 : ui64}, + computation_cycles = 5443 : ui64, + cycle_count = 8232 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<480x120x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 120 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 1138 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x1x1xbf16> loc(#loc134) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc134) + } -> tensor<1x480x1x1xbf16> loc(#loc134) + xten_nn.output %464 : tensor<1x480x1x1xbf16> loc(#loc134) + } -> tensor<1x480x1x1xbf16> loc(#loc134) + %279 = xten_nn.subgraph (%arg5 = %278: tensor<1x480x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 2, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 2, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 11, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 778 : ui64, ofm = 394 : ui64}, + computation_cycles = 401 : ui64, + cycle_count = 778 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 128 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x1x1xbf16> loc(#loc135) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc135) + } -> tensor<1x480x1x1xbf16> loc(#loc135) + xten_nn.output %464 : tensor<1x480x1x1xbf16> loc(#loc135) + } -> tensor<1x480x1x1xbf16> loc(#loc135) + %280 = xten_nn.subgraph (%arg5 = %279: tensor<1x480x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 11, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 7, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 522 : ui64, ofm = 266 : ui64}, + computation_cycles = 395 : ui64, + cycle_count = 522 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 128 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %464 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + %281 = xten_nn.subgraph (%arg5 = %280: tensor<1x480x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 7, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 15, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1034 : ui64, ofm = 522 : ui64}, + computation_cycles = 409 : ui64, + cycle_count = 1034 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 128 : ui32, + config.ifmsv_height = 4 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 43 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %464 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + %282 = xten_nn.subgraph (%arg5 = %281: tensor<1x480x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#28", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#29", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 480 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + %283 = xten_nn.subgraph (%arg5 = %282: tensor<1x480x12x20xbf16>, %arg6 = %274: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_202", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15440 : ui64, ifm2 = 15440 : ui64, ofm = 7760 : ui64}, + computation_cycles = 1656 : ui64, + cycle_count = 15440 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %465 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %464 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + %284 = xten_nn.subgraph (%arg5 = %283: tensor<1x480x12x20xbf16>, %arg6 = %79: tensor<112x480x1x1xbf16>, %arg7 = %78: tensor<112xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 96, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 96, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 23190 : ui64, ofm = 6942 : ui64, wts = 8370 : ui64}, + computation_cycles = 25063 : ui64, + cycle_count = 25063 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<112x480x1x1xbf16>, %arg10 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 96 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 5 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 7555 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x112x12x20xbf16> loc(#loc139) + xten_nn.output %465 : tensor<1x112x12x20xbf16> loc(#loc139) + } -> tensor<1x112x12x20xbf16> loc(#loc139) + xten_nn.output %464 : tensor<1x112x12x20xbf16> loc(#loc139) + } -> tensor<1x112x12x20xbf16> loc(#loc139) + %285 = xten_nn.subgraph (%arg5 = %284: tensor<1x112x12x20xbf16>, %arg6 = %77: tensor<672x112x1x1xbf16>, %arg7 = %76: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 64, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[56, 64, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 56, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 6204 : ui64, ofm = 18522 : ui64, wts = 12156 : ui64}, + computation_cycles = 34007 : ui64, + cycle_count = 34007 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 56 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 3464 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc140) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc140) + } -> tensor<1x672x12x20xbf16> loc(#loc140) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc140) + } -> tensor<1x672x12x20xbf16> loc(#loc140) + %286 = xten_nn.subgraph (%arg5 = %285: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2517 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc141) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc141) + } -> tensor<1x672x12x20xbf16> loc(#loc141) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc141) + } -> tensor<1x672x12x20xbf16> loc(#loc141) + %287 = xten_nn.subgraph (%arg5 = %286: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2599 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 79 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + %288 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2517 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + %289 = xten_nn.subgraph (%arg5 = %285: tensor<1x672x12x20xbf16>, %arg6 = %288: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ifm2 = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2199 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + %290 = xten_nn.subgraph (%arg5 = %289: tensor<1x672x12x20xbf16>, %arg6 = %75: tensor<672x1x3x3xbf16>, %arg7 = %74: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 8]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[0, 0, 0, 1]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 4]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 4, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 56658 : ui64, ofm = 16338 : ui64, wts = 1554 : ui64}, + computation_cycles = 26169 : ui64, + cycle_count = 56658 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x3x3xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 6 : ui8, + config.ifm_sv_width.size = 26 : ui8, + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 1099 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + %291 = xten_nn.subgraph (%arg5 = %290: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2517 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + %292 = xten_nn.subgraph (%arg5 = %291: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2599 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 79 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + %293 = xten_nn.subgraph (%arg5 = %292: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2517 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + %294 = xten_nn.subgraph (%arg5 = %290: tensor<1x672x12x20xbf16>, %arg6 = %293: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_221", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ifm2 = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2199 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + %295 = xten_nn.subgraph (%arg5 = %294: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#30", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#31", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x672x1x240xbf16> loc(#loc335) + xten_nn.output %464 : tensor<1x672x1x240xbf16> loc(#loc335) + } -> tensor<1x672x1x240xbf16> loc(#loc335) + %296 = xten_nn.subgraph (%arg5 = %295: tensor<1x672x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 56, 1, 32]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 56, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 86256 : ui64, ofm = 10992 : ui64}, + computation_cycles = 81517 : ui64, + cycle_count = 86256 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#33", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W", + config.sv_channel = 56 : ui32, + config.sv_height = 1 : ui32, + config.sv_width = 32 : ui32 + }, + estimation = { + compute_cycles = 2285 : ui64, + startup_cycles = 190 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x1x1xbf16> loc(#loc150) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc150) + } -> tensor<1x672x1x1xbf16> loc(#loc150) + xten_nn.output %464 : tensor<1x672x1x1xbf16> loc(#loc150) + } -> tensor<1x672x1x1xbf16> loc(#loc150) + %297 = xten_nn.subgraph (%arg5 = %296: tensor<1x672x1x1xbf16>, %arg6 = %73: tensor<168x672x1x1xbf16>, %arg7 = %72: tensor<168xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 96, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[48, 96, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[24, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[24, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 48, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 2758 : ui64, ofm = 906 : ui64, wts = 17542 : ui64}, + computation_cycles = 10332 : ui64, + cycle_count = 17542 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 96 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 7 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 48 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 9030 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x168x1x1xbf16> loc(#loc152) + xten_nn.output %465 : tensor<1x168x1x1xbf16> loc(#loc152) + } -> tensor<1x168x1x1xbf16> loc(#loc336) + xten_nn.output %464 : tensor<1x168x1x1xbf16> loc(#loc336) + } -> tensor<1x168x1x1xbf16> loc(#loc336) + %298 = xten_nn.subgraph (%arg5 = %297: tensor<1x168x1x1xbf16>, %arg6 = %71: tensor<672x168x1x1xbf16>, %arg7 = %70: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 168, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[32, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[32, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1354 : ui64, ofm = 8558 : ui64, wts = 15598 : ui64}, + computation_cycles = 16150 : ui64, + cycle_count = 16150 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 168 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 1300 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x1x1xbf16> loc(#loc153) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc153) + } -> tensor<1x672x1x1xbf16> loc(#loc153) + xten_nn.output %464 : tensor<1x672x1x1xbf16> loc(#loc153) + } -> tensor<1x672x1x1xbf16> loc(#loc153) + %299 = xten_nn.subgraph (%arg5 = %298: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 192, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 192, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 96, 7, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 778 : ui64, ofm = 394 : ui64}, + computation_cycles = 401 : ui64, + cycle_count = 778 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 192 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x1x1xbf16> loc(#loc154) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc154) + } -> tensor<1x672x1x1xbf16> loc(#loc154) + xten_nn.output %464 : tensor<1x672x1x1xbf16> loc(#loc154) + } -> tensor<1x672x1x1xbf16> loc(#loc154) + %300 = xten_nn.subgraph (%arg5 = %299: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 96, 7, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 352, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 522 : ui64, ofm = 266 : ui64}, + computation_cycles = 395 : ui64, + cycle_count = 522 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 256 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %464 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + %301 = xten_nn.subgraph (%arg5 = %300: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 352, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 352, 7, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1034 : ui64, ofm = 522 : ui64}, + computation_cycles = 409 : ui64, + cycle_count = 1034 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 256 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 43 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %464 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + %302 = xten_nn.subgraph (%arg5 = %301: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#34", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#35", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + %303 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>, %arg6 = %294: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_233", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ifm2 = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2199 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + %304 = xten_nn.subgraph (%arg5 = %303: tensor<1x672x12x20xbf16>, %arg6 = %69: tensor<112x672x1x1xbf16>, %arg7 = %68: tensor<112xbf16>, %arg8 = %284: tensor<1x112x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_234", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 96, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 96, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 32466 : ui64, ifm2 = 6174 : ui64, ofm = 3102 : ui64, wts = 11718 : ui64}, + computation_cycles = 35198 : ui64, + cycle_count = 35198 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x672x12x20xbf16>, %arg10 = %arg6: tensor<112x672x1x1xbf16>, %arg11 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_234", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_234", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 96 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 7 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 10577 : ui64, + startup_cycles = 160 : ui64 + }} { + %466 = tensor.empty() : tensor<1x112x12x20xbf16> loc(#loc158) + xten_nn.output %466 : tensor<1x112x12x20xbf16> loc(#loc158) + } -> tensor<1x112x12x20xbf16> loc(#loc158) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x112x12x20xbf16>, %arg10 = %arg8: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Add_235", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 32 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 75 : ui64, + startup_cycles = 22 : ui64 + }} { + %466 = tensor.empty() : tensor<1x112x12x20xbf16> loc(#loc159) + xten_nn.output %466 : tensor<1x112x12x20xbf16> loc(#loc159) + } -> tensor<1x112x12x20xbf16> loc(#loc159) + xten_nn.output %465 : tensor<1x112x12x20xbf16> loc(#loc159) + } -> tensor<1x112x12x20xbf16> loc(#loc337) + %305 = xten_nn.subgraph (%arg5 = %304: tensor<1x112x12x20xbf16>, %arg6 = %67: tensor<672x112x1x1xbf16>, %arg7 = %66: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 64, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[56, 64, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 56, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 6204 : ui64, ofm = 18522 : ui64, wts = 12156 : ui64}, + computation_cycles = 34007 : ui64, + cycle_count = 34007 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 56 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 3464 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc160) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc160) + } -> tensor<1x672x12x20xbf16> loc(#loc160) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc160) + } -> tensor<1x672x12x20xbf16> loc(#loc160) + %306 = xten_nn.subgraph (%arg5 = %305: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2517 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc161) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc161) + } -> tensor<1x672x12x20xbf16> loc(#loc161) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc161) + } -> tensor<1x672x12x20xbf16> loc(#loc161) + %307 = xten_nn.subgraph (%arg5 = %306: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2599 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 79 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + %308 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2517 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + %309 = xten_nn.subgraph (%arg5 = %305: tensor<1x672x12x20xbf16>, %arg6 = %308: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ifm2 = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2199 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + %310 = xten_nn.subgraph (%arg5 = %309: tensor<1x672x12x20xbf16>, %arg6 = %65: tensor<672x1x9x9xbf16>, %arg7 = %64: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 10, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 9, 9]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex>, + l3_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 2, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 323820 : ui64, ofm = 17388 : ui64, wts = 9618 : ui64}, + computation_cycles = 110589 : ui64, + cycle_count = 323820 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x9x9xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 10 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 739 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + %311 = xten_nn.subgraph (%arg5 = %310: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2517 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + %312 = xten_nn.subgraph (%arg5 = %311: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2599 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 79 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + %313 = xten_nn.subgraph (%arg5 = %312: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2517 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + %314 = xten_nn.subgraph (%arg5 = %310: tensor<1x672x12x20xbf16>, %arg6 = %313: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_253", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ifm2 = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2199 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + %315 = xten_nn.subgraph (%arg5 = %314: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#36", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#37", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x672x1x240xbf16> loc(#loc338) + xten_nn.output %464 : tensor<1x672x1x240xbf16> loc(#loc338) + } -> tensor<1x672x1x240xbf16> loc(#loc338) + %316 = xten_nn.subgraph (%arg5 = %315: tensor<1x672x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#38", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 56, 1, 32]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#39", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 56, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 86256 : ui64, ofm = 10992 : ui64}, + computation_cycles = 81517 : ui64, + cycle_count = 86256 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#38", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#39", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W", + config.sv_channel = 56 : ui32, + config.sv_height = 1 : ui32, + config.sv_width = 32 : ui32 + }, + estimation = { + compute_cycles = 2285 : ui64, + startup_cycles = 190 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x1x1xbf16> loc(#loc170) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc170) + } -> tensor<1x672x1x1xbf16> loc(#loc170) + xten_nn.output %464 : tensor<1x672x1x1xbf16> loc(#loc170) + } -> tensor<1x672x1x1xbf16> loc(#loc170) + %317 = xten_nn.subgraph (%arg5 = %316: tensor<1x672x1x1xbf16>, %arg6 = %63: tensor<168x672x1x1xbf16>, %arg7 = %62: tensor<168xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 96, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[48, 96, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[24, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[24, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 48, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 2758 : ui64, ofm = 906 : ui64, wts = 17542 : ui64}, + computation_cycles = 10332 : ui64, + cycle_count = 17542 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 96 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 7 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 48 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 9030 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x168x1x1xbf16> loc(#loc172) + xten_nn.output %465 : tensor<1x168x1x1xbf16> loc(#loc172) + } -> tensor<1x168x1x1xbf16> loc(#loc339) + xten_nn.output %464 : tensor<1x168x1x1xbf16> loc(#loc339) + } -> tensor<1x168x1x1xbf16> loc(#loc339) + %318 = xten_nn.subgraph (%arg5 = %317: tensor<1x168x1x1xbf16>, %arg6 = %61: tensor<672x168x1x1xbf16>, %arg7 = %60: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 168, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[32, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[32, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1354 : ui64, ofm = 8558 : ui64, wts = 15598 : ui64}, + computation_cycles = 16150 : ui64, + cycle_count = 16150 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 168 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 1300 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x1x1xbf16> loc(#loc173) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc173) + } -> tensor<1x672x1x1xbf16> loc(#loc173) + xten_nn.output %464 : tensor<1x672x1x1xbf16> loc(#loc173) + } -> tensor<1x672x1x1xbf16> loc(#loc173) + %319 = xten_nn.subgraph (%arg5 = %318: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 192, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 192, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 96, 7, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 778 : ui64, ofm = 394 : ui64}, + computation_cycles = 401 : ui64, + cycle_count = 778 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 192 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x1x1xbf16> loc(#loc174) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc174) + } -> tensor<1x672x1x1xbf16> loc(#loc174) + xten_nn.output %464 : tensor<1x672x1x1xbf16> loc(#loc174) + } -> tensor<1x672x1x1xbf16> loc(#loc174) + %320 = xten_nn.subgraph (%arg5 = %319: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 96, 7, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 352, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 522 : ui64, ofm = 266 : ui64}, + computation_cycles = 395 : ui64, + cycle_count = 522 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 256 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %464 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + %321 = xten_nn.subgraph (%arg5 = %320: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 352, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 352, 7, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1034 : ui64, ofm = 522 : ui64}, + computation_cycles = 409 : ui64, + cycle_count = 1034 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 256 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 43 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %464 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + %322 = xten_nn.subgraph (%arg5 = %321: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#41", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + %323 = xten_nn.subgraph (%arg5 = %322: tensor<1x672x12x20xbf16>, %arg6 = %314: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_265", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 21230 : ui64, ifm2 = 21230 : ui64, ofm = 10670 : ui64}, + computation_cycles = 2199 : ui64, + cycle_count = 21230 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %465 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %464 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + %324 = xten_nn.subgraph (%arg5 = %323: tensor<1x672x12x20xbf16>, %arg6 = %59: tensor<160x672x1x1xbf16>, %arg7 = %58: tensor<160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[40, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 48, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[0, 48, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 40, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 34830 : ui64, ofm = 4638 : ui64, wts = 15930 : ui64}, + computation_cycles = 47156 : ui64, + cycle_count = 47156 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<160x672x1x1xbf16>, %arg10 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 9 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 40 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 14364 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x160x12x20xbf16> loc(#loc178) + xten_nn.output %465 : tensor<1x160x12x20xbf16> loc(#loc178) + } -> tensor<1x160x12x20xbf16> loc(#loc178) + xten_nn.output %464 : tensor<1x160x12x20xbf16> loc(#loc178) + } -> tensor<1x160x12x20xbf16> loc(#loc178) + %325 = xten_nn.subgraph (%arg5 = %324: tensor<1x160x12x20xbf16>, %arg6 = %57: tensor<960x160x1x1xbf16>, %arg7 = %56: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[40, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 40, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7740 : ui64, ofm = 27828 : ui64, wts = 21240 : ui64}, + computation_cycles = 62753 : ui64, + cycle_count = 62753 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 40 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 3192 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc179) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc179) + } -> tensor<1x960x12x20xbf16> loc(#loc179) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc179) + } -> tensor<1x960x12x20xbf16> loc(#loc179) + %326 = xten_nn.subgraph (%arg5 = %325: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2461 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 111 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc180) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc180) + } -> tensor<1x960x12x20xbf16> loc(#loc180) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc180) + } -> tensor<1x960x12x20xbf16> loc(#loc180) + %327 = xten_nn.subgraph (%arg5 = %326: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2527 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 119 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + %328 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2461 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 111 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + %329 = xten_nn.subgraph (%arg5 = %325: tensor<1x960x12x20xbf16>, %arg6 = %328: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 38600 : ui64, ifm2 = 38600 : ui64, ofm = 19400 : ui64}, + computation_cycles = 3828 : ui64, + cycle_count = 77200 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + %330 = xten_nn.subgraph (%arg5 = %329: tensor<1x960x12x20xbf16>, %arg6 = %55: tensor<960x1x9x9xbf16>, %arg7 = %54: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 10, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 9, 9]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex>, + l3_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 2, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 462600 : ui64, ofm = 24840 : ui64, wts = 13740 : ui64}, + computation_cycles = 157893 : ui64, + cycle_count = 462600 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 10 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 739 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + %331 = xten_nn.subgraph (%arg5 = %330: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2461 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 111 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + %332 = xten_nn.subgraph (%arg5 = %331: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2527 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 119 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + %333 = xten_nn.subgraph (%arg5 = %332: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2461 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 111 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + %334 = xten_nn.subgraph (%arg5 = %330: tensor<1x960x12x20xbf16>, %arg6 = %333: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_284", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 38600 : ui64, ifm2 = 38600 : ui64, ofm = 19400 : ui64}, + computation_cycles = 3828 : ui64, + cycle_count = 77200 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + %335 = xten_nn.subgraph (%arg5 = %334: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#42", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#43", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x960x1x240xbf16> loc(#loc340) + xten_nn.output %464 : tensor<1x960x1x240xbf16> loc(#loc340) + } -> tensor<1x960x1x240xbf16> loc(#loc340) + %336 = xten_nn.subgraph (%arg5 = %335: tensor<1x960x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#44", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 40, 1, 48]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#45", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 40, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 115500 : ui64, ofm = 9900 : ui64}, + computation_cycles = 90253 : ui64, + cycle_count = 115500 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#44", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#45", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W", + config.sv_channel = 40 : ui32, + config.sv_height = 1 : ui32, + config.sv_width = 48 : ui32 + }, + estimation = { + compute_cycles = 2311 : ui64, + startup_cycles = 190 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc189) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc189) + } -> tensor<1x960x1x1xbf16> loc(#loc189) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc189) + } -> tensor<1x960x1x1xbf16> loc(#loc189) + %337 = xten_nn.subgraph (%arg5 = %336: tensor<1x960x1x1xbf16>, %arg6 = %53: tensor<240x960x1x1xbf16>, %arg7 = %52: tensor<240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 192, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 192, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7730 : ui64, ofm = 3112 : ui64, wts = 32200 : ui64}, + computation_cycles = 31903 : ui64, + cycle_count = 32200 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 192 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 5 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 7205 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x1x1xbf16> loc(#loc191) + xten_nn.output %465 : tensor<1x240x1x1xbf16> loc(#loc191) + } -> tensor<1x240x1x1xbf16> loc(#loc341) + xten_nn.output %464 : tensor<1x240x1x1xbf16> loc(#loc341) + } -> tensor<1x240x1x1xbf16> loc(#loc341) + %338 = xten_nn.subgraph (%arg5 = %337: tensor<1x240x1x1xbf16>, %arg6 = %51: tensor<960x240x1x1xbf16>, %arg7 = %50: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[48, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 48, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 990 : ui64, ofm = 4530 : ui64, wts = 31830 : ui64}, + computation_cycles = 19498 : ui64, + cycle_count = 31830 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 3 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 48 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 3420 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc192) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc192) + } -> tensor<1x960x1x1xbf16> loc(#loc192) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc192) + } -> tensor<1x960x1x1xbf16> loc(#loc192) + %339 = xten_nn.subgraph (%arg5 = %338: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 384, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 384, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 576, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 778 : ui64, ofm = 394 : ui64}, + computation_cycles = 401 : ui64, + cycle_count = 778 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 384 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc193) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc193) + } -> tensor<1x960x1x1xbf16> loc(#loc193) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc193) + } -> tensor<1x960x1x1xbf16> loc(#loc193) + %340 = xten_nn.subgraph (%arg5 = %339: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 576, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 64, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 522 : ui64, ofm = 266 : ui64}, + computation_cycles = 395 : ui64, + cycle_count = 522 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 256 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + %341 = xten_nn.subgraph (%arg5 = %340: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 64, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 64, 7, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1034 : ui64, ofm = 522 : ui64}, + computation_cycles = 409 : ui64, + cycle_count = 1034 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 256 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 43 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + %342 = xten_nn.subgraph (%arg5 = %341: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#46", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#47", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + %343 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>, %arg6 = %334: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_296", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28950 : ui64, ifm2 = 28950 : ui64, ofm = 14550 : ui64}, + computation_cycles = 2923 : ui64, + cycle_count = 28950 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + %344 = xten_nn.subgraph (%arg5 = %343: tensor<1x960x12x20xbf16>, %arg6 = %49: tensor<160x960x1x1xbf16>, %arg7 = %48: tensor<160xbf16>, %arg8 = %324: tensor<1x160x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_297", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 112, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[24, 112, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[32, 48, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[32, 48, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 48654 : ui64, ifm2 = 9276 : ui64, ofm = 4668 : ui64, wts = 26100 : ui64}, + computation_cycles = 89679 : ui64, + cycle_count = 89679 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_297", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_297", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 112 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 9 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 24 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 13590 : ui64, + startup_cycles = 182 : ui64 + }} { + %466 = tensor.empty() : tensor<1x160x12x20xbf16> loc(#loc197) + xten_nn.output %466 : tensor<1x160x12x20xbf16> loc(#loc197) + } -> tensor<1x160x12x20xbf16> loc(#loc197) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_298", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 32 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 59 : ui64, + startup_cycles = 22 : ui64 + }} { + %466 = tensor.empty() : tensor<1x160x12x20xbf16> loc(#loc198) + xten_nn.output %466 : tensor<1x160x12x20xbf16> loc(#loc198) + } -> tensor<1x160x12x20xbf16> loc(#loc198) + xten_nn.output %465 : tensor<1x160x12x20xbf16> loc(#loc198) + } -> tensor<1x160x12x20xbf16> loc(#loc342) + %345 = xten_nn.subgraph (%arg5 = %344: tensor<1x160x12x20xbf16>, %arg6 = %47: tensor<960x160x1x1xbf16>, %arg7 = %46: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[40, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 40, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7740 : ui64, ofm = 27828 : ui64, wts = 21240 : ui64}, + computation_cycles = 62753 : ui64, + cycle_count = 62753 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 40 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 3192 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc199) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc199) + } -> tensor<1x960x12x20xbf16> loc(#loc199) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc199) + } -> tensor<1x960x12x20xbf16> loc(#loc199) + %346 = xten_nn.subgraph (%arg5 = %345: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2461 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 111 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc200) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc200) + } -> tensor<1x960x12x20xbf16> loc(#loc200) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc200) + } -> tensor<1x960x12x20xbf16> loc(#loc200) + %347 = xten_nn.subgraph (%arg5 = %346: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2527 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 119 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + %348 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2461 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 111 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + %349 = xten_nn.subgraph (%arg5 = %345: tensor<1x960x12x20xbf16>, %arg6 = %348: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 38600 : ui64, ifm2 = 38600 : ui64, ofm = 19400 : ui64}, + computation_cycles = 3828 : ui64, + cycle_count = 77200 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + %350 = xten_nn.subgraph (%arg5 = %349: tensor<1x960x12x20xbf16>, %arg6 = %45: tensor<960x1x9x9xbf16>, %arg7 = %44: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 10, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 1, 9, 9]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex>, + l3_extend_end = dense<[0, 0, 0, 3]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 2, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 462600 : ui64, ofm = 24840 : ui64, wts = 13740 : ui64}, + computation_cycles = 157893 : ui64, + cycle_count = 462600 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 4, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 10 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8, + data_io.ifm.dimensions.batch.padding = 0 : ui8, + data_io.ifm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.batch.padding = 0 : ui8, + data_io.ofm.dimensions.batch.size = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8 + }, + estimation = { + compute_cycles = 739 : ui64, + startup_cycles = 30 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + %351 = xten_nn.subgraph (%arg5 = %350: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 4, 4]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2461 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 111 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + %352 = xten_nn.subgraph (%arg5 = %351: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2527 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 119 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + %353 = xten_nn.subgraph (%arg5 = %352: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2461 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 111 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + %354 = xten_nn.subgraph (%arg5 = %350: tensor<1x960x12x20xbf16>, %arg6 = %353: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_316", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 38600 : ui64, ifm2 = 38600 : ui64, ofm = 19400 : ui64}, + computation_cycles = 3828 : ui64, + cycle_count = 77200 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + %355 = xten_nn.subgraph (%arg5 = %354: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#48", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#49", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x960x1x240xbf16> loc(#loc343) + xten_nn.output %464 : tensor<1x960x1x240xbf16> loc(#loc343) + } -> tensor<1x960x1x240xbf16> loc(#loc343) + %356 = xten_nn.subgraph (%arg5 = %355: tensor<1x960x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 40, 1, 48]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#51", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 40, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 115500 : ui64, ofm = 9900 : ui64}, + computation_cycles = 90253 : ui64, + cycle_count = 115500 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#51", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W", + config.sv_channel = 40 : ui32, + config.sv_height = 1 : ui32, + config.sv_width = 48 : ui32 + }, + estimation = { + compute_cycles = 2311 : ui64, + startup_cycles = 190 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc209) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc209) + } -> tensor<1x960x1x1xbf16> loc(#loc209) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc209) + } -> tensor<1x960x1x1xbf16> loc(#loc209) + %357 = xten_nn.subgraph (%arg5 = %356: tensor<1x960x1x1xbf16>, %arg6 = %43: tensor<240x960x1x1xbf16>, %arg7 = %42: tensor<240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 192, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 192, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7730 : ui64, ofm = 3112 : ui64, wts = 32200 : ui64}, + computation_cycles = 31903 : ui64, + cycle_count = 32200 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 192 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 5 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 7205 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x240x1x1xbf16> loc(#loc211) + xten_nn.output %465 : tensor<1x240x1x1xbf16> loc(#loc211) + } -> tensor<1x240x1x1xbf16> loc(#loc344) + xten_nn.output %464 : tensor<1x240x1x1xbf16> loc(#loc344) + } -> tensor<1x240x1x1xbf16> loc(#loc344) + %358 = xten_nn.subgraph (%arg5 = %357: tensor<1x240x1x1xbf16>, %arg6 = %41: tensor<960x240x1x1xbf16>, %arg7 = %40: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[48, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 7]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 48, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 990 : ui64, ofm = 4530 : ui64, wts = 31830 : ui64}, + computation_cycles = 19498 : ui64, + cycle_count = 31830 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 8 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 3 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 48 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 3420 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc212) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc212) + } -> tensor<1x960x1x1xbf16> loc(#loc212) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc212) + } -> tensor<1x960x1x1xbf16> loc(#loc212) + %359 = xten_nn.subgraph (%arg5 = %358: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 384, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 31]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 384, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 576, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 778 : ui64, ofm = 394 : ui64}, + computation_cycles = 401 : ui64, + cycle_count = 778 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 384 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc213) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc213) + } -> tensor<1x960x1x1xbf16> loc(#loc213) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc213) + } -> tensor<1x960x1x1xbf16> loc(#loc213) + %360 = xten_nn.subgraph (%arg5 = %359: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 576, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 64, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 522 : ui64, ofm = 266 : ui64}, + computation_cycles = 395 : ui64, + cycle_count = 522 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 256 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 35 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + %361 = xten_nn.subgraph (%arg5 = %360: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 64, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 1, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 256, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 64, 7, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1034 : ui64, ofm = 522 : ui64}, + computation_cycles = 409 : ui64, + cycle_count = 1034 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 256 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 43 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + %362 = xten_nn.subgraph (%arg5 = %361: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#52", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#53", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + %363 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>, %arg6 = %354: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_328", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28950 : ui64, ifm2 = 28950 : ui64, ofm = 14550 : ui64}, + computation_cycles = 2923 : ui64, + cycle_count = 28950 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + %364 = xten_nn.subgraph (%arg5 = %363: tensor<1x960x12x20xbf16>, %arg6 = %39: tensor<160x960x1x1xbf16>, %arg7 = %38: tensor<160xbf16>, %arg8 = %344: tensor<1x160x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_329", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 112, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[24, 112, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[32, 48, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[32, 48, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 48654 : ui64, ifm2 = 9276 : ui64, ofm = 4668 : ui64, wts = 26100 : ui64}, + computation_cycles = 89679 : ui64, + cycle_count = 89679 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_329", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_329", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 112 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 9 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 24 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 13590 : ui64, + startup_cycles = 182 : ui64 + }} { + %466 = tensor.empty() : tensor<1x160x12x20xbf16> loc(#loc217) + xten_nn.output %466 : tensor<1x160x12x20xbf16> loc(#loc217) + } -> tensor<1x160x12x20xbf16> loc(#loc217) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_330", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 32 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 59 : ui64, + startup_cycles = 22 : ui64 + }} { + %466 = tensor.empty() : tensor<1x160x12x20xbf16> loc(#loc218) + xten_nn.output %466 : tensor<1x160x12x20xbf16> loc(#loc218) + } -> tensor<1x160x12x20xbf16> loc(#loc218) + xten_nn.output %465 : tensor<1x160x12x20xbf16> loc(#loc218) + } -> tensor<1x160x12x20xbf16> loc(#loc345) + %365 = xten_nn.subgraph (%arg5 = %364: tensor<1x160x12x20xbf16>, %arg6 = %37: tensor<960x160x1x1xbf16>, %arg7 = %36: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 32, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[40, 80, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 40, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7740 : ui64, ofm = 27828 : ui64, wts = 21240 : ui64}, + computation_cycles = 62753 : ui64, + cycle_count = 62753 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 80 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 40 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 3192 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc219) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc219) + } -> tensor<1x960x12x20xbf16> loc(#loc219) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc219) + } -> tensor<1x960x12x20xbf16> loc(#loc219) + %366 = xten_nn.subgraph (%arg5 = %365: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2461 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 111 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc220) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc220) + } -> tensor<1x960x12x20xbf16> loc(#loc220) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc220) + } -> tensor<1x960x12x20xbf16> loc(#loc220) + %367 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2527 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 119 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + %368 = xten_nn.subgraph (%arg5 = %367: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 80, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28890 : ui64, ofm = 14490 : ui64}, + computation_cycles = 2461 : ui64, + cycle_count = 28890 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifmsv_depth = 80 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }, + estimation = { + compute_cycles = 111 : ui64, + startup_cycles = 46 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + %369 = xten_nn.subgraph (%arg5 = %365: tensor<1x960x12x20xbf16>, %arg6 = %368: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_339", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 2, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_force_single_buffering = true, + l2_tile_count = dense<[1, 960, 6, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 38600 : ui64, ifm2 = 38600 : ui64, ofm = 19400 : ui64}, + computation_cycles = 3828 : ui64, + cycle_count = 77200 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %465 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %464 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + %370 = xten_nn.subgraph (%arg5 = %369: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#54", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#55", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x960x1x240xbf16> loc(#loc346) + xten_nn.output %464 : tensor<1x960x1x240xbf16> loc(#loc346) + } -> tensor<1x960x1x240xbf16> loc(#loc346) + %371 = xten_nn.subgraph (%arg5 = %370: tensor<1x960x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 40, 1, 48]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 40, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 115500 : ui64, ofm = 9900 : ui64}, + computation_cycles = 90253 : ui64, + cycle_count = 115500 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#57", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W", + config.sv_channel = 40 : ui32, + config.sv_height = 1 : ui32, + config.sv_width = 48 : ui32 + }, + estimation = { + compute_cycles = 2311 : ui64, + startup_cycles = 190 : ui64 + }} { + %465 = tensor.empty() : tensor<1x960x1x1xbf16> loc(#loc224) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc224) + } -> tensor<1x960x1x1xbf16> loc(#loc224) + xten_nn.output %464 : tensor<1x960x1x1xbf16> loc(#loc224) + } -> tensor<1x960x1x1xbf16> loc(#loc224) + %372 = xten_nn.subgraph (%arg5 = %371: tensor<1x960x1x1xbf16>, %arg6 = %35: tensor<128x960x1x1xbf16>, %arg7 = %34: tensor<128xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 192, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 192, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 15]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7730 : ui64, ofm = 1556 : ui64, wts = 16100 : ui64}, + computation_cycles = 16123 : ui64, + cycle_count = 16123 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[3, 1, 2, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<128x960x1x1xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 12 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 192 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 5 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 7205 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x128x1x1xbf16> loc(#loc225) + xten_nn.output %465 : tensor<1x128x1x1xbf16> loc(#loc225) + } -> tensor<1x128x1x1xbf16> loc(#loc225) + xten_nn.output %464 : tensor<1x128x1x1xbf16> loc(#loc225) + } -> tensor<1x128x1x1xbf16> loc(#loc225) + %373 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 15, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 63]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 15, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 63, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1034 : ui64, ofm = 522 : ui64}, + computation_cycles = 490 : ui64, + cycle_count = 1034 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 16 : ui32, + config.ifmsv_width = 1 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 139 : ui64, + startup_cycles = 31 : ui64 + }} { + %465 = tensor.empty() : tensor<1x128x1x1xbf16> loc(#loc226) + xten_nn.output %465 : tensor<1x128x1x1xbf16> loc(#loc226) + } -> tensor<1x128x1x1xbf16> loc(#loc226) + xten_nn.output %464 : tensor<1x128x1x1xbf16> loc(#loc226) + } -> tensor<1x128x1x1xbf16> loc(#loc226) + %374 = xten_nn.subgraph (%arg5 = %373: tensor<1x128x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#58", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#59", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 128 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %464 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc227) + %375 = xten_nn.subgraph (%arg5 = %369: tensor<1x960x12x20xbf16>, %arg6 = %33: tensor<128x960x1x1xbf16>, %arg7 = %32: tensor<128xbf16>, %arg8 = %374: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_340", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 96, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[32, 96, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 1, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 46380 : ui64, ifm2 = 6174 : ui64, ofm = 3102 : ui64, wts = 16740 : ui64}, + computation_cycles = 49946 : ui64, + cycle_count = 49946 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<128x960x1x1xbf16>, %arg11 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_340", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_341", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 96 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 32 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 10 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 32 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 15110 : ui64, + startup_cycles = 160 : ui64 + }} { + %466 = tensor.empty() : tensor<1x128x12x20xbf16> loc(#loc229) + xten_nn.output %466 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc348) + %465 = xten_nn.subgraph (%arg9 = %464: tensor<1x128x12x20xbf16>, %arg10 = %arg8: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Mul_345", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 1 : ui32, + config.ifmsv_width = 32 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 46 : ui64, + startup_cycles = 25 : ui64 + }} { + %466 = tensor.empty() : tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %466 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %465 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc347) + %376 = xten_nn.subgraph (%arg5 = %375: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_349_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc230) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc230) + } -> tensor<1x64x12x20xbf16> loc(#loc230) + %377 = xten_nn.subgraph (%arg5 = %375: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_349_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc230) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc230) + } -> tensor<1x64x12x20xbf16> loc(#loc230) + %378 = xten_nn.subgraph (%arg5 = %377: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_350", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_350", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x128x12x20xbf16> loc(#loc231) + xten_nn.output %464 : tensor<1x128x12x20xbf16> loc(#loc231) + } -> tensor<1x128x12x20xbf16> loc(#loc231) + %379 = xten_nn.subgraph (%arg5 = %378: tensor<1x128x12x20xbf16>, %arg6 = %31: tensor<128x128x3x3xbf16>, %arg7 = %30: tensor<128xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 14]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 5, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 10960 : ui64, ofm = 6676 : ui64, wts = 19616 : ui64}, + computation_cycles = 30551 : ui64, + cycle_count = 30551 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<128x128x3x3xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 5 : ui8, + config.ifm_sv_width.size = 34 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 8 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 14008 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x128x12x20xbf16> loc(#loc232) + xten_nn.output %465 : tensor<1x128x12x20xbf16> loc(#loc232) + } -> tensor<1x128x12x20xbf16> loc(#loc232) + xten_nn.output %464 : tensor<1x128x12x20xbf16> loc(#loc232) + } -> tensor<1x128x12x20xbf16> loc(#loc232) + %380 = xten_nn.subgraph (%arg5 = %379: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 32, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 3850 : ui64, ofm = 1930 : ui64}, + computation_cycles = 732 : ui64, + cycle_count = 3850 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 32 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 381 : ui64, + startup_cycles = 31 : ui64 + }} { + %465 = tensor.empty() : tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %465 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %464 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + %381 = xten_nn.subgraph (%arg5 = %380: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_353_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc234) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc234) + } -> tensor<1x64x12x20xbf16> loc(#loc234) + %382 = xten_nn.subgraph (%arg5 = %27: tensor<1x64x12x20xbf16>, %arg6 = %381: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [1 : index], + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1930 : ui64, ifm2 = 1930 : ui64, ofm = 970 : ui64}, + computation_cycles = 414 : ui64, + cycle_count = 1930 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 23 : ui64 + }} { + %465 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %465 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + %383 = xten_nn.subgraph (%arg5 = %382: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1930 : ui64, ifm2 = 1930 : ui64, ofm = 970 : ui64}, + computation_cycles = 389 : ui64, + cycle_count = 1930 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %465 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + %384 = xten_nn.subgraph (%arg5 = %380: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_353_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc234) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc234) + } -> tensor<1x64x12x20xbf16> loc(#loc234) + %385 = xten_nn.subgraph (%arg5 = %384: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1930 : ui64, ifm2 = 1930 : ui64, ofm = 970 : ui64}, + computation_cycles = 389 : ui64, + cycle_count = 1930 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc235) + xten_nn.output %465 : tensor<1x64x12x20xbf16> loc(#loc235) + } -> tensor<1x64x12x20xbf16> loc(#loc235) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc235) + } -> tensor<1x64x12x20xbf16> loc(#loc235) + %386 = xten_nn.subgraph (%arg5 = %377: tensor<1x64x12x20xbf16>, %arg6 = %385: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_355", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_355", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x128x12x20xbf16> loc(#loc236) + xten_nn.output %464 : tensor<1x128x12x20xbf16> loc(#loc236) + } -> tensor<1x128x12x20xbf16> loc(#loc236) + %387 = xten_nn.subgraph (%arg5 = %386: tensor<1x128x12x20xbf16>, %arg6 = %29: tensor<64x128x3x3xbf16>, %arg7 = %28: tensor<64xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 14]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 5, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 10960 : ui64, ofm = 3338 : ui64, wts = 9808 : ui64}, + computation_cycles = 15447 : ui64, + cycle_count = 15447 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<64x128x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 5 : ui8, + config.ifm_sv_width.size = 34 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 8 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 14008 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %465 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + %388 = xten_nn.subgraph (%arg5 = %387: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 0, 12]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1930 : ui64, ofm = 970 : ui64}, + computation_cycles = 2690 : ui64, + cycle_count = 2690 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 2346 : ui64, + startup_cycles = 24 : ui64 + }} { + %465 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %465 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + %389 = xten_nn.subgraph (%arg5 = %381: tensor<1x64x12x20xbf16>, %arg6 = %388: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1930 : ui64, ifm2 = 1930 : ui64, ofm = 970 : ui64}, + computation_cycles = 389 : ui64, + cycle_count = 1930 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc239) + xten_nn.output %465 : tensor<1x64x12x20xbf16> loc(#loc239) + } -> tensor<1x64x12x20xbf16> loc(#loc239) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc239) + } -> tensor<1x64x12x20xbf16> loc(#loc239) + %390 = xten_nn.subgraph (%arg5 = %383: tensor<1x64x12x20xbf16>, %arg6 = %389: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 20]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 1930 : ui64, ifm2 = 1930 : ui64, ofm = 970 : ui64}, + computation_cycles = 413 : ui64, + cycle_count = 1930 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 3 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 22 : ui64 + }} { + %465 = tensor.empty() : tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %465 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %464 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + %391 = xten_nn.subgraph (%arg5 = %376: tensor<1x64x12x20xbf16>, %arg6 = %390: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_363", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_363", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x128x12x20xbf16> loc(#loc242) + xten_nn.output %464 : tensor<1x128x12x20xbf16> loc(#loc242) + } -> tensor<1x128x12x20xbf16> loc(#loc242) + %392 = xten_nn.subgraph (%arg5 = %391: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Resize_365", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Resize_365", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 128 : ui32, + config.dim_2 = 12 : ui32, + config.dim_3 = 20 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.output_H = 24 : ui32, + config.output_W = 40 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x128x24x40xbf16> loc(#loc243) + xten_nn.output %464 : tensor<1x128x24x40xbf16> loc(#loc243) + } -> tensor<1x128x24x40xbf16> loc(#loc243) + %393 = xten_nn.subgraph (%arg5 = %392: tensor<1x128x24x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Slice_371", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + OutputName = "Slice_371", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 128 : ui32, + config.dim_h = 24 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 23 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x128x23x40xbf16> loc(#loc244) + xten_nn.output %464 : tensor<1x128x23x40xbf16> loc(#loc244) + } -> tensor<1x128x23x40xbf16> loc(#loc244) + %394 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 2, 80]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 60, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 3, 1, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 246720 : ui64, ofm = 31680 : ui64}, + computation_cycles = 26024 : ui64, + cycle_count = 740160 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 2 : ui8, + config.ifm_sv_width.size = 80 : ui8, + config.ksize = 2 : ui8, + config.manual_ba_pad_val = 0 : si8, + config.stride_log2 = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8, + data_io.ofm.dimensions.width.size = 40 : ui8 + }, + estimation = { + compute_cycles = 131 : ui64, + startup_cycles = 113 : ui64 + }} { + %465 = tensor.empty() : tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %465 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %464 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + %395 = xten_nn.subgraph (%arg5 = %394: tensor<1x3x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 12, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 90, 160]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 3, 6, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61640 : ui64, ofm = 7880 : ui64}, + computation_cycles = 6036 : ui64, + cycle_count = 61640 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x90x160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 12 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize = 2 : ui8, + config.manual_ba_pad_val = 0 : si8, + config.stride_log2 = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 150 : ui64, + startup_cycles = 113 : ui64 + }} { + %465 = tensor.empty() : tensor<1x3x45x80xbf16> loc(#loc245) + xten_nn.output %465 : tensor<1x3x45x80xbf16> loc(#loc245) + } -> tensor<1x3x45x80xbf16> loc(#loc245) + xten_nn.output %464 : tensor<1x3x45x80xbf16> loc(#loc245) + } -> tensor<1x3x45x80xbf16> loc(#loc245) + %396 = xten_nn.subgraph (%arg5 = %395: tensor<1x3x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 12, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 3, 6, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15410 : ui64, ofm = 1970 : ui64}, + computation_cycles = 1731 : ui64, + cycle_count = 15410 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 0]> : vector<4xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x45x80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 1], [0, 0]], + HWPaddingNotCounted = [[0, 1], [0, 0]], + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifm_sv_depth.size = 8 : ui8, + config.ifm_sv_height.size = 12 : ui8, + config.ifm_sv_width.size = 16 : ui8, + config.ksize = 2 : ui8, + config.manual_ba_pad_val = 0 : si8, + config.stride_log2 = 1 : ui8, + data_io.ofm.dimensions.width.padding = 0 : ui8, + data_io.ofm.dimensions.width.size = 8 : ui8 + }, + estimation = { + compute_cycles = 150 : ui64, + startup_cycles = 113 : ui64 + }} { + %465 = tensor.empty() : tensor<1x3x23x40xbf16> loc(#loc246) + xten_nn.output %465 : tensor<1x3x23x40xbf16> loc(#loc246) + } -> tensor<1x3x23x40xbf16> loc(#loc246) + xten_nn.output %464 : tensor<1x3x23x40xbf16> loc(#loc246) + } -> tensor<1x3x23x40xbf16> loc(#loc246) + %397 = xten_nn.subgraph (%arg5 = %393: tensor<1x128x23x40xbf16>, %arg6 = %219: tensor<1x40x23x40xbf16>, %arg7 = %396: tensor<1x3x23x40xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index, 2 : index], + LayerName = "Concat_372", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_372", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x171x23x40xbf16> loc(#loc247) + xten_nn.output %464 : tensor<1x171x23x40xbf16> loc(#loc247) + } -> tensor<1x171x23x40xbf16> loc(#loc247) + %398 = xten_nn.subgraph (%arg5 = %397: tensor<1x171x23x40xbf16>, %arg6 = %26: tensor<80x171x3x3xbf16>, %arg7 = %25: tensor<80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 10]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 5, 40]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 176, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[48, 5, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[48, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 8]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 48, 1, 8]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 44220 : ui64, ofm = 20008 : ui64, wts = 26972 : ui64}, + computation_cycles = 110123 : ui64, + cycle_count = 110123 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x171x23x40xbf16>, %arg9 = %arg6: tensor<80x171x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 5 : ui8, + config.ifm_sv_width.size = 50 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 11 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 48 : ui8 + }, + estimation = { + compute_cycles = 25938 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x80x23x40xbf16> loc(#loc249) + xten_nn.output %465 : tensor<1x80x23x40xbf16> loc(#loc249) + } -> tensor<1x80x23x40xbf16> loc(#loc349) + xten_nn.output %464 : tensor<1x80x23x40xbf16> loc(#loc349) + } -> tensor<1x80x23x40xbf16> loc(#loc349) + %399 = xten_nn.subgraph (%arg5 = %398: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_375_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %400 = xten_nn.subgraph (%arg5 = %398: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_375_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %401 = xten_nn.subgraph (%arg5 = %400: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_376", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_376", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x80x23x40xbf16> loc(#loc251) + xten_nn.output %464 : tensor<1x80x23x40xbf16> loc(#loc251) + } -> tensor<1x80x23x40xbf16> loc(#loc251) + %402 = xten_nn.subgraph (%arg5 = %401: tensor<1x80x23x40xbf16>, %arg6 = %24: tensor<80x80x3x3xbf16>, %arg7 = %23: tensor<80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 10]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 5, 40]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[48, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[48, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 8]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 48, 1, 8]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 20100 : ui64, ofm = 20008 : ui64, wts = 12260 : ui64}, + computation_cycles = 50243 : ui64, + cycle_count = 50243 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<80x80x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 5 : ui8, + config.ifm_sv_width.size = 50 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 5 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 48 : ui8 + }, + estimation = { + compute_cycles = 11790 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %465 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %464 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + %403 = xten_nn.subgraph (%arg5 = %402: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 24, 2, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 48, 1, 8]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 24, 2, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 11550 : ui64, ofm = 5790 : ui64}, + computation_cycles = 1768 : ui64, + cycle_count = 11550 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 24 : ui32, + config.ifmsv_height = 2 : ui32, + config.ifmsv_width = 40 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 381 : ui64, + startup_cycles = 31 : ui64 + }} { + %465 = tensor.empty() : tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %465 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %464 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + %404 = xten_nn.subgraph (%arg5 = %403: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_379_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %405 = xten_nn.subgraph (%arg5 = %20: tensor<1x40x23x40xbf16>, %arg6 = %404: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [1 : index], + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7720 : ui64, ifm2 = 7720 : ui64, ofm = 3880 : ui64}, + computation_cycles = 1038 : ui64, + cycle_count = 7720 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 23 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %465 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + %406 = xten_nn.subgraph (%arg5 = %405: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7720 : ui64, ifm2 = 7720 : ui64, ofm = 3880 : ui64}, + computation_cycles = 932 : ui64, + cycle_count = 7720 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %465 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + %407 = xten_nn.subgraph (%arg5 = %403: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_379_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %408 = xten_nn.subgraph (%arg5 = %407: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7720 : ui64, ifm2 = 7720 : ui64, ofm = 3880 : ui64}, + computation_cycles = 932 : ui64, + cycle_count = 7720 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %465 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + %409 = xten_nn.subgraph (%arg5 = %400: tensor<1x40x23x40xbf16>, %arg6 = %408: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_381", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_381", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x80x23x40xbf16> loc(#loc256) + xten_nn.output %464 : tensor<1x80x23x40xbf16> loc(#loc256) + } -> tensor<1x80x23x40xbf16> loc(#loc256) + %410 = xten_nn.subgraph (%arg5 = %409: tensor<1x80x23x40xbf16>, %arg6 = %22: tensor<40x80x3x3xbf16>, %arg7 = %21: tensor<40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 10]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 5, 40]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[24, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[24, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 0, 0, 8]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 3, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 8]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 20100 : ui64, ofm = 10004 : ui64, wts = 6130 : ui64}, + computation_cycles = 25293 : ui64, + cycle_count = 25293 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<40x80x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 5 : ui8, + config.ifm_sv_width.size = 50 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 5 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 48 : ui8 + }, + estimation = { + compute_cycles = 11790 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc257) + xten_nn.output %465 : tensor<1x40x23x40xbf16> loc(#loc257) + } -> tensor<1x40x23x40xbf16> loc(#loc257) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc257) + } -> tensor<1x40x23x40xbf16> loc(#loc257) + %411 = xten_nn.subgraph (%arg5 = %410: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 8]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7700 : ui64, ofm = 3860 : ui64}, + computation_cycles = 9373 : ui64, + cycle_count = 9373 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 4446 : ui64, + startup_cycles = 24 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %465 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + %412 = xten_nn.subgraph (%arg5 = %404: tensor<1x40x23x40xbf16>, %arg6 = %411: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7720 : ui64, ifm2 = 7720 : ui64, ofm = 3880 : ui64}, + computation_cycles = 932 : ui64, + cycle_count = 7720 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %465 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + %413 = xten_nn.subgraph (%arg5 = %406: tensor<1x40x23x40xbf16>, %arg6 = %412: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 6, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 7720 : ui64, ifm2 = 7720 : ui64, ofm = 3880 : ui64}, + computation_cycles = 1037 : ui64, + cycle_count = 7720 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 6 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 22 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %465 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %464 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + %414 = xten_nn.subgraph (%arg5 = %399: tensor<1x40x23x40xbf16>, %arg6 = %413: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_389", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_389", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x80x23x40xbf16> loc(#loc262) + xten_nn.output %464 : tensor<1x80x23x40xbf16> loc(#loc262) + } -> tensor<1x80x23x40xbf16> loc(#loc262) + %415 = xten_nn.subgraph (%arg5 = %414: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Resize_391", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Resize_391", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 80 : ui32, + config.dim_2 = 23 : ui32, + config.dim_3 = 40 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.output_H = 46 : ui32, + config.output_W = 80 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x80x46x80xbf16> loc(#loc263) + xten_nn.output %464 : tensor<1x80x46x80xbf16> loc(#loc263) + } -> tensor<1x80x46x80xbf16> loc(#loc263) + %416 = xten_nn.subgraph (%arg5 = %415: tensor<1x80x46x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Slice_397", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + OutputName = "Slice_397", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 80 : ui32, + config.dim_h = 46 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 45 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x80x45x80xbf16> loc(#loc264) + xten_nn.output %464 : tensor<1x80x45x80xbf16> loc(#loc264) + } -> tensor<1x80x45x80xbf16> loc(#loc264) + %417 = xten_nn.subgraph (%arg5 = %416: tensor<1x80x45x80xbf16>, %arg6 = %183: tensor<1x24x45x80xbf16>, %arg7 = %395: tensor<1x3x45x80xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index, 2 : index], + LayerName = "Concat_398", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x107x45x80xbf16> loc(#loc265) + xten_nn.output %464 : tensor<1x107x45x80xbf16> loc(#loc265) + } -> tensor<1x107x45x80xbf16> loc(#loc265) + %418 = xten_nn.subgraph (%arg5 = %417: tensor<1x107x45x80xbf16>, %arg6 = %19: tensor<40x107x3x3xbf16>, %arg7 = %18: tensor<40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 14, 18]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 112, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[24, 5, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[24, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 12, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 70910 : ui64, ofm = 31410 : ui64, wts = 8582 : ui64}, + computation_cycles = 107863 : ui64, + cycle_count = 107863 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x107x45x80xbf16>, %arg9 = %arg6: tensor<40x107x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 14 : ui8, + config.ifm_sv_width.size = 18 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 7 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 20545 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x45x80xbf16> loc(#loc267) + xten_nn.output %465 : tensor<1x40x45x80xbf16> loc(#loc267) + } -> tensor<1x40x45x80xbf16> loc(#loc350) + xten_nn.output %464 : tensor<1x40x45x80xbf16> loc(#loc350) + } -> tensor<1x40x45x80xbf16> loc(#loc350) + %419 = xten_nn.subgraph (%arg5 = %418: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_401_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %420 = xten_nn.subgraph (%arg5 = %418: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_401_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %421 = xten_nn.subgraph (%arg5 = %420: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_402", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_402", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x40x45x80xbf16> loc(#loc269) + xten_nn.output %464 : tensor<1x40x45x80xbf16> loc(#loc269) + } -> tensor<1x40x45x80xbf16> loc(#loc269) + %422 = xten_nn.subgraph (%arg5 = %421: tensor<1x40x45x80xbf16>, %arg6 = %17: tensor<40x40x3x3xbf16>, %arg7 = %16: tensor<40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 14, 18]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[24, 8, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[24, 8, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 12, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 30390 : ui64, ofm = 31410 : ui64, wts = 3678 : ui64}, + computation_cycles = 46423 : ui64, + cycle_count = 46423 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<40x40x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 0 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 14 : ui8, + config.ifm_sv_width.size = 18 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 3 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 16 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 16 : ui8 + }, + estimation = { + compute_cycles = 8805 : ui64, + startup_cycles = 160 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %465 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %464 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + %423 = xten_nn.subgraph (%arg5 = %422: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 16, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 30800 : ui64, ofm = 15440 : ui64}, + computation_cycles = 4358 : ui64, + cycle_count = 30800 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 16 : ui32, + config.ifmsv_height = 12 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 381 : ui64, + startup_cycles = 31 : ui64 + }} { + %465 = tensor.empty() : tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %465 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %464 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + %424 = xten_nn.subgraph (%arg5 = %423: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_405_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %425 = xten_nn.subgraph (%arg5 = %13: tensor<1x20x45x80xbf16>, %arg6 = %424: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [1 : index], + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15440 : ui64, ifm2 = 15440 : ui64, ofm = 7760 : ui64}, + computation_cycles = 1870 : ui64, + cycle_count = 15440 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 12 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 23 : ui64 + }} { + %465 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %465 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + %426 = xten_nn.subgraph (%arg5 = %425: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15440 : ui64, ifm2 = 15440 : ui64, ofm = 7760 : ui64}, + computation_cycles = 1656 : ui64, + cycle_count = 15440 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 12 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %465 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + %427 = xten_nn.subgraph (%arg5 = %423: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_405_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %428 = xten_nn.subgraph (%arg5 = %427: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15440 : ui64, ifm2 = 15440 : ui64, ofm = 7760 : ui64}, + computation_cycles = 1656 : ui64, + cycle_count = 15440 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 12 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %465 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + %429 = xten_nn.subgraph (%arg5 = %420: tensor<1x20x45x80xbf16>, %arg6 = %428: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_407", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_407", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x40x45x80xbf16> loc(#loc274) + xten_nn.output %464 : tensor<1x40x45x80xbf16> loc(#loc274) + } -> tensor<1x40x45x80xbf16> loc(#loc274) + %430 = xten_nn.subgraph (%arg5 = %429: tensor<1x40x45x80xbf16>, %arg6 = %15: tensor<20x40x3x3xbf16>, %arg7 = %14: tensor<20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 14, 34]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 8, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[12, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[12, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 12, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 16]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 28710 : ui64, ofm = 10014 : ui64, wts = 1650 : ui64}, + computation_cycles = 36245 : ui64, + cycle_count = 36245 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<20x40x3x3xbf16>, %arg10 = %arg7: tensor<20xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 8 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 14 : ui8, + config.ifm_sv_width.size = 34 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 5 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 8 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 11275 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc275) + xten_nn.output %465 : tensor<1x20x45x80xbf16> loc(#loc275) + } -> tensor<1x20x45x80xbf16> loc(#loc275) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc275) + } -> tensor<1x20x45x80xbf16> loc(#loc275) + %431 = xten_nn.subgraph (%arg5 = %430: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 16]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 12, 20]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15400 : ui64, ofm = 7720 : ui64}, + computation_cycles = 18539 : ui64, + cycle_count = 18539 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 12 : ui32, + config.ifmsv_width = 20 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 4446 : ui64, + startup_cycles = 24 : ui64 + }} { + %465 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %465 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + %432 = xten_nn.subgraph (%arg5 = %424: tensor<1x20x45x80xbf16>, %arg6 = %431: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15440 : ui64, ifm2 = 15440 : ui64, ofm = 7760 : ui64}, + computation_cycles = 1656 : ui64, + cycle_count = 15440 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 12 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 44 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %465 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + %433 = xten_nn.subgraph (%arg5 = %426: tensor<1x20x45x80xbf16>, %arg6 = %432: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 12, 10]> : vector<4xindex>, + l2_extend_end = dense<[0, 8, 3, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 15440 : ui64, ifm2 = 15440 : ui64, ofm = 7760 : ui64}, + computation_cycles = 1869 : ui64, + cycle_count = 15440 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 12 : ui32, + config.ifmsv_width = 10 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 71 : ui64, + startup_cycles = 22 : ui64 + }} { + %465 = tensor.empty() : tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %465 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %464 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + %434 = xten_nn.subgraph (%arg5 = %419: tensor<1x20x45x80xbf16>, %arg6 = %433: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_415", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_415", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x40x45x80xbf16> loc(#loc280) + xten_nn.output %464 : tensor<1x40x45x80xbf16> loc(#loc280) + } -> tensor<1x40x45x80xbf16> loc(#loc280) + %435 = xten_nn.subgraph (%arg5 = %434: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Resize_417", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Resize_417", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 40 : ui32, + config.dim_2 = 45 : ui32, + config.dim_3 = 80 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.output_H = 90 : ui32, + config.output_W = 160 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x40x90x160xbf16> loc(#loc281) + xten_nn.output %464 : tensor<1x40x90x160xbf16> loc(#loc281) + } -> tensor<1x40x90x160xbf16> loc(#loc281) + %436 = xten_nn.subgraph (%arg5 = %435: tensor<1x40x90x160xbf16>, %arg6 = %175: tensor<1x16x90x160xbf16>, %arg7 = %394: tensor<1x3x90x160xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index, 2 : index], + LayerName = "Concat_418", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_418", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x59x90x160xbf16> loc(#loc282) + xten_nn.output %464 : tensor<1x59x90x160xbf16> loc(#loc282) + } -> tensor<1x59x90x160xbf16> loc(#loc282) + %437 = xten_nn.subgraph (%arg5 = %436: tensor<1x59x90x160xbf16>, %arg6 = %12: tensor<32x59x3x3xbf16>, %arg7 = %11: tensor<32xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 7, 34]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 64, 20, 160]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 5, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 32, 18, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 191400 : ui64, ofm = 38650 : ui64, wts = 2472 : ui64}, + computation_cycles = 190465 : ui64, + cycle_count = 957000 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x59x90x160xbf16>, %arg9 = %arg6: tensor<32x59x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 7 : ui8, + config.ifm_sv_width.size = 34 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 4 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 8 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 7056 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x32x90x160xbf16> loc(#loc284) + xten_nn.output %465 : tensor<1x32x90x160xbf16> loc(#loc284) + } -> tensor<1x32x90x160xbf16> loc(#loc351) + xten_nn.output %464 : tensor<1x32x90x160xbf16> loc(#loc351) + } -> tensor<1x32x90x160xbf16> loc(#loc351) + %438 = xten_nn.subgraph (%arg5 = %437: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_421_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %439 = xten_nn.subgraph (%arg5 = %437: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_421_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %440 = xten_nn.subgraph (%arg5 = %439: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_422", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_422", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x32x90x160xbf16> loc(#loc286) + xten_nn.output %464 : tensor<1x32x90x160xbf16> loc(#loc286) + } -> tensor<1x32x90x160xbf16> loc(#loc286) + %441 = xten_nn.subgraph (%arg5 = %440: tensor<1x32x90x160xbf16>, %arg6 = %10: tensor<32x32x3x3xbf16>, %arg7 = %9: tensor<32xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 34]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 32, 32, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 4, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 32, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 98520 : ui64, ofm = 38700 : ui64, wts = 1236 : ui64}, + computation_cycles = 99125 : ui64, + cycle_count = 297375 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<32x32x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 6 : ui8, + config.ifm_sv_width.size = 34 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 8 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 3018 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %465 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %464 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + %442 = xten_nn.subgraph (%arg5 = %441: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 32]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 32, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 8, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 0, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 32, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61590 : ui64, ofm = 30870 : ui64}, + computation_cycles = 8314 : ui64, + cycle_count = 184770 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 8 : ui32, + config.ifmsv_width = 32 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 403 : ui64, + startup_cycles = 31 : ui64 + }} { + %465 = tensor.empty() : tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %465 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %464 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + %443 = xten_nn.subgraph (%arg5 = %442: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_425_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %444 = xten_nn.subgraph (%arg5 = %6: tensor<1x16x90x160xbf16>, %arg6 = %443: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [1 : index], + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61740 : ui64, ifm2 = 61740 : ui64, ofm = 31020 : ui64}, + computation_cycles = 6566 : ui64, + cycle_count = 185220 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 8 : ui32, + config.ifmsv_width = 16 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 75 : ui64, + startup_cycles = 23 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + %445 = xten_nn.subgraph (%arg5 = %444: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61740 : ui64, ifm2 = 61740 : ui64, ofm = 31020 : ui64}, + computation_cycles = 5698 : ui64, + cycle_count = 185220 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 8 : ui32, + config.ifmsv_width = 16 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 46 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + %446 = xten_nn.subgraph (%arg5 = %442: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_425_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %447 = xten_nn.subgraph (%arg5 = %446: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61740 : ui64, ifm2 = 61740 : ui64, ofm = 31020 : ui64}, + computation_cycles = 5698 : ui64, + cycle_count = 185220 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 8 : ui32, + config.ifmsv_width = 16 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 46 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + %448 = xten_nn.subgraph (%arg5 = %439: tensor<1x16x90x160xbf16>, %arg6 = %447: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_427", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_427", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x32x90x160xbf16> loc(#loc291) + xten_nn.output %464 : tensor<1x32x90x160xbf16> loc(#loc291) + } -> tensor<1x32x90x160xbf16> loc(#loc291) + %449 = xten_nn.subgraph (%arg5 = %448: tensor<1x32x90x160xbf16>, %arg6 = %8: tensor<16x32x3x3xbf16>, %arg7 = %7: tensor<16xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 6, 34]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 32, 32, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 4, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 98520 : ui64, ofm = 38700 : ui64, wts = 1236 : ui64}, + computation_cycles = 99125 : ui64, + cycle_count = 297375 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<16x32x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 6 : ui8, + config.ifm_sv_width.size = 34 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 2 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 8 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 3018 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc292) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc292) + } -> tensor<1x16x90x160xbf16> loc(#loc292) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc292) + } -> tensor<1x16x90x160xbf16> loc(#loc292) + %450 = xten_nn.subgraph (%arg5 = %449: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 23, 8]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 23, 8]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 59080 : ui64, ofm = 29640 : ui64}, + computation_cycles = 72267 : ui64, + cycle_count = 72267 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 23 : ui32, + config.ifmsv_width = 8 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 3466 : ui64, + startup_cycles = 24 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + %451 = xten_nn.subgraph (%arg5 = %443: tensor<1x16x90x160xbf16>, %arg6 = %450: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61740 : ui64, ifm2 = 61740 : ui64, ofm = 31020 : ui64}, + computation_cycles = 5698 : ui64, + cycle_count = 185220 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 8 : ui32, + config.ifmsv_width = 16 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 46 : ui64, + startup_cycles = 25 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + %452 = xten_nn.subgraph (%arg5 = %445: tensor<1x16x90x160xbf16>, %arg6 = %451: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 8, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 2, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 30, 160]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 61740 : ui64, ifm2 = 61740 : ui64, ofm = 31020 : ui64}, + computation_cycles = 6565 : ui64, + cycle_count = 185220 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 8 : ui32, + config.ifmsv_width = 16 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 75 : ui64, + startup_cycles = 22 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %465 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %464 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + %453 = xten_nn.subgraph (%arg5 = %438: tensor<1x16x90x160xbf16>, %arg6 = %452: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_435", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_435", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x32x90x160xbf16> loc(#loc297) + xten_nn.output %464 : tensor<1x32x90x160xbf16> loc(#loc297) + } -> tensor<1x32x90x160xbf16> loc(#loc297) + %454 = xten_nn.subgraph (%arg5 = %453: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Resize_437", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Resize_437", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 32 : ui32, + config.dim_2 = 90 : ui32, + config.dim_3 = 160 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.output_H = 180 : ui32, + config.output_W = 320 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x32x180x320xbf16> loc(#loc298) + xten_nn.output %464 : tensor<1x32x180x320xbf16> loc(#loc298) + } -> tensor<1x32x180x320xbf16> loc(#loc298) + %455 = xten_nn.subgraph (%arg5 = %454: tensor<1x32x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_438", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Concat_438", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN", + herd_size = dense<4> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x35x180x320xbf16> loc(#loc299) + xten_nn.output %464 : tensor<1x35x180x320xbf16> loc(#loc299) + } -> tensor<1x35x180x320xbf16> loc(#loc299) + %456 = xten_nn.subgraph (%arg5 = %455: tensor<1x35x180x320xbf16>, %arg6 = %5: tensor<16x35x3x3xbf16>, %arg7 = %4: tensor<16xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 8, 6, 66]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 40, 17, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 8, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[16, 5, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[16, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 4, 64]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 1, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 15, 320]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 478200 : ui64, ofm = 154200 : ui64, wts = 1650 : ui64}, + computation_cycles = 534365 : ui64, + cycle_count = 6412380 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x35x180x320xbf16>, %arg9 = %arg6: tensor<16x35x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 8 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 6 : ui8, + config.ifm_sv_width.size = 66 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 5 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 8 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 64 : ui8 + }, + estimation = { + compute_cycles = 8215 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x180x320xbf16> loc(#loc301) + xten_nn.output %465 : tensor<1x16x180x320xbf16> loc(#loc301) + } -> tensor<1x16x180x320xbf16> loc(#loc352) + xten_nn.output %464 : tensor<1x16x180x320xbf16> loc(#loc352) + } -> tensor<1x16x180x320xbf16> loc(#loc352) + %457 = xten_nn.subgraph (%arg5 = %456: tensor<1x16x180x320xbf16>, %arg6 = %3: tensor<16x16x3x3xbf16>, %arg7 = %2: tensor<16xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 7, 34]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 22, 320]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[8, 16, 3, 3]> : vector<4xindex>, + l2_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex>, + l3_extend_end = dense<[16, 0, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<0> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 8, 5, 32]> : vector<4xindex>, + l2_extend_end = dense<[0, 16, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 172260 : ui64, ofm = 139140 : ui64, wts = 618 : ui64}, + computation_cycles = 171455 : ui64, + cycle_count = 1550340 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<16x16x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 16 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 7 : ui8, + config.ifm_sv_width.size = 34 : ui8, + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 8 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 32 : ui8 + }, + estimation = { + compute_cycles = 1764 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x16x180x320xbf16> loc(#loc303) + xten_nn.output %465 : tensor<1x16x180x320xbf16> loc(#loc303) + } -> tensor<1x16x180x320xbf16> loc(#loc353) + xten_nn.output %464 : tensor<1x16x180x320xbf16> loc(#loc353) + } -> tensor<1x16x180x320xbf16> loc(#loc353) + %458 = xten_nn.subgraph (%arg5 = %457: tensor<1x16x180x320xbf16>, %arg6 = %1: tensor<4x16x1x1xbf16>, %arg7 = %0: tensor<4xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 48, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[1, 16, 1, 64]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 16, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l1_extend_end = dense<[4, 48, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<-1> : vector<2xindex>, + l1_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex>, + l2_extend_end = dense<[28, 48, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex>, + l3_extend_end = dense<[28, 48, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 4, 1, 64]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 463050 : ui64, ofm = 232650 : ui64, wts = 298 : ui64}, + computation_cycles = 264965 : ui64, + cycle_count = 4167450 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3, 4, 0]> : vector<5xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<4x16x1x1xbf16>, %arg10 = %arg7: tensor<4xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = 64 : ui8, + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.force_och_gran_8 = 1 : ui8, + config.ifm_sv_depth.size = 64 : ui16, + config.ifm_sv_height.padding = 0 : ui8, + config.ifm_sv_height.size = 1 : ui8, + config.ifm_sv_width.size = 64 : ui8, + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.num_ifm_depth_iters = 1 : ui8, + config.ofm_offset_packed.left = 0 : ui4, + config.ofm_offset_packed.top = 0 : ui4, + config.ofm_sv_depth.size = 8 : ui16, + config.ofm_width_pad = 0 : ui8, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8, + data_io.ofm.dimensions.width.size = 64 : ui8 + }, + estimation = { + compute_cycles = 1039 : ui64, + startup_cycles = 182 : ui64 + }} { + %465 = tensor.empty() : tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %465 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %464 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + %459 = xten_nn.subgraph (%arg5 = %458: tensor<1x4x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_444_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 4 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 3 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x1x180x320xbf16> loc(#loc305) + xten_nn.output %464 : tensor<1x1x180x320xbf16> loc(#loc305) + } -> tensor<1x1x180x320xbf16> loc(#loc305) + %460 = xten_nn.subgraph (%arg5 = %459: tensor<1x1x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 1, 5, 40]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 1, 5, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 231120 : ui64, ofm = 115920 : ui64}, + computation_cycles = 18655 : ui64, + cycle_count = 2080080 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 5 : ui32, + config.ifmsv_width = 40 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 119 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %465 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %464 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + %461 = xten_nn.subgraph (%arg5 = %458: tensor<1x4x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_444_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }, + herd_size = dense<1> : vector<2xindex>} { + %464 = tensor.empty() : tensor<1x3x180x320xbf16> loc(#loc305) + xten_nn.output %464 : tensor<1x3x180x320xbf16> loc(#loc305) + } -> tensor<1x3x180x320xbf16> loc(#loc305) + %462 = xten_nn.subgraph (%arg5 = %461: tensor<1x3x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 5, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 5, 16]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 3, 5, 16]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 232200 : ui64, ifm2 = 232200 : ui64, ofm = 117000 : ui64}, + computation_cycles = 34045 : ui64, + cycle_count = 2089800 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 5 : ui32, + config.ifmsv_width = 16 : ui32, + config.num_kernel_iters = 0 : ui16 + }, + estimation = { + compute_cycles = 51 : ui64, + startup_cycles = 22 : ui64 + }} { + %465 = tensor.empty() : tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %465 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %464 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + %463 = xten_nn.subgraph (%arg5 = %462: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_fullbroadcast = dense<[1, 0]> : vector<2xindex>, + l1_tile_count = dense<[1, 3, 5, 40]> : vector<4xindex>, + l2_extend_end = dense<0> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l1_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l1_extend_start = dense<0> : vector<4xindex>, + l1_tile_count = dense<[1, 3, 5, 40]> : vector<4xindex>, + l2_extend_end = dense<[0, 24, 0, 0]> : vector<4xindex>, + l2_extend_start = dense<0> : vector<4xindex>, + l2_tile_count = dense<[1, 8, 20, 320]> : vector<4xindex>, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + estimation = { + buffer_cycles = {ifm = 231120 : ui64, ofm = 115920 : ui64}, + computation_cycles = 18655 : ui64, + cycle_count = 2080080 : ui64 + }, + herd_size = dense<4> : vector<2xindex>, + l1_tile_permute = dense<[2, 1, 3]> : vector<3xindex>, + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %464 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.ifmsv_depth = 8 : ui32, + config.ifmsv_height = 5 : ui32, + config.ifmsv_width = 40 : ui32, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }, + estimation = { + compute_cycles = 119 : ui64, + startup_cycles = 40 : ui64 + }} { + %465 = tensor.empty() : tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %465 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %464 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + return %452, %433, %413, %390, %463, %460 : tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16> loc(#loc308) + } loc(#loc308) +} loc(#loc) +#loc1 = loc("Div_2") +#loc2 = loc("Sub_431") +#loc3 = loc("Sub_411") +#loc4 = loc("Sub_385") +#loc5 = loc("Sub_359") +#loc6 = loc("Div_16") +#loc7 = loc("Sub_14") +#loc8 = loc("Initializer_398") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("Conv_17") +#loc14 = loc("Add_19") +#loc15 = loc("Clip_22") +#loc16 = loc("Div_24") +#loc17 = loc("Mul_25") +#loc18 = loc("Conv_26") +#loc19 = loc("Relu_27") +#loc20 = loc("Conv_28") +#loc21 = loc("Add_29") +#loc22 = loc("Conv_30") +#loc23 = loc("Relu_31") +#loc24 = loc("Conv_32") +#loc25 = loc("Relu_33") +#loc26 = loc("Conv_34") +#loc27 = loc("Conv_35") +#loc28 = loc("Relu_36") +#loc29 = loc("Conv_37") +#loc30 = loc("Relu_38") +#loc31 = loc("Conv_39") +#loc32 = loc("Add_40") +#loc33 = loc("Conv_41") +#loc34 = loc("Relu_42") +#loc35 = loc("Conv_43") +#loc36 = loc("Relu_44") +#loc37 = loc("GlobalAveragePool_45") +#loc38 = loc("Conv_46") +#loc39 = loc("Relu_47") +#loc40 = loc("Conv_48") +#loc41 = loc("Add_50") +#loc42 = loc("Clip_53") +#loc43 = loc("Div_55") +#loc44 = loc("Mul_56") +#loc45 = loc("Conv_57") +#loc46 = loc("Conv_58") +#loc47 = loc("Relu_59") +#loc48 = loc("Conv_60") +#loc49 = loc("Relu_61") +#loc50 = loc("GlobalAveragePool_62") +#loc51 = loc("Conv_63") +#loc52 = loc("Relu_64") +#loc53 = loc("Conv_65") +#loc54 = loc("Add_67") +#loc55 = loc("Clip_70") +#loc56 = loc("Div_72") +#loc57 = loc("Mul_73") +#loc58 = loc("Conv_74") +#loc59 = loc("Add_75") +#loc60 = loc("Conv_76") +#loc61 = loc("Relu_77") +#loc62 = loc("Conv_78") +#loc63 = loc("Relu_79") +#loc64 = loc("GlobalAveragePool_80") +#loc65 = loc("Conv_81") +#loc66 = loc("Relu_82") +#loc67 = loc("Conv_83") +#loc68 = loc("Add_85") +#loc69 = loc("Clip_88") +#loc70 = loc("Div_90") +#loc71 = loc("Mul_91") +#loc72 = loc("Conv_92") +#loc73 = loc("Add_93") +#loc74 = loc("Conv_94") +#loc75 = loc("Add_96") +#loc76 = loc("Clip_99") +#loc77 = loc("Div_101") +#loc78 = loc("Mul_102") +#loc79 = loc("Conv_103") +#loc80 = loc("Add_105") +#loc81 = loc("Clip_108") +#loc82 = loc("Div_110") +#loc83 = loc("Mul_111") +#loc84 = loc("Conv_112") +#loc85 = loc("Conv_113") +#loc86 = loc("Add_115") +#loc87 = loc("Clip_118") +#loc88 = loc("Div_120") +#loc89 = loc("Mul_121") +#loc90 = loc("Conv_122") +#loc91 = loc("Add_124") +#loc92 = loc("Clip_127") +#loc93 = loc("Div_129") +#loc94 = loc("Mul_130") +#loc95 = loc("Conv_131") +#loc96 = loc("Add_132") +#loc97 = loc("Conv_133") +#loc98 = loc("Add_135") +#loc99 = loc("Clip_138") +#loc100 = loc("Div_140") +#loc101 = loc("Mul_141") +#loc102 = loc("Conv_142") +#loc103 = loc("Add_144") +#loc104 = loc("Clip_147") +#loc105 = loc("Div_149") +#loc106 = loc("Mul_150") +#loc107 = loc("Conv_151") +#loc108 = loc("Add_152") +#loc109 = loc("Conv_153") +#loc110 = loc("Add_155") +#loc111 = loc("Clip_158") +#loc112 = loc("Div_160") +#loc113 = loc("Mul_161") +#loc114 = loc("Conv_162") +#loc115 = loc("Add_164") +#loc116 = loc("Clip_167") +#loc117 = loc("Div_169") +#loc118 = loc("Mul_170") +#loc119 = loc("Conv_171") +#loc120 = loc("Add_172") +#loc121 = loc("Conv_173") +#loc122 = loc("Add_175") +#loc123 = loc("Clip_178") +#loc124 = loc("Div_180") +#loc125 = loc("Mul_181") +#loc126 = loc("Conv_182") +#loc127 = loc("Add_184") +#loc128 = loc("Clip_187") +#loc129 = loc("Div_189") +#loc130 = loc("Mul_190") +#loc131 = loc("GlobalAveragePool_191") +#loc132 = loc("Conv_192") +#loc133 = loc("Relu_193") +#loc134 = loc("Conv_194") +#loc135 = loc("Add_196") +#loc136 = loc("Clip_199") +#loc137 = loc("Div_201") +#loc138 = loc("Mul_202") +#loc139 = loc("Conv_203") +#loc140 = loc("Conv_204") +#loc141 = loc("Add_206") +#loc142 = loc("Clip_209") +#loc143 = loc("Div_211") +#loc144 = loc("Mul_212") +#loc145 = loc("Conv_213") +#loc146 = loc("Add_215") +#loc147 = loc("Clip_218") +#loc148 = loc("Div_220") +#loc149 = loc("Mul_221") +#loc150 = loc("GlobalAveragePool_222") +#loc151 = loc("Conv_223") +#loc152 = loc("Relu_224") +#loc153 = loc("Conv_225") +#loc154 = loc("Add_227") +#loc155 = loc("Clip_230") +#loc156 = loc("Div_232") +#loc157 = loc("Mul_233") +#loc158 = loc("Conv_234") +#loc159 = loc("Add_235") +#loc160 = loc("Conv_236") +#loc161 = loc("Add_238") +#loc162 = loc("Clip_241") +#loc163 = loc("Div_243") +#loc164 = loc("Mul_244") +#loc165 = loc("Conv_245") +#loc166 = loc("Add_247") +#loc167 = loc("Clip_250") +#loc168 = loc("Div_252") +#loc169 = loc("Mul_253") +#loc170 = loc("GlobalAveragePool_254") +#loc171 = loc("Conv_255") +#loc172 = loc("Relu_256") +#loc173 = loc("Conv_257") +#loc174 = loc("Add_259") +#loc175 = loc("Clip_262") +#loc176 = loc("Div_264") +#loc177 = loc("Mul_265") +#loc178 = loc("Conv_266") +#loc179 = loc("Conv_267") +#loc180 = loc("Add_269") +#loc181 = loc("Clip_272") +#loc182 = loc("Div_274") +#loc183 = loc("Mul_275") +#loc184 = loc("Conv_276") +#loc185 = loc("Add_278") +#loc186 = loc("Clip_281") +#loc187 = loc("Div_283") +#loc188 = loc("Mul_284") +#loc189 = loc("GlobalAveragePool_285") +#loc190 = loc("Conv_286") +#loc191 = loc("Relu_287") +#loc192 = loc("Conv_288") +#loc193 = loc("Add_290") +#loc194 = loc("Clip_293") +#loc195 = loc("Div_295") +#loc196 = loc("Mul_296") +#loc197 = loc("Conv_297") +#loc198 = loc("Add_298") +#loc199 = loc("Conv_299") +#loc200 = loc("Add_301") +#loc201 = loc("Clip_304") +#loc202 = loc("Div_306") +#loc203 = loc("Mul_307") +#loc204 = loc("Conv_308") +#loc205 = loc("Add_310") +#loc206 = loc("Clip_313") +#loc207 = loc("Div_315") +#loc208 = loc("Mul_316") +#loc209 = loc("GlobalAveragePool_317") +#loc210 = loc("Conv_318") +#loc211 = loc("Relu_319") +#loc212 = loc("Conv_320") +#loc213 = loc("Add_322") +#loc214 = loc("Clip_325") +#loc215 = loc("Div_327") +#loc216 = loc("Mul_328") +#loc217 = loc("Conv_329") +#loc218 = loc("Add_330") +#loc219 = loc("Conv_331") +#loc220 = loc("Add_333") +#loc221 = loc("Clip_336") +#loc222 = loc("Div_338") +#loc223 = loc("Mul_339") +#loc224 = loc("GlobalAveragePool_342") +#loc225 = loc("Conv_343") +#loc226 = loc("Sigmoid_344") +#loc227 = loc("Mul_345") +#loc228 = loc("Conv_340") +#loc229 = loc("Relu_341") +#loc230 = loc("Split_349") +#loc231 = loc("Concat_350") +#loc232 = loc("Conv_351") +#loc233 = loc("Sigmoid_352") +#loc234 = loc("Split_353") +#loc235 = loc("Mul_354") +#loc236 = loc("Concat_355") +#loc237 = loc("Conv_356") +#loc238 = loc("Tanh_357") +#loc239 = loc("Mul_361") +#loc240 = loc("Mul_360") +#loc241 = loc("Add_362") +#loc242 = loc("Concat_363") +#loc243 = loc("Resize_365") +#loc244 = loc("Slice_371") +#loc245 = loc("AveragePool_347") +#loc246 = loc("AveragePool_348") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_380") +#loc256 = loc("Concat_381") +#loc257 = loc("Conv_382") +#loc258 = loc("Tanh_383") +#loc259 = loc("Mul_387") +#loc260 = loc("Mul_386") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_406") +#loc274 = loc("Concat_407") +#loc275 = loc("Conv_408") +#loc276 = loc("Tanh_409") +#loc277 = loc("Mul_413") +#loc278 = loc("Mul_412") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_426") +#loc291 = loc("Concat_427") +#loc292 = loc("Conv_428") +#loc293 = loc("Tanh_429") +#loc294 = loc("Mul_433") +#loc295 = loc("Mul_432") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_446") +#loc307 = loc("Clip_447") +#loc308 = loc(fused[#loc1, #loc2, #loc3, #loc4, #loc5, #loc6, #loc7, #loc8, #loc9, #loc10, #loc11, #loc12, #loc13, #loc14, #loc15, #loc16, #loc17, #loc18, #loc19, #loc20, #loc21, #loc22, #loc23, #loc24, #loc25, #loc26, #loc27, #loc28, #loc29, #loc30, #loc31, #loc32, #loc33, #loc34, #loc35, #loc36, #loc37, #loc38, #loc39, #loc40, #loc41, #loc42, #loc43, #loc44, #loc45, #loc46, #loc47, #loc48, #loc49, #loc50, #loc51, #loc52, #loc53, #loc54, #loc55, #loc56, #loc57, #loc58, #loc59, #loc60, #loc61, #loc62, #loc63, #loc64, #loc65, #loc66, #loc67, #loc68, #loc69, #loc70, #loc71, #loc72, #loc73, #loc74, #loc75, #loc76, #loc77, #loc78, #loc79, #loc80, #loc81, #loc82, #loc83, #loc84, #loc85, #loc86, #loc87, #loc88, #loc89, #loc90, #loc91, #loc92, #loc93, #loc94, #loc95, #loc96, #loc97, #loc98, #loc99, #loc100, #loc101, #loc102, #loc103, #loc104, #loc105, #loc106, #loc107, #loc108, #loc109, #loc110, #loc111, #loc112, #loc113, #loc114, #loc115, #loc116, #loc117, #loc118, #loc119, #loc120, #loc121, #loc122, #loc123, #loc124, #loc125, #loc126, #loc127, #loc128, #loc129, #loc130, #loc131, #loc132, #loc133, #loc134, #loc135, #loc136, #loc137, #loc138, #loc139, #loc140, #loc141, #loc142, #loc143, #loc144, #loc145, #loc146, #loc147, #loc148, #loc149, #loc150, #loc151, #loc152, #loc153, #loc154, #loc155, #loc156, #loc157, #loc158, #loc159, #loc160, #loc161, #loc162, #loc163, #loc164, #loc165, #loc166, #loc167, #loc168, #loc169, #loc170, #loc171, #loc172, #loc173, #loc174, #loc175, #loc176, #loc177, #loc178, #loc179, #loc180, #loc181, #loc182, #loc183, #loc184, #loc185, #loc186, #loc187, #loc188, #loc189, #loc190, #loc191, #loc192, #loc193, #loc194, #loc195, #loc196, #loc197, #loc198, #loc199, #loc200, #loc201, #loc202, #loc203, #loc204, #loc205, #loc206, #loc207, #loc208, #loc209, #loc210, #loc211, #loc212, #loc213, #loc214, #loc215, #loc216, #loc217, #loc218, #loc219, #loc220, #loc221, #loc222, #loc223, #loc224, #loc225, #loc226, #loc227, #loc228, #loc229, #loc230, #loc231, #loc232, #loc233, #loc234, #loc235, #loc236, #loc237, #loc238, #loc239, #loc240, #loc241, #loc242, #loc243, #loc244, #loc245, #loc246, #loc247, #loc248, #loc249, #loc250, #loc251, #loc252, #loc253, #loc254, #loc255, #loc256, #loc257, #loc258, #loc259, #loc260, #loc261, #loc262, #loc263, #loc264, #loc265, #loc266, #loc267, #loc268, #loc269, #loc270, #loc271, #loc272, #loc273, #loc274, #loc275, #loc276, #loc277, #loc278, #loc279, #loc280, #loc281, #loc282, #loc283, #loc284, #loc285, #loc286, #loc287, #loc288, #loc289, #loc290, #loc291, #loc292, #loc293, #loc294, #loc295, #loc296, #loc297, #loc298, #loc299, #loc300, #loc301, #loc302, #loc303, #loc304, #loc305, #loc306, #loc307]) +#loc309 = loc(fused[#loc7, #loc8]) +#loc310 = loc(fused[#loc11, #loc9, #loc12]) +#loc311 = loc(fused[#loc9, #loc12, #loc11]) +#loc312 = loc(fused[#loc18, #loc19]) +#loc313 = loc(fused[#loc20, #loc21]) +#loc314 = loc(fused[#loc22, #loc23]) +#loc315 = loc(fused[#loc24, #loc25]) +#loc316 = loc(fused[#loc27, #loc28]) +#loc317 = loc(fused[#loc29, #loc30]) +#loc318 = loc(fused[#loc31, #loc32]) +#loc319 = loc(fused[#loc33, #loc34]) +#loc320 = loc(fused[#loc35, #loc36]) +#loc321 = loc(fused[#loc38, #loc39]) +#loc322 = loc(fused[#loc46, #loc47]) +#loc323 = loc(fused[#loc48, #loc49]) +#loc324 = loc(fused[#loc51, #loc52]) +#loc325 = loc(fused[#loc58, #loc59]) +#loc326 = loc(fused[#loc60, #loc61]) +#loc327 = loc(fused[#loc62, #loc63]) +#loc328 = loc(fused[#loc65, #loc66]) +#loc329 = loc(fused[#loc72, #loc73]) +#loc330 = loc(fused[#loc95, #loc96]) +#loc331 = loc(fused[#loc107, #loc108]) +#loc332 = loc(fused[#loc119, #loc120]) +#loc333 = loc(fused[#loc130, #loc131]) +#loc334 = loc(fused[#loc132, #loc133]) +#loc335 = loc(fused[#loc149, #loc150]) +#loc336 = loc(fused[#loc151, #loc152]) +#loc337 = loc(fused[#loc158, #loc159]) +#loc338 = loc(fused[#loc169, #loc170]) +#loc339 = loc(fused[#loc171, #loc172]) +#loc340 = loc(fused[#loc188, #loc189]) +#loc341 = loc(fused[#loc190, #loc191]) +#loc342 = loc(fused[#loc197, #loc198]) +#loc343 = loc(fused[#loc208, #loc209]) +#loc344 = loc(fused[#loc210, #loc211]) +#loc345 = loc(fused[#loc217, #loc218]) +#loc346 = loc(fused[#loc223, #loc224]) +#loc347 = loc(fused[#loc228, #loc229, #loc227]) +#loc348 = loc(fused[#loc228, #loc229]) +#loc349 = loc(fused[#loc248, #loc249]) +#loc350 = loc(fused[#loc266, #loc267]) +#loc351 = loc(fused[#loc283, #loc284]) +#loc352 = loc(fused[#loc300, #loc301]) +#loc353 = loc(fused[#loc302, #loc303]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/par.subgraph.pre-dse.mlir b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/par.subgraph.pre-dse.mlir new file mode 100644 index 0000000000000000000000000000000000000000..777941f7a624f9e607e84e992d143386659969a3 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/par.subgraph.pre-dse.mlir @@ -0,0 +1,25253 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func private @forward(%arg0: tensor<1x180x320x4xbf16> loc(unknown), %arg1: tensor<1x16x90x160xbf16> loc(unknown), %arg2: tensor<1x20x45x80xbf16> loc(unknown), %arg3: tensor<1x40x23x40xbf16> loc(unknown), %arg4: tensor<1x64x12x20xbf16> loc(unknown)) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) attributes { + max_heap_size = 2240 : ui32, + max_stack_size = 2368 : ui32, + stack_heap_start_address = 45696 : ui32, + total_stack_heap_region_size = 6912 : ui32} { + %0 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/biases"} -> tensor<4xbf16> loc(#loc) + %1 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/weights"} -> tensor<4x16x1x1xbf16> loc(#loc) + %2 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/biases"} -> tensor<16xbf16> loc(#loc) + %3 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/weights"} -> tensor<16x16x3x3xbf16> loc(#loc) + %4 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/biases"} -> tensor<16xbf16> loc(#loc) + %5 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/weights"} -> tensor<16x35x3x3xbf16> loc(#loc) + %6 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_431/Constant_0_0"} -> tensor<1x16x90x160xbf16> loc(#loc2) + %7 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/biases"} -> tensor<16xbf16> loc(#loc) + %8 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/weights"} -> tensor<16x32x3x3xbf16> loc(#loc) + %9 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/biases"} -> tensor<32xbf16> loc(#loc) + %10 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/weights"} -> tensor<32x32x3x3xbf16> loc(#loc) + %11 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/biases"} -> tensor<32xbf16> loc(#loc) + %12 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/weights"} -> tensor<32x59x3x3xbf16> loc(#loc) + %13 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_411/Constant_0_0"} -> tensor<1x20x45x80xbf16> loc(#loc3) + %14 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/biases"} -> tensor<20xbf16> loc(#loc) + %15 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/weights"} -> tensor<20x40x3x3xbf16> loc(#loc) + %16 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/biases"} -> tensor<40xbf16> loc(#loc) + %17 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/weights"} -> tensor<40x40x3x3xbf16> loc(#loc) + %18 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/biases"} -> tensor<40xbf16> loc(#loc) + %19 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/weights"} -> tensor<40x107x3x3xbf16> loc(#loc) + %20 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_385/Constant_0_0"} -> tensor<1x40x23x40xbf16> loc(#loc4) + %21 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/biases"} -> tensor<40xbf16> loc(#loc) + %22 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/weights"} -> tensor<40x80x3x3xbf16> loc(#loc) + %23 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/biases"} -> tensor<80xbf16> loc(#loc) + %24 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/weights"} -> tensor<80x80x3x3xbf16> loc(#loc) + %25 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/biases"} -> tensor<80xbf16> loc(#loc) + %26 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/weights"} -> tensor<80x171x3x3xbf16> loc(#loc) + %27 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_359/Constant_0_0"} -> tensor<1x64x12x20xbf16> loc(#loc5) + %28 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/biases"} -> tensor<64xbf16> loc(#loc) + %29 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/weights"} -> tensor<64x128x3x3xbf16> loc(#loc) + %30 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/biases"} -> tensor<128xbf16> loc(#loc) + %31 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/weights"} -> tensor<128x128x3x3xbf16> loc(#loc) + %32 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/biases"} -> tensor<128xbf16> loc(#loc) + %33 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %34 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/biases"} -> tensor<128xbf16> loc(#loc) + %35 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %36 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/biases"} -> tensor<960xbf16> loc(#loc) + %37 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %38 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/biases"} -> tensor<160xbf16> loc(#loc) + %39 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %40 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/biases"} -> tensor<960xbf16> loc(#loc) + %41 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %42 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/biases"} -> tensor<240xbf16> loc(#loc) + %43 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %44 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/biases"} -> tensor<960xbf16> loc(#loc) + %45 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %46 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/biases"} -> tensor<960xbf16> loc(#loc) + %47 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %48 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/biases"} -> tensor<160xbf16> loc(#loc) + %49 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %50 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/biases"} -> tensor<960xbf16> loc(#loc) + %51 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %52 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/biases"} -> tensor<240xbf16> loc(#loc) + %53 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %54 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/biases"} -> tensor<960xbf16> loc(#loc) + %55 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %56 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/biases"} -> tensor<960xbf16> loc(#loc) + %57 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %58 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/biases"} -> tensor<160xbf16> loc(#loc) + %59 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/weights"} -> tensor<160x672x1x1xbf16> loc(#loc) + %60 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/biases"} -> tensor<672xbf16> loc(#loc) + %61 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %62 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/biases"} -> tensor<168xbf16> loc(#loc) + %63 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %64 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/biases"} -> tensor<672xbf16> loc(#loc) + %65 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/weights"} -> tensor<672x1x9x9xbf16> loc(#loc) + %66 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/biases"} -> tensor<672xbf16> loc(#loc) + %67 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %68 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/biases"} -> tensor<112xbf16> loc(#loc) + %69 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/weights"} -> tensor<112x672x1x1xbf16> loc(#loc) + %70 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/biases"} -> tensor<672xbf16> loc(#loc) + %71 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %72 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/biases"} -> tensor<168xbf16> loc(#loc) + %73 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %74 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/biases"} -> tensor<672xbf16> loc(#loc) + %75 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/weights"} -> tensor<672x1x3x3xbf16> loc(#loc) + %76 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/biases"} -> tensor<672xbf16> loc(#loc) + %77 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %78 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/biases"} -> tensor<112xbf16> loc(#loc) + %79 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/weights"} -> tensor<112x480x1x1xbf16> loc(#loc) + %80 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/biases"} -> tensor<480xbf16> loc(#loc) + %81 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/weights"} -> tensor<480x120x1x1xbf16> loc(#loc) + %82 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/biases"} -> tensor<120xbf16> loc(#loc) + %83 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/weights"} -> tensor<120x480x1x1xbf16> loc(#loc) + %84 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/biases"} -> tensor<480xbf16> loc(#loc) + %85 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/weights"} -> tensor<480x1x3x3xbf16> loc(#loc) + %86 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/biases"} -> tensor<480xbf16> loc(#loc) + %87 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/weights"} -> tensor<480x80x1x1xbf16> loc(#loc) + %88 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/biases"} -> tensor<80xbf16> loc(#loc) + %89 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %90 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/biases"} -> tensor<184xbf16> loc(#loc) + %91 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %92 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/biases"} -> tensor<184xbf16> loc(#loc) + %93 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %94 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/biases"} -> tensor<80xbf16> loc(#loc) + %95 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %96 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/biases"} -> tensor<184xbf16> loc(#loc) + %97 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %98 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/biases"} -> tensor<184xbf16> loc(#loc) + %99 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %100 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/biases"} -> tensor<80xbf16> loc(#loc) + %101 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/weights"} -> tensor<80x200x1x1xbf16> loc(#loc) + %102 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/biases"} -> tensor<200xbf16> loc(#loc) + %103 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/weights"} -> tensor<200x1x3x3xbf16> loc(#loc) + %104 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/biases"} -> tensor<200xbf16> loc(#loc) + %105 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/weights"} -> tensor<200x80x1x1xbf16> loc(#loc) + %106 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/biases"} -> tensor<80xbf16> loc(#loc) + %107 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/weights"} -> tensor<80x240x1x1xbf16> loc(#loc) + %108 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/biases"} -> tensor<240xbf16> loc(#loc) + %109 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/weights"} -> tensor<240x1x3x3xbf16> loc(#loc) + %110 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/biases"} -> tensor<240xbf16> loc(#loc) + %111 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/weights"} -> tensor<240x40x1x1xbf16> loc(#loc) + %112 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/biases"} -> tensor<40xbf16> loc(#loc) + %113 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %114 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/biases"} -> tensor<120xbf16> loc(#loc) + %115 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %116 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/biases"} -> tensor<32xbf16> loc(#loc) + %117 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %118 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/biases"} -> tensor<120xbf16> loc(#loc) + %119 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %120 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/biases"} -> tensor<120xbf16> loc(#loc) + %121 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %122 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/biases"} -> tensor<40xbf16> loc(#loc) + %123 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %124 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/biases"} -> tensor<120xbf16> loc(#loc) + %125 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %126 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/biases"} -> tensor<32xbf16> loc(#loc) + %127 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %128 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/biases"} -> tensor<120xbf16> loc(#loc) + %129 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %130 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/biases"} -> tensor<120xbf16> loc(#loc) + %131 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %132 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/biases"} -> tensor<40xbf16> loc(#loc) + %133 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/weights"} -> tensor<40x72x1x1xbf16> loc(#loc) + %134 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/biases"} -> tensor<72xbf16> loc(#loc) + %135 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %136 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/biases"} -> tensor<24xbf16> loc(#loc) + %137 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %138 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/biases"} -> tensor<72xbf16> loc(#loc) + %139 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/weights"} -> tensor<72x1x5x5xbf16> loc(#loc) + %140 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/biases"} -> tensor<72xbf16> loc(#loc) + %141 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %142 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/biases"} -> tensor<24xbf16> loc(#loc) + %143 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %144 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/biases"} -> tensor<72xbf16> loc(#loc) + %145 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/weights"} -> tensor<72x1x3x3xbf16> loc(#loc) + %146 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/biases"} -> tensor<72xbf16> loc(#loc) + %147 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %148 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/biases"} -> tensor<24xbf16> loc(#loc) + %149 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/weights"} -> tensor<24x64x1x1xbf16> loc(#loc) + %150 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/biases"} -> tensor<64xbf16> loc(#loc) + %151 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/weights"} -> tensor<64x1x3x3xbf16> loc(#loc) + %152 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/biases"} -> tensor<64xbf16> loc(#loc) + %153 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/weights"} -> tensor<64x16x1x1xbf16> loc(#loc) + %154 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/biases"} -> tensor<16xbf16> loc(#loc) + %155 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/weights"} -> tensor<16x16x1x1xbf16> loc(#loc) + %156 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/biases"} -> tensor<16xbf16> loc(#loc) + %157 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/weights"} -> tensor<16x1x3x3xbf16> loc(#loc) + %158 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/biases"} -> tensor<16xbf16> loc(#loc) + %159 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/weights"} -> tensor<16x3x3x3xbf16> loc(#loc) + %160 = xten_nn.load_external_const {file = "constants.h5", key = "Div_16/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc6) + %161 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_14/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc309) + %162 = xten_nn.subgraph (%arg5 = %arg0: tensor<1x180x320x4xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.906250e-03 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.906250e-03> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc1) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_2", + OutputName = "Div_2", + shift = 0 : i8} : (tensor<1x180x320x4xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %463 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %461 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + %163 = xten_nn.subgraph (%arg5 = %162: tensor<1x180x320x4xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Slice_7", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Slice_7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "W", + config.dim_c = 184 : ui32, + config.dim_h = 320 : ui32, + config.dim_w = 4 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_7", + OutputName = "Slice_7", + size = array, + start = array} : (tensor<1x180x320x4xbf16>) -> tensor<1x180x320x3xbf16> loc(#loc9) + xten_nn.output %461 : tensor<1x180x320x3xbf16> loc(#loc9) + } -> tensor<1x180x320x3xbf16> loc(#loc9) + %164 = xten_nn.subgraph (%arg5 = %163: tensor<1x180x320x3xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Generated-#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "BufferPadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_0_padded = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_1_padded = 23 : ui32, + config.dim_2 = 3 : ui32, + config.dim_2_padded = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_padded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x180x320x3xbf16> loc(#loc10) + } -> tensor<1x180x320x3xbf16> loc(#loc10) + %165 = xten_nn.subgraph (%arg5 = %164: tensor<1x180x320x3xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#2", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Generated-#3", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_2 = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 10 : ui32 + }} { + %461 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc11) + %462 = tosa.transpose %arg5, %461 : (tensor<1x180x320x3xbf16>, tensor<4xi32>) -> tensor<1x3x180x320xbf16> loc(#loc311) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc311) + } -> tensor<1x3x180x320xbf16> loc(#loc310) + %166 = xten_nn.subgraph (%arg5 = %165: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#4", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Generated-#5", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 184 : ui32, + config.dim_0_unpadded = 180 : ui32, + config.dim_1 = 1 : ui32, + config.dim_1_unpadded = 1 : ui32, + config.dim_2 = 320 : ui32, + config.dim_2_unpadded = 320 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x3x180x320xbf16> loc(#loc10) + } -> tensor<1x3x180x320xbf16> loc(#loc10) + %167 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>, %arg6 = %161: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Initializer_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Initializer_398", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Sub_14", OutputName = "Initializer_398"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc309) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc309) + } -> tensor<1x3x180x320xbf16> loc(#loc309) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc309) + } -> tensor<1x3x180x320xbf16> loc(#loc309) + %168 = xten_nn.subgraph (%arg5 = %167: tensor<1x3x180x320xbf16>, %arg6 = %160: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + OutputName = "Div_16", + PartOfLayerName = "Div_16", + shift = 0 : i8} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + %169 = xten_nn.subgraph (%arg5 = %168: tensor<1x3x180x320xbf16>, %arg6 = %159: tensor<16x3x3x3xbf16>, %arg7 = %158: tensor<16xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x3x180x320xbf16>, %arg9 = %arg6: tensor<16x3x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 2 : ui8, + config.stride_w = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x3x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x3xbf16> loc(#loc13) + %465 = tosa.transpose %arg8, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc13) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_17", + PartOfOutputName = "Conv_17", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>, tensor<16x3x3x3xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc13) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc13) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc13) + } -> tensor<1x16x90x160xbf16> loc(#loc13) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc13) + } -> tensor<1x16x90x160xbf16> loc(#loc13) + %170 = xten_nn.subgraph (%arg5 = %169: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_19", OutputName = "Add_19"} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc14) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc14) + } -> tensor<1x16x90x160xbf16> loc(#loc14) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc14) + } -> tensor<1x16x90x160xbf16> loc(#loc14) + %171 = xten_nn.subgraph (%arg5 = %170: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_22", + OutputName = "Clip_22", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + %172 = xten_nn.subgraph (%arg5 = %171: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_24", + OutputName = "Div_24", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + %173 = xten_nn.subgraph (%arg5 = %169: tensor<1x16x90x160xbf16>, %arg6 = %172: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_25", + OutputName = "Mul_25", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + %174 = xten_nn.subgraph (%arg5 = %173: tensor<1x16x90x160xbf16>, %arg6 = %157: tensor<16x1x3x3xbf16>, %arg7 = %156: tensor<16xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<16x1x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc312) + %465 = tosa.transpose %arg9, %464 : (tensor<16x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x16x1xbf16> loc(#loc312) + %466 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc312) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_26", + PartOfOutputName = "Conv_26", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<3x3x16x1xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc18) + %468 = tosa.clamp %467 { + LayerName = "Relu_27", + OutputName = "Relu_27", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc19) + %469 = tosa.transpose %468, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc312) + xten_nn.output %469 : tensor<1x16x90x160xbf16> loc(#loc19) + } -> tensor<1x16x90x160xbf16> loc(#loc312) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc312) + } -> tensor<1x16x90x160xbf16> loc(#loc312) + %175 = xten_nn.subgraph (%arg5 = %174: tensor<1x16x90x160xbf16>, %arg6 = %155: tensor<16x16x1x1xbf16>, %arg7 = %154: tensor<16xbf16>, %arg8 = %173: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_28", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x16x90x160xbf16>, %arg10 = %arg6: tensor<16x16x1x1xbf16>, %arg11 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_28", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_28", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc20) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<16x16x1x1xbf16>) -> tensor<16x1x1x16xbf16> loc(#loc20) + %466 = tosa.transpose %arg9, %464 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_28", + PartOfOutputName = "Conv_28", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<16x1x1x16xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %468 = tosa.transpose %467, %463 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc20) + xten_nn.output %468 : tensor<1x16x90x160xbf16> loc(#loc20) + } -> tensor<1x16x90x160xbf16> loc(#loc20) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x16x90x160xbf16>, %arg10 = %arg8: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_29", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_29", OutputName = "Add_29"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc313) + %176 = xten_nn.subgraph (%arg5 = %175: tensor<1x16x90x160xbf16>, %arg6 = %153: tensor<64x16x1x1xbf16>, %arg7 = %152: tensor<64xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<64x16x1x1xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc314) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<64x16x1x1xbf16>) -> tensor<64x1x1x16xbf16> loc(#loc314) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc314) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_30", + PartOfOutputName = "Conv_30", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<64x1x1x16xbf16>, tensor<64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc22) + %467 = tosa.clamp %466 { + LayerName = "Relu_31", + OutputName = "Relu_31", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc23) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x64xbf16>, tensor<4xi32>) -> tensor<1x64x90x160xbf16> loc(#loc314) + xten_nn.output %468 : tensor<1x64x90x160xbf16> loc(#loc23) + } -> tensor<1x64x90x160xbf16> loc(#loc314) + xten_nn.output %461 : tensor<1x64x90x160xbf16> loc(#loc314) + } -> tensor<1x64x90x160xbf16> loc(#loc314) + %177 = xten_nn.subgraph (%arg5 = %176: tensor<1x64x90x160xbf16>, %arg6 = %151: tensor<64x1x3x3xbf16>, %arg7 = %150: tensor<64xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x90x160xbf16>, %arg9 = %arg6: tensor<64x1x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc315) + %465 = tosa.transpose %arg9, %464 : (tensor<64x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x64x1xbf16> loc(#loc315) + %466 = tosa.transpose %arg8, %463 : (tensor<1x64x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x64xbf16> loc(#loc315) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_32", + PartOfOutputName = "Conv_32", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x64xbf16>, tensor<3x3x64x1xbf16>, tensor<64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc24) + %468 = tosa.clamp %467 { + LayerName = "Relu_33", + OutputName = "Relu_33", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc25) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x64xbf16>, tensor<4xi32>) -> tensor<1x64x45x80xbf16> loc(#loc315) + xten_nn.output %469 : tensor<1x64x45x80xbf16> loc(#loc25) + } -> tensor<1x64x45x80xbf16> loc(#loc315) + xten_nn.output %461 : tensor<1x64x45x80xbf16> loc(#loc315) + } -> tensor<1x64x45x80xbf16> loc(#loc315) + %178 = xten_nn.subgraph (%arg5 = %177: tensor<1x64x45x80xbf16>, %arg6 = %149: tensor<24x64x1x1xbf16>, %arg7 = %148: tensor<24xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x45x80xbf16>, %arg9 = %arg6: tensor<24x64x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc26) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x64x1x1xbf16>) -> tensor<24x1x1x64xbf16> loc(#loc26) + %465 = tosa.transpose %arg8, %463 : (tensor<1x64x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x64xbf16> loc(#loc26) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_34", + PartOfOutputName = "Conv_34", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x64xbf16>, tensor<24x1x1x64xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc26) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc26) + xten_nn.output %467 : tensor<1x24x45x80xbf16> loc(#loc26) + } -> tensor<1x24x45x80xbf16> loc(#loc26) + xten_nn.output %461 : tensor<1x24x45x80xbf16> loc(#loc26) + } -> tensor<1x24x45x80xbf16> loc(#loc26) + %179 = xten_nn.subgraph (%arg5 = %178: tensor<1x24x45x80xbf16>, %arg6 = %147: tensor<72x24x1x1xbf16>, %arg7 = %146: tensor<72xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc316) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc316) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc316) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_35", + PartOfOutputName = "Conv_35", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc27) + %467 = tosa.clamp %466 { + LayerName = "Relu_36", + OutputName = "Relu_36", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc28) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc316) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc28) + } -> tensor<1x72x45x80xbf16> loc(#loc316) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc316) + } -> tensor<1x72x45x80xbf16> loc(#loc316) + %180 = xten_nn.subgraph (%arg5 = %179: tensor<1x72x45x80xbf16>, %arg6 = %145: tensor<72x1x3x3xbf16>, %arg7 = %144: tensor<72xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x3x3xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc317) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x72x1xbf16> loc(#loc317) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc317) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_37", + PartOfOutputName = "Conv_37", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<3x3x72x1xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc29) + %468 = tosa.clamp %467 { + LayerName = "Relu_38", + OutputName = "Relu_38", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc30) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc317) + xten_nn.output %469 : tensor<1x72x45x80xbf16> loc(#loc30) + } -> tensor<1x72x45x80xbf16> loc(#loc317) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc317) + } -> tensor<1x72x45x80xbf16> loc(#loc317) + %181 = xten_nn.subgraph (%arg5 = %180: tensor<1x72x45x80xbf16>, %arg6 = %143: tensor<24x72x1x1xbf16>, %arg7 = %142: tensor<24xbf16>, %arg8 = %178: tensor<1x24x45x80xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_39", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x72x45x80xbf16>, %arg10 = %arg6: tensor<24x72x1x1xbf16>, %arg11 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_39", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_39", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc31) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc31) + %466 = tosa.transpose %arg9, %464 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc31) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_39", + PartOfOutputName = "Conv_39", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc31) + %468 = tosa.transpose %467, %463 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc31) + xten_nn.output %468 : tensor<1x24x45x80xbf16> loc(#loc31) + } -> tensor<1x24x45x80xbf16> loc(#loc31) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x24x45x80xbf16>, %arg10 = %arg8: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Add_40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_40", OutputName = "Add_40"} : (tensor<1x24x45x80xbf16>, tensor<1x24x45x80xbf16>) -> tensor<1x24x45x80xbf16> loc(#loc32) + xten_nn.output %463 : tensor<1x24x45x80xbf16> loc(#loc32) + } -> tensor<1x24x45x80xbf16> loc(#loc32) + xten_nn.output %462 : tensor<1x24x45x80xbf16> loc(#loc32) + } -> tensor<1x24x45x80xbf16> loc(#loc318) + %182 = xten_nn.subgraph (%arg5 = %181: tensor<1x24x45x80xbf16>, %arg6 = %141: tensor<72x24x1x1xbf16>, %arg7 = %140: tensor<72xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc319) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc319) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc319) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_41", + PartOfOutputName = "Conv_41", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc33) + %467 = tosa.clamp %466 { + LayerName = "Relu_42", + OutputName = "Relu_42", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc34) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc319) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc34) + } -> tensor<1x72x45x80xbf16> loc(#loc319) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc319) + } -> tensor<1x72x45x80xbf16> loc(#loc319) + %183 = xten_nn.subgraph (%arg5 = %182: tensor<1x72x45x80xbf16>, %arg6 = %139: tensor<72x1x5x5xbf16>, %arg7 = %138: tensor<72xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x5x5xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 1]], + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc320) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x72x1xbf16> loc(#loc320) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc320) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_43", + PartOfOutputName = "Conv_43", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<5x5x72x1xbf16>, tensor<72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc35) + %468 = tosa.clamp %467 { + LayerName = "Relu_44", + OutputName = "Relu_44", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc36) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x72xbf16>, tensor<4xi32>) -> tensor<1x72x23x40xbf16> loc(#loc320) + xten_nn.output %469 : tensor<1x72x23x40xbf16> loc(#loc36) + } -> tensor<1x72x23x40xbf16> loc(#loc320) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc320) + } -> tensor<1x72x23x40xbf16> loc(#loc320) + %184 = xten_nn.subgraph (%arg5 = %183: tensor<1x72x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#6", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 9 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x72x23x40xbf16>) -> tensor<1x72x1x920xbf16> loc(#loc37) + xten_nn.output %461 : tensor<1x72x1x920xbf16> loc(#loc37) + } -> tensor<1x72x1x920xbf16> loc(#loc37) + %185 = xten_nn.subgraph (%arg5 = %184: tensor<1x72x1x920xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#8", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#9", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x920xbf16>) attributes { + LayerName = "Generated-#8", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#9", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 72 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x72x1x920xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc37) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc37) + } -> tensor<1x72x1x1xbf16> loc(#loc37) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc37) + } -> tensor<1x72x1x1xbf16> loc(#loc37) + %186 = xten_nn.subgraph (%arg5 = %185: tensor<1x72x1x1xbf16>, %arg6 = %137: tensor<24x72x1x1xbf16>, %arg7 = %136: tensor<24xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x1x1xbf16>, %arg9 = %arg6: tensor<24x72x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc321) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc321) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_46", + PartOfOutputName = "Conv_46", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc38) + %465 = tosa.clamp %464 { + LayerName = "Relu_47", + OutputName = "Relu_47", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc39) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x24xbf16>) -> tensor<1x24x1x1xbf16> loc(#loc321) + xten_nn.output %466 : tensor<1x24x1x1xbf16> loc(#loc39) + } -> tensor<1x24x1x1xbf16> loc(#loc321) + xten_nn.output %461 : tensor<1x24x1x1xbf16> loc(#loc321) + } -> tensor<1x24x1x1xbf16> loc(#loc321) + %187 = xten_nn.subgraph (%arg5 = %186: tensor<1x24x1x1xbf16>, %arg6 = %135: tensor<72x24x1x1xbf16>, %arg7 = %134: tensor<72xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x1x1xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc40) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x24x1x1xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc40) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_48", + PartOfOutputName = "Conv_48", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc40) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x72xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc40) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc40) + } -> tensor<1x72x1x1xbf16> loc(#loc40) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc40) + } -> tensor<1x72x1x1xbf16> loc(#loc40) + %188 = xten_nn.subgraph (%arg5 = %187: tensor<1x72x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_50", OutputName = "Add_50"} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc41) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc41) + } -> tensor<1x72x1x1xbf16> loc(#loc41) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc41) + } -> tensor<1x72x1x1xbf16> loc(#loc41) + %189 = xten_nn.subgraph (%arg5 = %188: tensor<1x72x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_53", + OutputName = "Clip_53", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + %190 = xten_nn.subgraph (%arg5 = %189: tensor<1x72x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_55", + OutputName = "Div_55", + shift = 0 : i8} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + %191 = xten_nn.subgraph (%arg5 = %190: tensor<1x72x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#10", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#11", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 72 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + %192 = xten_nn.subgraph (%arg5 = %191: tensor<1x72x23x40xbf16>, %arg6 = %183: tensor<1x72x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x72x23x40xbf16>, %arg8 = %arg6: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_56", + OutputName = "Mul_56", + shift = 0 : i8} : (tensor<1x72x23x40xbf16>, tensor<1x72x23x40xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %462 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + %193 = xten_nn.subgraph (%arg5 = %192: tensor<1x72x23x40xbf16>, %arg6 = %133: tensor<40x72x1x1xbf16>, %arg7 = %132: tensor<40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x23x40xbf16>, %arg9 = %arg6: tensor<40x72x1x1xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc45) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<40x72x1x1xbf16>) -> tensor<40x1x1x72xbf16> loc(#loc45) + %465 = tosa.transpose %arg8, %463 : (tensor<1x72x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x72xbf16> loc(#loc45) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_57", + PartOfOutputName = "Conv_57", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x72xbf16>, tensor<40x1x1x72xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc45) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc45) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc45) + } -> tensor<1x40x23x40xbf16> loc(#loc45) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc45) + } -> tensor<1x40x23x40xbf16> loc(#loc45) + %194 = xten_nn.subgraph (%arg5 = %193: tensor<1x40x23x40xbf16>, %arg6 = %131: tensor<120x40x1x1xbf16>, %arg7 = %130: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc322) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc322) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc322) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_58", + PartOfOutputName = "Conv_58", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc46) + %467 = tosa.clamp %466 { + LayerName = "Relu_59", + OutputName = "Relu_59", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc47) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc322) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc47) + } -> tensor<1x120x23x40xbf16> loc(#loc322) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc322) + } -> tensor<1x120x23x40xbf16> loc(#loc322) + %195 = xten_nn.subgraph (%arg5 = %194: tensor<1x120x23x40xbf16>, %arg6 = %129: tensor<120x1x5x5xbf16>, %arg7 = %128: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc323) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc323) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc323) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_60", + PartOfOutputName = "Conv_60", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc48) + %468 = tosa.clamp %467 { + LayerName = "Relu_61", + OutputName = "Relu_61", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc49) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc323) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc49) + } -> tensor<1x120x23x40xbf16> loc(#loc323) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc323) + } -> tensor<1x120x23x40xbf16> loc(#loc323) + %196 = xten_nn.subgraph (%arg5 = %195: tensor<1x120x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#12", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#13", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc50) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc50) + } -> tensor<1x120x1x920xbf16> loc(#loc50) + %197 = xten_nn.subgraph (%arg5 = %196: tensor<1x120x1x920xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#15", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#15", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc50) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc50) + } -> tensor<1x120x1x1xbf16> loc(#loc50) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc50) + } -> tensor<1x120x1x1xbf16> loc(#loc50) + %198 = xten_nn.subgraph (%arg5 = %197: tensor<1x120x1x1xbf16>, %arg6 = %127: tensor<32x120x1x1xbf16>, %arg7 = %126: tensor<32xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc324) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc324) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_63", + PartOfOutputName = "Conv_63", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc51) + %465 = tosa.clamp %464 { + LayerName = "Relu_64", + OutputName = "Relu_64", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc52) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc324) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc52) + } -> tensor<1x32x1x1xbf16> loc(#loc324) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc324) + } -> tensor<1x32x1x1xbf16> loc(#loc324) + %199 = xten_nn.subgraph (%arg5 = %198: tensor<1x32x1x1xbf16>, %arg6 = %125: tensor<120x32x1x1xbf16>, %arg7 = %124: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc53) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc53) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_65", + PartOfOutputName = "Conv_65", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc53) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc53) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc53) + } -> tensor<1x120x1x1xbf16> loc(#loc53) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc53) + } -> tensor<1x120x1x1xbf16> loc(#loc53) + %200 = xten_nn.subgraph (%arg5 = %199: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_67", OutputName = "Add_67"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc54) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc54) + } -> tensor<1x120x1x1xbf16> loc(#loc54) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc54) + } -> tensor<1x120x1x1xbf16> loc(#loc54) + %201 = xten_nn.subgraph (%arg5 = %200: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_70", + OutputName = "Clip_70", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + %202 = xten_nn.subgraph (%arg5 = %201: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_72", + OutputName = "Div_72", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + %203 = xten_nn.subgraph (%arg5 = %202: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#16", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#17", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + %204 = xten_nn.subgraph (%arg5 = %203: tensor<1x120x23x40xbf16>, %arg6 = %195: tensor<1x120x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_73", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_73", + OutputName = "Mul_73", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + %205 = xten_nn.subgraph (%arg5 = %204: tensor<1x120x23x40xbf16>, %arg6 = %123: tensor<40x120x1x1xbf16>, %arg7 = %122: tensor<40xbf16>, %arg8 = %193: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_74", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_74", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_74", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc58) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc58) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc58) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_74", + PartOfOutputName = "Conv_74", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc58) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc58) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc58) + } -> tensor<1x40x23x40xbf16> loc(#loc58) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_75", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_75", OutputName = "Add_75"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc59) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc59) + } -> tensor<1x40x23x40xbf16> loc(#loc59) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc59) + } -> tensor<1x40x23x40xbf16> loc(#loc325) + %206 = xten_nn.subgraph (%arg5 = %205: tensor<1x40x23x40xbf16>, %arg6 = %121: tensor<120x40x1x1xbf16>, %arg7 = %120: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc326) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc326) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc326) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_76", + PartOfOutputName = "Conv_76", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc60) + %467 = tosa.clamp %466 { + LayerName = "Relu_77", + OutputName = "Relu_77", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc61) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc326) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc61) + } -> tensor<1x120x23x40xbf16> loc(#loc326) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc326) + } -> tensor<1x120x23x40xbf16> loc(#loc326) + %207 = xten_nn.subgraph (%arg5 = %206: tensor<1x120x23x40xbf16>, %arg6 = %119: tensor<120x1x5x5xbf16>, %arg7 = %118: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc327) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc327) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc327) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_78", + PartOfOutputName = "Conv_78", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc62) + %468 = tosa.clamp %467 { + LayerName = "Relu_79", + OutputName = "Relu_79", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc63) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc327) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc63) + } -> tensor<1x120x23x40xbf16> loc(#loc327) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc327) + } -> tensor<1x120x23x40xbf16> loc(#loc327) + %208 = xten_nn.subgraph (%arg5 = %207: tensor<1x120x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#18", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#19", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc64) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc64) + } -> tensor<1x120x1x920xbf16> loc(#loc64) + %209 = xten_nn.subgraph (%arg5 = %208: tensor<1x120x1x920xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#20", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#21", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#20", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#21", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc64) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc64) + } -> tensor<1x120x1x1xbf16> loc(#loc64) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc64) + } -> tensor<1x120x1x1xbf16> loc(#loc64) + %210 = xten_nn.subgraph (%arg5 = %209: tensor<1x120x1x1xbf16>, %arg6 = %117: tensor<32x120x1x1xbf16>, %arg7 = %116: tensor<32xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc328) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc328) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_81", + PartOfOutputName = "Conv_81", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc65) + %465 = tosa.clamp %464 { + LayerName = "Relu_82", + OutputName = "Relu_82", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc66) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc328) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc66) + } -> tensor<1x32x1x1xbf16> loc(#loc328) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc328) + } -> tensor<1x32x1x1xbf16> loc(#loc328) + %211 = xten_nn.subgraph (%arg5 = %210: tensor<1x32x1x1xbf16>, %arg6 = %115: tensor<120x32x1x1xbf16>, %arg7 = %114: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc67) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc67) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_83", + PartOfOutputName = "Conv_83", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc67) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc67) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc67) + } -> tensor<1x120x1x1xbf16> loc(#loc67) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc67) + } -> tensor<1x120x1x1xbf16> loc(#loc67) + %212 = xten_nn.subgraph (%arg5 = %211: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_85", OutputName = "Add_85"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc68) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc68) + } -> tensor<1x120x1x1xbf16> loc(#loc68) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc68) + } -> tensor<1x120x1x1xbf16> loc(#loc68) + %213 = xten_nn.subgraph (%arg5 = %212: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_88", + OutputName = "Clip_88", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + %214 = xten_nn.subgraph (%arg5 = %213: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_90", + OutputName = "Div_90", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + %215 = xten_nn.subgraph (%arg5 = %214: tensor<1x120x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#22", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#23", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + %216 = xten_nn.subgraph (%arg5 = %215: tensor<1x120x23x40xbf16>, %arg6 = %207: tensor<1x120x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_91", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_91", + OutputName = "Mul_91", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + %217 = xten_nn.subgraph (%arg5 = %216: tensor<1x120x23x40xbf16>, %arg6 = %113: tensor<40x120x1x1xbf16>, %arg7 = %112: tensor<40xbf16>, %arg8 = %205: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_92", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_92", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_92", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc72) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc72) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc72) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_92", + PartOfOutputName = "Conv_92", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc72) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc72) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc72) + } -> tensor<1x40x23x40xbf16> loc(#loc72) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_93", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_93", OutputName = "Add_93"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc73) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc73) + } -> tensor<1x40x23x40xbf16> loc(#loc73) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc73) + } -> tensor<1x40x23x40xbf16> loc(#loc329) + %218 = xten_nn.subgraph (%arg5 = %217: tensor<1x40x23x40xbf16>, %arg6 = %111: tensor<240x40x1x1xbf16>, %arg7 = %110: tensor<240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<240x40x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc74) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x40x1x1xbf16>) -> tensor<240x1x1x40xbf16> loc(#loc74) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc74) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_94", + PartOfOutputName = "Conv_94", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<240x1x1x40xbf16>, tensor<240xbf16>) -> tensor<1x23x40x240xbf16> loc(#loc74) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x240xbf16>, tensor<4xi32>) -> tensor<1x240x23x40xbf16> loc(#loc74) + xten_nn.output %467 : tensor<1x240x23x40xbf16> loc(#loc74) + } -> tensor<1x240x23x40xbf16> loc(#loc74) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc74) + } -> tensor<1x240x23x40xbf16> loc(#loc74) + %219 = xten_nn.subgraph (%arg5 = %218: tensor<1x240x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_96", OutputName = "Add_96"} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc75) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc75) + } -> tensor<1x240x23x40xbf16> loc(#loc75) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc75) + } -> tensor<1x240x23x40xbf16> loc(#loc75) + %220 = xten_nn.subgraph (%arg5 = %219: tensor<1x240x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_99", + OutputName = "Clip_99", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + %221 = xten_nn.subgraph (%arg5 = %220: tensor<1x240x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_101", + OutputName = "Div_101", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + %222 = xten_nn.subgraph (%arg5 = %218: tensor<1x240x23x40xbf16>, %arg6 = %221: tensor<1x240x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x23x40xbf16>, %arg8 = %arg6: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_102", + OutputName = "Mul_102", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + %223 = xten_nn.subgraph (%arg5 = %222: tensor<1x240x23x40xbf16>, %arg6 = %109: tensor<240x1x3x3xbf16>, %arg7 = %108: tensor<240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x23x40xbf16>, %arg9 = %arg6: tensor<240x1x3x3xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 0]], + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc79) + %465 = tosa.transpose %arg9, %464 : (tensor<240x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x240x1xbf16> loc(#loc79) + %466 = tosa.transpose %arg8, %463 : (tensor<1x240x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x240xbf16> loc(#loc79) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_103", + PartOfOutputName = "Conv_103", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x240xbf16>, tensor<3x3x240x1xbf16>, tensor<240xbf16>) -> tensor<1x12x20x240xbf16> loc(#loc79) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x240xbf16>, tensor<4xi32>) -> tensor<1x240x12x20xbf16> loc(#loc79) + xten_nn.output %468 : tensor<1x240x12x20xbf16> loc(#loc79) + } -> tensor<1x240x12x20xbf16> loc(#loc79) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc79) + } -> tensor<1x240x12x20xbf16> loc(#loc79) + %224 = xten_nn.subgraph (%arg5 = %223: tensor<1x240x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_105", OutputName = "Add_105"} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc80) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc80) + } -> tensor<1x240x12x20xbf16> loc(#loc80) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc80) + } -> tensor<1x240x12x20xbf16> loc(#loc80) + %225 = xten_nn.subgraph (%arg5 = %224: tensor<1x240x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_108", + OutputName = "Clip_108", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + %226 = xten_nn.subgraph (%arg5 = %225: tensor<1x240x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_110", + OutputName = "Div_110", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + %227 = xten_nn.subgraph (%arg5 = %223: tensor<1x240x12x20xbf16>, %arg6 = %226: tensor<1x240x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x12x20xbf16>, %arg8 = %arg6: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_111", + OutputName = "Mul_111", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + %228 = xten_nn.subgraph (%arg5 = %227: tensor<1x240x12x20xbf16>, %arg6 = %107: tensor<80x240x1x1xbf16>, %arg7 = %106: tensor<80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x12x20xbf16>, %arg9 = %arg6: tensor<80x240x1x1xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc84) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<80x240x1x1xbf16>) -> tensor<80x1x1x240xbf16> loc(#loc84) + %465 = tosa.transpose %arg8, %463 : (tensor<1x240x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x240xbf16> loc(#loc84) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_112", + PartOfOutputName = "Conv_112", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x240xbf16>, tensor<80x1x1x240xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc84) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc84) + xten_nn.output %467 : tensor<1x80x12x20xbf16> loc(#loc84) + } -> tensor<1x80x12x20xbf16> loc(#loc84) + xten_nn.output %461 : tensor<1x80x12x20xbf16> loc(#loc84) + } -> tensor<1x80x12x20xbf16> loc(#loc84) + %229 = xten_nn.subgraph (%arg5 = %228: tensor<1x80x12x20xbf16>, %arg6 = %105: tensor<200x80x1x1xbf16>, %arg7 = %104: tensor<200xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<200x80x1x1xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc85) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<200x80x1x1xbf16>) -> tensor<200x1x1x80xbf16> loc(#loc85) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc85) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_113", + PartOfOutputName = "Conv_113", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<200x1x1x80xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc85) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc85) + xten_nn.output %467 : tensor<1x200x12x20xbf16> loc(#loc85) + } -> tensor<1x200x12x20xbf16> loc(#loc85) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc85) + } -> tensor<1x200x12x20xbf16> loc(#loc85) + %230 = xten_nn.subgraph (%arg5 = %229: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_115", OutputName = "Add_115"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc86) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc86) + } -> tensor<1x200x12x20xbf16> loc(#loc86) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc86) + } -> tensor<1x200x12x20xbf16> loc(#loc86) + %231 = xten_nn.subgraph (%arg5 = %230: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_118", + OutputName = "Clip_118", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + %232 = xten_nn.subgraph (%arg5 = %231: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_120", + OutputName = "Div_120", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + %233 = xten_nn.subgraph (%arg5 = %229: tensor<1x200x12x20xbf16>, %arg6 = %232: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_121", + OutputName = "Mul_121", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + %234 = xten_nn.subgraph (%arg5 = %233: tensor<1x200x12x20xbf16>, %arg6 = %103: tensor<200x1x3x3xbf16>, %arg7 = %102: tensor<200xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x200x12x20xbf16>, %arg9 = %arg6: tensor<200x1x3x3xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc90) + %465 = tosa.transpose %arg9, %464 : (tensor<200x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x200x1xbf16> loc(#loc90) + %466 = tosa.transpose %arg8, %463 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc90) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_122", + PartOfOutputName = "Conv_122", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<3x3x200x1xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc90) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %468 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + %235 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_124", OutputName = "Add_124"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + %236 = xten_nn.subgraph (%arg5 = %235: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_127", + OutputName = "Clip_127", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + %237 = xten_nn.subgraph (%arg5 = %236: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_129", + OutputName = "Div_129", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + %238 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>, %arg6 = %237: tensor<1x200x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_130", + OutputName = "Mul_130", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + %239 = xten_nn.subgraph (%arg5 = %238: tensor<1x200x12x20xbf16>, %arg6 = %101: tensor<80x200x1x1xbf16>, %arg7 = %100: tensor<80xbf16>, %arg8 = %228: tensor<1x80x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_131", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x200x12x20xbf16>, %arg10 = %arg6: tensor<80x200x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_131", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_131", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc95) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x200x1x1xbf16>) -> tensor<80x1x1x200xbf16> loc(#loc95) + %466 = tosa.transpose %arg9, %464 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc95) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_131", + PartOfOutputName = "Conv_131", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<80x1x1x200xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc95) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc95) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc95) + } -> tensor<1x80x12x20xbf16> loc(#loc95) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_132", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_132", OutputName = "Add_132"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc96) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc96) + } -> tensor<1x80x12x20xbf16> loc(#loc96) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc96) + } -> tensor<1x80x12x20xbf16> loc(#loc330) + %240 = xten_nn.subgraph (%arg5 = %239: tensor<1x80x12x20xbf16>, %arg6 = %99: tensor<184x80x1x1xbf16>, %arg7 = %98: tensor<184xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc97) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc97) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc97) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_133", + PartOfOutputName = "Conv_133", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc97) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc97) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc97) + } -> tensor<1x184x12x20xbf16> loc(#loc97) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc97) + } -> tensor<1x184x12x20xbf16> loc(#loc97) + %241 = xten_nn.subgraph (%arg5 = %240: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_135", OutputName = "Add_135"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc98) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc98) + } -> tensor<1x184x12x20xbf16> loc(#loc98) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc98) + } -> tensor<1x184x12x20xbf16> loc(#loc98) + %242 = xten_nn.subgraph (%arg5 = %241: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_138", + OutputName = "Clip_138", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + %243 = xten_nn.subgraph (%arg5 = %242: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_140", + OutputName = "Div_140", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + %244 = xten_nn.subgraph (%arg5 = %240: tensor<1x184x12x20xbf16>, %arg6 = %243: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_141", + OutputName = "Mul_141", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + %245 = xten_nn.subgraph (%arg5 = %244: tensor<1x184x12x20xbf16>, %arg6 = %97: tensor<184x1x3x3xbf16>, %arg7 = %96: tensor<184xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc102) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc102) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc102) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_142", + PartOfOutputName = "Conv_142", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc102) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + %246 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_144", OutputName = "Add_144"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + %247 = xten_nn.subgraph (%arg5 = %246: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_147", + OutputName = "Clip_147", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + %248 = xten_nn.subgraph (%arg5 = %247: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_149", + OutputName = "Div_149", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + %249 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>, %arg6 = %248: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_150", + OutputName = "Mul_150", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + %250 = xten_nn.subgraph (%arg5 = %249: tensor<1x184x12x20xbf16>, %arg6 = %95: tensor<80x184x1x1xbf16>, %arg7 = %94: tensor<80xbf16>, %arg8 = %239: tensor<1x80x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_151", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_151", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_151", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc107) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc107) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc107) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_151", + PartOfOutputName = "Conv_151", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc107) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc107) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc107) + } -> tensor<1x80x12x20xbf16> loc(#loc107) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_152", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_152", OutputName = "Add_152"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc108) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc108) + } -> tensor<1x80x12x20xbf16> loc(#loc108) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc108) + } -> tensor<1x80x12x20xbf16> loc(#loc331) + %251 = xten_nn.subgraph (%arg5 = %250: tensor<1x80x12x20xbf16>, %arg6 = %93: tensor<184x80x1x1xbf16>, %arg7 = %92: tensor<184xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc109) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc109) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc109) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_153", + PartOfOutputName = "Conv_153", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc109) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + %252 = xten_nn.subgraph (%arg5 = %251: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_155", OutputName = "Add_155"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + %253 = xten_nn.subgraph (%arg5 = %252: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_158", + OutputName = "Clip_158", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + %254 = xten_nn.subgraph (%arg5 = %253: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_160", + OutputName = "Div_160", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + %255 = xten_nn.subgraph (%arg5 = %251: tensor<1x184x12x20xbf16>, %arg6 = %254: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_161", + OutputName = "Mul_161", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + %256 = xten_nn.subgraph (%arg5 = %255: tensor<1x184x12x20xbf16>, %arg6 = %91: tensor<184x1x3x3xbf16>, %arg7 = %90: tensor<184xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc114) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc114) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc114) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_162", + PartOfOutputName = "Conv_162", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc114) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + %257 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_164", OutputName = "Add_164"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + %258 = xten_nn.subgraph (%arg5 = %257: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_167", + OutputName = "Clip_167", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + %259 = xten_nn.subgraph (%arg5 = %258: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_169", + OutputName = "Div_169", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + %260 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>, %arg6 = %259: tensor<1x184x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_170", + OutputName = "Mul_170", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + %261 = xten_nn.subgraph (%arg5 = %260: tensor<1x184x12x20xbf16>, %arg6 = %89: tensor<80x184x1x1xbf16>, %arg7 = %88: tensor<80xbf16>, %arg8 = %250: tensor<1x80x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_171", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_171", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_171", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc119) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc119) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc119) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_171", + PartOfOutputName = "Conv_171", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc119) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc119) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc119) + } -> tensor<1x80x12x20xbf16> loc(#loc119) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_172", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_172", OutputName = "Add_172"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc120) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc120) + } -> tensor<1x80x12x20xbf16> loc(#loc120) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc120) + } -> tensor<1x80x12x20xbf16> loc(#loc332) + %262 = xten_nn.subgraph (%arg5 = %261: tensor<1x80x12x20xbf16>, %arg6 = %87: tensor<480x80x1x1xbf16>, %arg7 = %86: tensor<480xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<480x80x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc121) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x80x1x1xbf16>) -> tensor<480x1x1x80xbf16> loc(#loc121) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc121) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_173", + PartOfOutputName = "Conv_173", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<480x1x1x80xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc121) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc121) + xten_nn.output %467 : tensor<1x480x12x20xbf16> loc(#loc121) + } -> tensor<1x480x12x20xbf16> loc(#loc121) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc121) + } -> tensor<1x480x12x20xbf16> loc(#loc121) + %263 = xten_nn.subgraph (%arg5 = %262: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_175", OutputName = "Add_175"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc122) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc122) + } -> tensor<1x480x12x20xbf16> loc(#loc122) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc122) + } -> tensor<1x480x12x20xbf16> loc(#loc122) + %264 = xten_nn.subgraph (%arg5 = %263: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_178", + OutputName = "Clip_178", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + %265 = xten_nn.subgraph (%arg5 = %264: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_180", + OutputName = "Div_180", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + %266 = xten_nn.subgraph (%arg5 = %262: tensor<1x480x12x20xbf16>, %arg6 = %265: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_181", + OutputName = "Mul_181", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + %267 = xten_nn.subgraph (%arg5 = %266: tensor<1x480x12x20xbf16>, %arg6 = %85: tensor<480x1x3x3xbf16>, %arg7 = %84: tensor<480xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<480x1x3x3xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc126) + %465 = tosa.transpose %arg9, %464 : (tensor<480x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x480x1xbf16> loc(#loc126) + %466 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc126) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_182", + PartOfOutputName = "Conv_182", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<3x3x480x1xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc126) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %468 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + %268 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_184", OutputName = "Add_184"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + %269 = xten_nn.subgraph (%arg5 = %268: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_187", + OutputName = "Clip_187", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + %270 = xten_nn.subgraph (%arg5 = %269: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_189", + OutputName = "Div_189", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + %271 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>, %arg6 = %270: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_190", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_190", + OutputName = "Mul_190", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + %272 = xten_nn.subgraph (%arg5 = %271: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#24", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#25", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 60 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x1x240xbf16> loc(#loc333) + xten_nn.output %461 : tensor<1x480x1x240xbf16> loc(#loc333) + } -> tensor<1x480x1x240xbf16> loc(#loc333) + %273 = xten_nn.subgraph (%arg5 = %272: tensor<1x480x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x240xbf16>) attributes { + LayerName = "Generated-#26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#27", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 480 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x480x1x240xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc131) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc131) + } -> tensor<1x480x1x1xbf16> loc(#loc131) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc131) + } -> tensor<1x480x1x1xbf16> loc(#loc131) + %274 = xten_nn.subgraph (%arg5 = %273: tensor<1x480x1x1xbf16>, %arg6 = %83: tensor<120x480x1x1xbf16>, %arg7 = %82: tensor<120xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x1x1xbf16>, %arg9 = %arg6: tensor<120x480x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x480x1x1xbf16>) -> tensor<120x1x1x480xbf16> loc(#loc334) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc334) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_192", + PartOfOutputName = "Conv_192", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x480xbf16>, tensor<120x1x1x480xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc132) + %465 = tosa.clamp %464 { + LayerName = "Relu_193", + OutputName = "Relu_193", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc133) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc334) + xten_nn.output %466 : tensor<1x120x1x1xbf16> loc(#loc133) + } -> tensor<1x120x1x1xbf16> loc(#loc334) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc334) + } -> tensor<1x120x1x1xbf16> loc(#loc334) + %275 = xten_nn.subgraph (%arg5 = %274: tensor<1x120x1x1xbf16>, %arg6 = %81: tensor<480x120x1x1xbf16>, %arg7 = %80: tensor<480xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<480x120x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x120x1x1xbf16>) -> tensor<480x1x1x120xbf16> loc(#loc134) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc134) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_194", + PartOfOutputName = "Conv_194", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<480x1x1x120xbf16>, tensor<480xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc134) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x480xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc134) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc134) + } -> tensor<1x480x1x1xbf16> loc(#loc134) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc134) + } -> tensor<1x480x1x1xbf16> loc(#loc134) + %276 = xten_nn.subgraph (%arg5 = %275: tensor<1x480x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_196", OutputName = "Add_196"} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc135) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc135) + } -> tensor<1x480x1x1xbf16> loc(#loc135) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc135) + } -> tensor<1x480x1x1xbf16> loc(#loc135) + %277 = xten_nn.subgraph (%arg5 = %276: tensor<1x480x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_199", + OutputName = "Clip_199", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + %278 = xten_nn.subgraph (%arg5 = %277: tensor<1x480x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_201", + OutputName = "Div_201", + shift = 0 : i8} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + %279 = xten_nn.subgraph (%arg5 = %278: tensor<1x480x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#28", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#29", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 480 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + %280 = xten_nn.subgraph (%arg5 = %279: tensor<1x480x12x20xbf16>, %arg6 = %271: tensor<1x480x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_202", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_202", + OutputName = "Mul_202", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + %281 = xten_nn.subgraph (%arg5 = %280: tensor<1x480x12x20xbf16>, %arg6 = %79: tensor<112x480x1x1xbf16>, %arg7 = %78: tensor<112xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<112x480x1x1xbf16>, %arg10 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc139) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<112x480x1x1xbf16>) -> tensor<112x1x1x480xbf16> loc(#loc139) + %465 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc139) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_203", + PartOfOutputName = "Conv_203", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<112x1x1x480xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc139) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc139) + xten_nn.output %467 : tensor<1x112x12x20xbf16> loc(#loc139) + } -> tensor<1x112x12x20xbf16> loc(#loc139) + xten_nn.output %461 : tensor<1x112x12x20xbf16> loc(#loc139) + } -> tensor<1x112x12x20xbf16> loc(#loc139) + %282 = xten_nn.subgraph (%arg5 = %281: tensor<1x112x12x20xbf16>, %arg6 = %77: tensor<672x112x1x1xbf16>, %arg7 = %76: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc140) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc140) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc140) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_204", + PartOfOutputName = "Conv_204", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc140) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc140) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc140) + } -> tensor<1x672x12x20xbf16> loc(#loc140) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc140) + } -> tensor<1x672x12x20xbf16> loc(#loc140) + %283 = xten_nn.subgraph (%arg5 = %282: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_206", OutputName = "Add_206"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc141) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc141) + } -> tensor<1x672x12x20xbf16> loc(#loc141) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc141) + } -> tensor<1x672x12x20xbf16> loc(#loc141) + %284 = xten_nn.subgraph (%arg5 = %283: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_209", + OutputName = "Clip_209", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + %285 = xten_nn.subgraph (%arg5 = %284: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_211", + OutputName = "Div_211", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + %286 = xten_nn.subgraph (%arg5 = %282: tensor<1x672x12x20xbf16>, %arg6 = %285: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_212", + OutputName = "Mul_212", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + %287 = xten_nn.subgraph (%arg5 = %286: tensor<1x672x12x20xbf16>, %arg6 = %75: tensor<672x1x3x3xbf16>, %arg7 = %74: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x3x3xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc145) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x672x1xbf16> loc(#loc145) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc145) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_213", + PartOfOutputName = "Conv_213", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<3x3x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc145) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + %288 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_215", OutputName = "Add_215"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + %289 = xten_nn.subgraph (%arg5 = %288: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_218", + OutputName = "Clip_218", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + %290 = xten_nn.subgraph (%arg5 = %289: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_220", + OutputName = "Div_220", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + %291 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>, %arg6 = %290: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_221", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_221", + OutputName = "Mul_221", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + %292 = xten_nn.subgraph (%arg5 = %291: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#30", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#31", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc335) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc335) + } -> tensor<1x672x1x240xbf16> loc(#loc335) + %293 = xten_nn.subgraph (%arg5 = %292: tensor<1x672x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#33", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc150) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc150) + } -> tensor<1x672x1x1xbf16> loc(#loc150) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc150) + } -> tensor<1x672x1x1xbf16> loc(#loc150) + %294 = xten_nn.subgraph (%arg5 = %293: tensor<1x672x1x1xbf16>, %arg6 = %73: tensor<168x672x1x1xbf16>, %arg7 = %72: tensor<168xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc336) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc336) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_223", + PartOfOutputName = "Conv_223", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc151) + %465 = tosa.clamp %464 { + LayerName = "Relu_224", + OutputName = "Relu_224", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc152) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc336) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc152) + } -> tensor<1x168x1x1xbf16> loc(#loc336) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc336) + } -> tensor<1x168x1x1xbf16> loc(#loc336) + %295 = xten_nn.subgraph (%arg5 = %294: tensor<1x168x1x1xbf16>, %arg6 = %71: tensor<672x168x1x1xbf16>, %arg7 = %70: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc153) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc153) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_225", + PartOfOutputName = "Conv_225", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc153) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc153) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc153) + } -> tensor<1x672x1x1xbf16> loc(#loc153) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc153) + } -> tensor<1x672x1x1xbf16> loc(#loc153) + %296 = xten_nn.subgraph (%arg5 = %295: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_227", OutputName = "Add_227"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc154) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc154) + } -> tensor<1x672x1x1xbf16> loc(#loc154) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc154) + } -> tensor<1x672x1x1xbf16> loc(#loc154) + %297 = xten_nn.subgraph (%arg5 = %296: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_230", + OutputName = "Clip_230", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + %298 = xten_nn.subgraph (%arg5 = %297: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_232", + OutputName = "Div_232", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + %299 = xten_nn.subgraph (%arg5 = %298: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#34", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#35", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + %300 = xten_nn.subgraph (%arg5 = %299: tensor<1x672x12x20xbf16>, %arg6 = %291: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_233", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_233", + OutputName = "Mul_233", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + %301 = xten_nn.subgraph (%arg5 = %300: tensor<1x672x12x20xbf16>, %arg6 = %69: tensor<112x672x1x1xbf16>, %arg7 = %68: tensor<112xbf16>, %arg8 = %281: tensor<1x112x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_234", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x672x12x20xbf16>, %arg10 = %arg6: tensor<112x672x1x1xbf16>, %arg11 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_234", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_234", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc158) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<112x672x1x1xbf16>) -> tensor<112x1x1x672xbf16> loc(#loc158) + %466 = tosa.transpose %arg9, %464 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc158) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_234", + PartOfOutputName = "Conv_234", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<112x1x1x672xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc158) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc158) + xten_nn.output %468 : tensor<1x112x12x20xbf16> loc(#loc158) + } -> tensor<1x112x12x20xbf16> loc(#loc158) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x112x12x20xbf16>, %arg10 = %arg8: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Add_235", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_235", OutputName = "Add_235"} : (tensor<1x112x12x20xbf16>, tensor<1x112x12x20xbf16>) -> tensor<1x112x12x20xbf16> loc(#loc159) + xten_nn.output %463 : tensor<1x112x12x20xbf16> loc(#loc159) + } -> tensor<1x112x12x20xbf16> loc(#loc159) + xten_nn.output %462 : tensor<1x112x12x20xbf16> loc(#loc159) + } -> tensor<1x112x12x20xbf16> loc(#loc337) + %302 = xten_nn.subgraph (%arg5 = %301: tensor<1x112x12x20xbf16>, %arg6 = %67: tensor<672x112x1x1xbf16>, %arg7 = %66: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc160) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc160) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc160) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_236", + PartOfOutputName = "Conv_236", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc160) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc160) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc160) + } -> tensor<1x672x12x20xbf16> loc(#loc160) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc160) + } -> tensor<1x672x12x20xbf16> loc(#loc160) + %303 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_238", OutputName = "Add_238"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc161) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc161) + } -> tensor<1x672x12x20xbf16> loc(#loc161) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc161) + } -> tensor<1x672x12x20xbf16> loc(#loc161) + %304 = xten_nn.subgraph (%arg5 = %303: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_241", + OutputName = "Clip_241", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + %305 = xten_nn.subgraph (%arg5 = %304: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_243", + OutputName = "Div_243", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + %306 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>, %arg6 = %305: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_244", + OutputName = "Mul_244", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + %307 = xten_nn.subgraph (%arg5 = %306: tensor<1x672x12x20xbf16>, %arg6 = %65: tensor<672x1x9x9xbf16>, %arg7 = %64: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x9x9xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc165) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x672x1xbf16> loc(#loc165) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc165) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_245", + PartOfOutputName = "Conv_245", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<9x9x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc165) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + %308 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_247", OutputName = "Add_247"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + %309 = xten_nn.subgraph (%arg5 = %308: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_250", + OutputName = "Clip_250", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + %310 = xten_nn.subgraph (%arg5 = %309: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_252", + OutputName = "Div_252", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + %311 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>, %arg6 = %310: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_253", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_253", + OutputName = "Mul_253", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + %312 = xten_nn.subgraph (%arg5 = %311: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#36", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#37", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc338) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc338) + } -> tensor<1x672x1x240xbf16> loc(#loc338) + %313 = xten_nn.subgraph (%arg5 = %312: tensor<1x672x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#38", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#39", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#38", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#39", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc170) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc170) + } -> tensor<1x672x1x1xbf16> loc(#loc170) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc170) + } -> tensor<1x672x1x1xbf16> loc(#loc170) + %314 = xten_nn.subgraph (%arg5 = %313: tensor<1x672x1x1xbf16>, %arg6 = %63: tensor<168x672x1x1xbf16>, %arg7 = %62: tensor<168xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc339) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc339) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_255", + PartOfOutputName = "Conv_255", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc171) + %465 = tosa.clamp %464 { + LayerName = "Relu_256", + OutputName = "Relu_256", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc172) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc339) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc172) + } -> tensor<1x168x1x1xbf16> loc(#loc339) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc339) + } -> tensor<1x168x1x1xbf16> loc(#loc339) + %315 = xten_nn.subgraph (%arg5 = %314: tensor<1x168x1x1xbf16>, %arg6 = %61: tensor<672x168x1x1xbf16>, %arg7 = %60: tensor<672xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc173) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc173) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_257", + PartOfOutputName = "Conv_257", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc173) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc173) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc173) + } -> tensor<1x672x1x1xbf16> loc(#loc173) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc173) + } -> tensor<1x672x1x1xbf16> loc(#loc173) + %316 = xten_nn.subgraph (%arg5 = %315: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_259", OutputName = "Add_259"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc174) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc174) + } -> tensor<1x672x1x1xbf16> loc(#loc174) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc174) + } -> tensor<1x672x1x1xbf16> loc(#loc174) + %317 = xten_nn.subgraph (%arg5 = %316: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_262", + OutputName = "Clip_262", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + %318 = xten_nn.subgraph (%arg5 = %317: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_264", + OutputName = "Div_264", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + %319 = xten_nn.subgraph (%arg5 = %318: tensor<1x672x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#41", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + %320 = xten_nn.subgraph (%arg5 = %319: tensor<1x672x12x20xbf16>, %arg6 = %311: tensor<1x672x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_265", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_265", + OutputName = "Mul_265", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + %321 = xten_nn.subgraph (%arg5 = %320: tensor<1x672x12x20xbf16>, %arg6 = %59: tensor<160x672x1x1xbf16>, %arg7 = %58: tensor<160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<160x672x1x1xbf16>, %arg10 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc178) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<160x672x1x1xbf16>) -> tensor<160x1x1x672xbf16> loc(#loc178) + %465 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc178) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_266", + PartOfOutputName = "Conv_266", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<160x1x1x672xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc178) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc178) + xten_nn.output %467 : tensor<1x160x12x20xbf16> loc(#loc178) + } -> tensor<1x160x12x20xbf16> loc(#loc178) + xten_nn.output %461 : tensor<1x160x12x20xbf16> loc(#loc178) + } -> tensor<1x160x12x20xbf16> loc(#loc178) + %322 = xten_nn.subgraph (%arg5 = %321: tensor<1x160x12x20xbf16>, %arg6 = %57: tensor<960x160x1x1xbf16>, %arg7 = %56: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc179) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc179) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc179) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_267", + PartOfOutputName = "Conv_267", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc179) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc179) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc179) + } -> tensor<1x960x12x20xbf16> loc(#loc179) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc179) + } -> tensor<1x960x12x20xbf16> loc(#loc179) + %323 = xten_nn.subgraph (%arg5 = %322: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_269", OutputName = "Add_269"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc180) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc180) + } -> tensor<1x960x12x20xbf16> loc(#loc180) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc180) + } -> tensor<1x960x12x20xbf16> loc(#loc180) + %324 = xten_nn.subgraph (%arg5 = %323: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_272", + OutputName = "Clip_272", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + %325 = xten_nn.subgraph (%arg5 = %324: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_274", + OutputName = "Div_274", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + %326 = xten_nn.subgraph (%arg5 = %322: tensor<1x960x12x20xbf16>, %arg6 = %325: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_275", + OutputName = "Mul_275", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + %327 = xten_nn.subgraph (%arg5 = %326: tensor<1x960x12x20xbf16>, %arg6 = %55: tensor<960x1x9x9xbf16>, %arg7 = %54: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc184) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc184) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc184) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_276", + PartOfOutputName = "Conv_276", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc184) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + %328 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_278", OutputName = "Add_278"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + %329 = xten_nn.subgraph (%arg5 = %328: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_281", + OutputName = "Clip_281", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + %330 = xten_nn.subgraph (%arg5 = %329: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_283", + OutputName = "Div_283", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + %331 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>, %arg6 = %330: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_284", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_284", + OutputName = "Mul_284", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + %332 = xten_nn.subgraph (%arg5 = %331: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#42", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#43", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc340) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc340) + } -> tensor<1x960x1x240xbf16> loc(#loc340) + %333 = xten_nn.subgraph (%arg5 = %332: tensor<1x960x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#44", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#45", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#44", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#45", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc189) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc189) + } -> tensor<1x960x1x1xbf16> loc(#loc189) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc189) + } -> tensor<1x960x1x1xbf16> loc(#loc189) + %334 = xten_nn.subgraph (%arg5 = %333: tensor<1x960x1x1xbf16>, %arg6 = %53: tensor<240x960x1x1xbf16>, %arg7 = %52: tensor<240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc341) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc341) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_286", + PartOfOutputName = "Conv_286", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc190) + %465 = tosa.clamp %464 { + LayerName = "Relu_287", + OutputName = "Relu_287", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc191) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc341) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc191) + } -> tensor<1x240x1x1xbf16> loc(#loc341) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc341) + } -> tensor<1x240x1x1xbf16> loc(#loc341) + %335 = xten_nn.subgraph (%arg5 = %334: tensor<1x240x1x1xbf16>, %arg6 = %51: tensor<960x240x1x1xbf16>, %arg7 = %50: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc192) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc192) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_288", + PartOfOutputName = "Conv_288", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc192) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc192) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc192) + } -> tensor<1x960x1x1xbf16> loc(#loc192) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc192) + } -> tensor<1x960x1x1xbf16> loc(#loc192) + %336 = xten_nn.subgraph (%arg5 = %335: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_290", OutputName = "Add_290"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc193) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc193) + } -> tensor<1x960x1x1xbf16> loc(#loc193) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc193) + } -> tensor<1x960x1x1xbf16> loc(#loc193) + %337 = xten_nn.subgraph (%arg5 = %336: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_293", + OutputName = "Clip_293", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + %338 = xten_nn.subgraph (%arg5 = %337: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_295", + OutputName = "Div_295", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + %339 = xten_nn.subgraph (%arg5 = %338: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#46", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#47", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + %340 = xten_nn.subgraph (%arg5 = %339: tensor<1x960x12x20xbf16>, %arg6 = %331: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_296", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_296", + OutputName = "Mul_296", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + %341 = xten_nn.subgraph (%arg5 = %340: tensor<1x960x12x20xbf16>, %arg6 = %49: tensor<160x960x1x1xbf16>, %arg7 = %48: tensor<160xbf16>, %arg8 = %321: tensor<1x160x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_297", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_297", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_297", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc197) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc197) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc197) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_297", + PartOfOutputName = "Conv_297", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc197) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc197) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc197) + } -> tensor<1x160x12x20xbf16> loc(#loc197) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_298", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_298", OutputName = "Add_298"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc198) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc198) + } -> tensor<1x160x12x20xbf16> loc(#loc198) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc198) + } -> tensor<1x160x12x20xbf16> loc(#loc342) + %342 = xten_nn.subgraph (%arg5 = %341: tensor<1x160x12x20xbf16>, %arg6 = %47: tensor<960x160x1x1xbf16>, %arg7 = %46: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc199) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc199) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc199) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_299", + PartOfOutputName = "Conv_299", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc199) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc199) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc199) + } -> tensor<1x960x12x20xbf16> loc(#loc199) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc199) + } -> tensor<1x960x12x20xbf16> loc(#loc199) + %343 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_301", OutputName = "Add_301"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc200) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc200) + } -> tensor<1x960x12x20xbf16> loc(#loc200) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc200) + } -> tensor<1x960x12x20xbf16> loc(#loc200) + %344 = xten_nn.subgraph (%arg5 = %343: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_304", + OutputName = "Clip_304", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + %345 = xten_nn.subgraph (%arg5 = %344: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_306", + OutputName = "Div_306", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + %346 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>, %arg6 = %345: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_307", + OutputName = "Mul_307", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + %347 = xten_nn.subgraph (%arg5 = %346: tensor<1x960x12x20xbf16>, %arg6 = %45: tensor<960x1x9x9xbf16>, %arg7 = %44: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc204) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc204) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc204) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_308", + PartOfOutputName = "Conv_308", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc204) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + %348 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_310", OutputName = "Add_310"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + %349 = xten_nn.subgraph (%arg5 = %348: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_313", + OutputName = "Clip_313", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + %350 = xten_nn.subgraph (%arg5 = %349: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_315", + OutputName = "Div_315", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + %351 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>, %arg6 = %350: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_316", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_316", + OutputName = "Mul_316", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + %352 = xten_nn.subgraph (%arg5 = %351: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#48", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#49", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc343) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc343) + } -> tensor<1x960x1x240xbf16> loc(#loc343) + %353 = xten_nn.subgraph (%arg5 = %352: tensor<1x960x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#51", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#51", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc209) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc209) + } -> tensor<1x960x1x1xbf16> loc(#loc209) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc209) + } -> tensor<1x960x1x1xbf16> loc(#loc209) + %354 = xten_nn.subgraph (%arg5 = %353: tensor<1x960x1x1xbf16>, %arg6 = %43: tensor<240x960x1x1xbf16>, %arg7 = %42: tensor<240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc344) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc344) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_318", + PartOfOutputName = "Conv_318", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc210) + %465 = tosa.clamp %464 { + LayerName = "Relu_319", + OutputName = "Relu_319", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc211) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc344) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc211) + } -> tensor<1x240x1x1xbf16> loc(#loc344) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc344) + } -> tensor<1x240x1x1xbf16> loc(#loc344) + %355 = xten_nn.subgraph (%arg5 = %354: tensor<1x240x1x1xbf16>, %arg6 = %41: tensor<960x240x1x1xbf16>, %arg7 = %40: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc212) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc212) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_320", + PartOfOutputName = "Conv_320", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc212) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc212) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc212) + } -> tensor<1x960x1x1xbf16> loc(#loc212) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc212) + } -> tensor<1x960x1x1xbf16> loc(#loc212) + %356 = xten_nn.subgraph (%arg5 = %355: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_322", OutputName = "Add_322"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc213) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc213) + } -> tensor<1x960x1x1xbf16> loc(#loc213) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc213) + } -> tensor<1x960x1x1xbf16> loc(#loc213) + %357 = xten_nn.subgraph (%arg5 = %356: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_325", + OutputName = "Clip_325", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + %358 = xten_nn.subgraph (%arg5 = %357: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_327", + OutputName = "Div_327", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + %359 = xten_nn.subgraph (%arg5 = %358: tensor<1x960x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#52", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#53", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + %360 = xten_nn.subgraph (%arg5 = %359: tensor<1x960x12x20xbf16>, %arg6 = %351: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_328", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_328", + OutputName = "Mul_328", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + %361 = xten_nn.subgraph (%arg5 = %360: tensor<1x960x12x20xbf16>, %arg6 = %39: tensor<160x960x1x1xbf16>, %arg7 = %38: tensor<160xbf16>, %arg8 = %341: tensor<1x160x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_329", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_329", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_329", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc217) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc217) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc217) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_329", + PartOfOutputName = "Conv_329", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc217) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc217) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc217) + } -> tensor<1x160x12x20xbf16> loc(#loc217) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_330", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_330", OutputName = "Add_330"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc218) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc218) + } -> tensor<1x160x12x20xbf16> loc(#loc218) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc218) + } -> tensor<1x160x12x20xbf16> loc(#loc345) + %362 = xten_nn.subgraph (%arg5 = %361: tensor<1x160x12x20xbf16>, %arg6 = %37: tensor<960x160x1x1xbf16>, %arg7 = %36: tensor<960xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc219) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc219) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc219) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_331", + PartOfOutputName = "Conv_331", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc219) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc219) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc219) + } -> tensor<1x960x12x20xbf16> loc(#loc219) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc219) + } -> tensor<1x960x12x20xbf16> loc(#loc219) + %363 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_333", OutputName = "Add_333"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc220) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc220) + } -> tensor<1x960x12x20xbf16> loc(#loc220) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc220) + } -> tensor<1x960x12x20xbf16> loc(#loc220) + %364 = xten_nn.subgraph (%arg5 = %363: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_336", + OutputName = "Clip_336", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + %365 = xten_nn.subgraph (%arg5 = %364: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_338", + OutputName = "Div_338", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + %366 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>, %arg6 = %365: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_339", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_339", + OutputName = "Mul_339", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + %367 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#54", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#55", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc346) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc346) + } -> tensor<1x960x1x240xbf16> loc(#loc346) + %368 = xten_nn.subgraph (%arg5 = %367: tensor<1x960x1x240xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#57", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc224) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc224) + } -> tensor<1x960x1x1xbf16> loc(#loc224) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc224) + } -> tensor<1x960x1x1xbf16> loc(#loc224) + %369 = xten_nn.subgraph (%arg5 = %368: tensor<1x960x1x1xbf16>, %arg6 = %35: tensor<128x960x1x1xbf16>, %arg7 = %34: tensor<128xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<128x960x1x1xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc225) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc225) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_343", + PartOfOutputName = "Conv_343", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x1x1x128xbf16> loc(#loc225) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x128xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc225) + xten_nn.output %465 : tensor<1x128x1x1xbf16> loc(#loc225) + } -> tensor<1x128x1x1xbf16> loc(#loc225) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc225) + } -> tensor<1x128x1x1xbf16> loc(#loc225) + %370 = xten_nn.subgraph (%arg5 = %369: tensor<1x128x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_344", OutputName = "Sigmoid_344"} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc226) + xten_nn.output %462 : tensor<1x128x1x1xbf16> loc(#loc226) + } -> tensor<1x128x1x1xbf16> loc(#loc226) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc226) + } -> tensor<1x128x1x1xbf16> loc(#loc226) + %371 = xten_nn.subgraph (%arg5 = %370: tensor<1x128x1x1xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Generated-#58", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#59", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 128 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc227) + %372 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>, %arg6 = %33: tensor<128x960x1x1xbf16>, %arg7 = %32: tensor<128xbf16>, %arg8 = %371: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index, 3 : index], + LayerName = "Conv_340", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<128x960x1x1xbf16>, %arg11 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_340", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_341", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc348) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc348) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc348) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_340", + PartOfOutputName = "Conv_340", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc228) + %468 = tosa.clamp %467 { + LayerName = "Relu_341", + OutputName = "Relu_341", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x12x20x128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc229) + %469 = tosa.transpose %468, %463 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc348) + xten_nn.output %469 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc348) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x128x12x20xbf16>, %arg10 = %arg8: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Mul_345", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.mul %arg9, %arg10 { + LayerName = "Mul_345", + OutputName = "Mul_345", + shift = 0 : i8} : (tensor<1x128x12x20xbf16>, tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %463 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc347) + %373 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_349_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc230) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc230) + } -> tensor<1x64x12x20xbf16> loc(#loc230) + %374 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_349_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc230) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc230) + } -> tensor<1x64x12x20xbf16> loc(#loc230) + %375 = xten_nn.subgraph (%arg5 = %374: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_350", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_350", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_350", + OutputName = "Concat_350", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc231) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc231) + } -> tensor<1x128x12x20xbf16> loc(#loc231) + %376 = xten_nn.subgraph (%arg5 = %375: tensor<1x128x12x20xbf16>, %arg6 = %31: tensor<128x128x3x3xbf16>, %arg7 = %30: tensor<128xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<128x128x3x3xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<128x128x3x3xbf16>, tensor<4xi32>) -> tensor<128x3x3x128xbf16> loc(#loc232) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc232) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_351", + PartOfOutputName = "Conv_351", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<128x3x3x128xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc232) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc232) + xten_nn.output %467 : tensor<1x128x12x20xbf16> loc(#loc232) + } -> tensor<1x128x12x20xbf16> loc(#loc232) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc232) + } -> tensor<1x128x12x20xbf16> loc(#loc232) + %377 = xten_nn.subgraph (%arg5 = %376: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_352", OutputName = "Sigmoid_352"} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + %378 = xten_nn.subgraph (%arg5 = %377: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_353_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc234) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc234) + } -> tensor<1x64x12x20xbf16> loc(#loc234) + %379 = xten_nn.subgraph (%arg5 = %27: tensor<1x64x12x20xbf16>, %arg6 = %378: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [1 : index], + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_359", OutputName = "Sub_359"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + %380 = xten_nn.subgraph (%arg5 = %379: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_360", + OutputName = "Mul_360", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + %381 = xten_nn.subgraph (%arg5 = %377: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_353_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc234) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc234) + } -> tensor<1x64x12x20xbf16> loc(#loc234) + %382 = xten_nn.subgraph (%arg5 = %381: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_354", + OutputName = "Mul_354", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc235) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc235) + } -> tensor<1x64x12x20xbf16> loc(#loc235) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc235) + } -> tensor<1x64x12x20xbf16> loc(#loc235) + %383 = xten_nn.subgraph (%arg5 = %374: tensor<1x64x12x20xbf16>, %arg6 = %382: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_355", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_355", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_355", + OutputName = "Concat_355", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc236) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc236) + } -> tensor<1x128x12x20xbf16> loc(#loc236) + %384 = xten_nn.subgraph (%arg5 = %383: tensor<1x128x12x20xbf16>, %arg6 = %29: tensor<64x128x3x3xbf16>, %arg7 = %28: tensor<64xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<64x128x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<64x128x3x3xbf16>, tensor<4xi32>) -> tensor<64x3x3x128xbf16> loc(#loc237) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc237) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_356", + PartOfOutputName = "Conv_356", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<64x3x3x128xbf16>, tensor<64xbf16>) -> tensor<1x12x20x64xbf16> loc(#loc237) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x64xbf16>, tensor<4xi32>) -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %467 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + %385 = xten_nn.subgraph (%arg5 = %384: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_357", OutputName = "Tanh_357"} : (tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + %386 = xten_nn.subgraph (%arg5 = %378: tensor<1x64x12x20xbf16>, %arg6 = %385: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_361", + OutputName = "Mul_361", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc239) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc239) + } -> tensor<1x64x12x20xbf16> loc(#loc239) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc239) + } -> tensor<1x64x12x20xbf16> loc(#loc239) + %387 = xten_nn.subgraph (%arg5 = %380: tensor<1x64x12x20xbf16>, %arg6 = %386: tensor<1x64x12x20xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_362", OutputName = "Add_362"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + %388 = xten_nn.subgraph (%arg5 = %373: tensor<1x64x12x20xbf16>, %arg6 = %387: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_363", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_363", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_363", + OutputName = "Concat_363", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc242) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc242) + } -> tensor<1x128x12x20xbf16> loc(#loc242) + %389 = xten_nn.subgraph (%arg5 = %388: tensor<1x128x12x20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Resize_365", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Resize_365", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 128 : ui32, + config.dim_2 = 12 : ui32, + config.dim_3 = 20 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.output_H = 24 : ui32, + config.output_W = 40 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_365", + OutputName = "Resize_365", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x24x40xbf16> loc(#loc243) + xten_nn.output %461 : tensor<1x128x24x40xbf16> loc(#loc243) + } -> tensor<1x128x24x40xbf16> loc(#loc243) + %390 = xten_nn.subgraph (%arg5 = %389: tensor<1x128x24x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Slice_371", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + OutputName = "Slice_371", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 128 : ui32, + config.dim_h = 24 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 23 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_371", + OutputName = "Slice_371", + size = array, + start = array} : (tensor<1x128x24x40xbf16>) -> tensor<1x128x23x40xbf16> loc(#loc244) + xten_nn.output %461 : tensor<1x128x23x40xbf16> loc(#loc244) + } -> tensor<1x128x23x40xbf16> loc(#loc244) + %391 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc12) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc12) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_346", + PartOfOutputName = "AveragePool_346", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>) -> tensor<1x90x160x3xbf16> loc(#loc12) + %466 = tosa.transpose %465, %462 : (tensor<1x90x160x3xbf16>, tensor<4xi32>) -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %466 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %461 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + %392 = xten_nn.subgraph (%arg5 = %391: tensor<1x3x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x90x160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc245) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x3xbf16> loc(#loc245) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_347", + PartOfOutputName = "AveragePool_347", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x90x160x3xbf16>) -> tensor<1x45x80x3xbf16> loc(#loc245) + %466 = tosa.transpose %465, %462 : (tensor<1x45x80x3xbf16>, tensor<4xi32>) -> tensor<1x3x45x80xbf16> loc(#loc245) + xten_nn.output %466 : tensor<1x3x45x80xbf16> loc(#loc245) + } -> tensor<1x3x45x80xbf16> loc(#loc245) + xten_nn.output %461 : tensor<1x3x45x80xbf16> loc(#loc245) + } -> tensor<1x3x45x80xbf16> loc(#loc245) + %393 = xten_nn.subgraph (%arg5 = %392: tensor<1x3x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x45x80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 1], [0, 0]], + HWPaddingNotCounted = [[0, 1], [0, 0]], + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc246) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x3xbf16> loc(#loc246) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_348", + PartOfOutputName = "AveragePool_348", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x45x80x3xbf16>) -> tensor<1x23x40x3xbf16> loc(#loc246) + %466 = tosa.transpose %465, %462 : (tensor<1x23x40x3xbf16>, tensor<4xi32>) -> tensor<1x3x23x40xbf16> loc(#loc246) + xten_nn.output %466 : tensor<1x3x23x40xbf16> loc(#loc246) + } -> tensor<1x3x23x40xbf16> loc(#loc246) + xten_nn.output %461 : tensor<1x3x23x40xbf16> loc(#loc246) + } -> tensor<1x3x23x40xbf16> loc(#loc246) + %394 = xten_nn.subgraph (%arg5 = %390: tensor<1x128x23x40xbf16>, %arg6 = %217: tensor<1x40x23x40xbf16>, %arg7 = %393: tensor<1x3x23x40xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index, 2 : index], + LayerName = "Concat_372", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_372", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_372", + OutputName = "Concat_372", + axis = 1 : i32} : (tensor<1x128x23x40xbf16>, tensor<1x40x23x40xbf16>, tensor<1x3x23x40xbf16>) -> tensor<1x171x23x40xbf16> loc(#loc247) + xten_nn.output %461 : tensor<1x171x23x40xbf16> loc(#loc247) + } -> tensor<1x171x23x40xbf16> loc(#loc247) + %395 = xten_nn.subgraph (%arg5 = %394: tensor<1x171x23x40xbf16>, %arg6 = %26: tensor<80x171x3x3xbf16>, %arg7 = %25: tensor<80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x171x23x40xbf16>, %arg9 = %arg6: tensor<80x171x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x171x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x171xbf16> loc(#loc349) + %465 = tosa.transpose %arg8, %463 : (tensor<1x171x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x171xbf16> loc(#loc349) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_373", + PartOfOutputName = "Conv_373", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x171xbf16>, tensor<80x3x3x171xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc248) + %467 = tosa.clamp %466 { + LayerName = "Relu_374", + OutputName = "Relu_374", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc249) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc349) + xten_nn.output %468 : tensor<1x80x23x40xbf16> loc(#loc249) + } -> tensor<1x80x23x40xbf16> loc(#loc349) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc349) + } -> tensor<1x80x23x40xbf16> loc(#loc349) + %396 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_375_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %397 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_375_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %398 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_376", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_376", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_376", + OutputName = "Concat_376", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc251) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc251) + } -> tensor<1x80x23x40xbf16> loc(#loc251) + %399 = xten_nn.subgraph (%arg5 = %398: tensor<1x80x23x40xbf16>, %arg6 = %24: tensor<80x80x3x3xbf16>, %arg7 = %23: tensor<80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<80x80x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x80x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x80xbf16> loc(#loc252) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_377", + PartOfOutputName = "Conv_377", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<80x3x3x80xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %467 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + %400 = xten_nn.subgraph (%arg5 = %399: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_378", OutputName = "Sigmoid_378"} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %462 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + %401 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_379_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %402 = xten_nn.subgraph (%arg5 = %20: tensor<1x40x23x40xbf16>, %arg6 = %401: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [1 : index], + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_385", OutputName = "Sub_385"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + %403 = xten_nn.subgraph (%arg5 = %402: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_386", + OutputName = "Mul_386", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + %404 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_379_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %405 = xten_nn.subgraph (%arg5 = %404: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_380", + OutputName = "Mul_380", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + %406 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %405: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_381", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_381", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_381", + OutputName = "Concat_381", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc256) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc256) + } -> tensor<1x80x23x40xbf16> loc(#loc256) + %407 = xten_nn.subgraph (%arg5 = %406: tensor<1x80x23x40xbf16>, %arg6 = %22: tensor<40x80x3x3xbf16>, %arg7 = %21: tensor<40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<40x80x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x80x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x80xbf16> loc(#loc257) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc257) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_382", + PartOfOutputName = "Conv_382", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<40x3x3x80xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc257) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc257) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc257) + } -> tensor<1x40x23x40xbf16> loc(#loc257) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc257) + } -> tensor<1x40x23x40xbf16> loc(#loc257) + %408 = xten_nn.subgraph (%arg5 = %407: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_383", OutputName = "Tanh_383"} : (tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + %409 = xten_nn.subgraph (%arg5 = %401: tensor<1x40x23x40xbf16>, %arg6 = %408: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_387", + OutputName = "Mul_387", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + %410 = xten_nn.subgraph (%arg5 = %403: tensor<1x40x23x40xbf16>, %arg6 = %409: tensor<1x40x23x40xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_388", OutputName = "Add_388"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + %411 = xten_nn.subgraph (%arg5 = %396: tensor<1x40x23x40xbf16>, %arg6 = %410: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_389", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_389", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_389", + OutputName = "Concat_389", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc262) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc262) + } -> tensor<1x80x23x40xbf16> loc(#loc262) + %412 = xten_nn.subgraph (%arg5 = %411: tensor<1x80x23x40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Resize_391", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Resize_391", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 80 : ui32, + config.dim_2 = 23 : ui32, + config.dim_3 = 40 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.output_H = 46 : ui32, + config.output_W = 80 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_391", + OutputName = "Resize_391", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x46x80xbf16> loc(#loc263) + xten_nn.output %461 : tensor<1x80x46x80xbf16> loc(#loc263) + } -> tensor<1x80x46x80xbf16> loc(#loc263) + %413 = xten_nn.subgraph (%arg5 = %412: tensor<1x80x46x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Slice_397", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + OutputName = "Slice_397", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 80 : ui32, + config.dim_h = 46 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 45 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_397", + OutputName = "Slice_397", + size = array, + start = array} : (tensor<1x80x46x80xbf16>) -> tensor<1x80x45x80xbf16> loc(#loc264) + xten_nn.output %461 : tensor<1x80x45x80xbf16> loc(#loc264) + } -> tensor<1x80x45x80xbf16> loc(#loc264) + %414 = xten_nn.subgraph (%arg5 = %413: tensor<1x80x45x80xbf16>, %arg6 = %181: tensor<1x24x45x80xbf16>, %arg7 = %392: tensor<1x3x45x80xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index, 2 : index], + LayerName = "Concat_398", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_398", + OutputName = "Concat_398", + axis = 1 : i32} : (tensor<1x80x45x80xbf16>, tensor<1x24x45x80xbf16>, tensor<1x3x45x80xbf16>) -> tensor<1x107x45x80xbf16> loc(#loc265) + xten_nn.output %461 : tensor<1x107x45x80xbf16> loc(#loc265) + } -> tensor<1x107x45x80xbf16> loc(#loc265) + %415 = xten_nn.subgraph (%arg5 = %414: tensor<1x107x45x80xbf16>, %arg6 = %19: tensor<40x107x3x3xbf16>, %arg7 = %18: tensor<40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x107x45x80xbf16>, %arg9 = %arg6: tensor<40x107x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x107x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x107xbf16> loc(#loc350) + %465 = tosa.transpose %arg8, %463 : (tensor<1x107x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x107xbf16> loc(#loc350) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_399", + PartOfOutputName = "Conv_399", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x107xbf16>, tensor<40x3x3x107xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc266) + %467 = tosa.clamp %466 { + LayerName = "Relu_400", + OutputName = "Relu_400", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc267) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc350) + xten_nn.output %468 : tensor<1x40x45x80xbf16> loc(#loc267) + } -> tensor<1x40x45x80xbf16> loc(#loc350) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc350) + } -> tensor<1x40x45x80xbf16> loc(#loc350) + %416 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_401_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %417 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_401_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %418 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_402", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_402", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_402", + OutputName = "Concat_402", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc269) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc269) + } -> tensor<1x40x45x80xbf16> loc(#loc269) + %419 = xten_nn.subgraph (%arg5 = %418: tensor<1x40x45x80xbf16>, %arg6 = %17: tensor<40x40x3x3xbf16>, %arg7 = %16: tensor<40xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<40x40x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x40x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x40xbf16> loc(#loc270) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_403", + PartOfOutputName = "Conv_403", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<40x3x3x40xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %467 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + %420 = xten_nn.subgraph (%arg5 = %419: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_404", OutputName = "Sigmoid_404"} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %462 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + %421 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_405_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %422 = xten_nn.subgraph (%arg5 = %13: tensor<1x20x45x80xbf16>, %arg6 = %421: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [1 : index], + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_411", OutputName = "Sub_411"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + %423 = xten_nn.subgraph (%arg5 = %422: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_412", + OutputName = "Mul_412", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + %424 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_405_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %425 = xten_nn.subgraph (%arg5 = %424: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_406", + OutputName = "Mul_406", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + %426 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %425: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_407", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_407", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_407", + OutputName = "Concat_407", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc274) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc274) + } -> tensor<1x40x45x80xbf16> loc(#loc274) + %427 = xten_nn.subgraph (%arg5 = %426: tensor<1x40x45x80xbf16>, %arg6 = %15: tensor<20x40x3x3xbf16>, %arg7 = %14: tensor<20xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<20x40x3x3xbf16>, %arg10 = %arg7: tensor<20xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<20x40x3x3xbf16>, tensor<4xi32>) -> tensor<20x3x3x40xbf16> loc(#loc275) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc275) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_408", + PartOfOutputName = "Conv_408", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<20x3x3x40xbf16>, tensor<20xbf16>) -> tensor<1x45x80x20xbf16> loc(#loc275) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x20xbf16>, tensor<4xi32>) -> tensor<1x20x45x80xbf16> loc(#loc275) + xten_nn.output %467 : tensor<1x20x45x80xbf16> loc(#loc275) + } -> tensor<1x20x45x80xbf16> loc(#loc275) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc275) + } -> tensor<1x20x45x80xbf16> loc(#loc275) + %428 = xten_nn.subgraph (%arg5 = %427: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_409", OutputName = "Tanh_409"} : (tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + %429 = xten_nn.subgraph (%arg5 = %421: tensor<1x20x45x80xbf16>, %arg6 = %428: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_413", + OutputName = "Mul_413", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + %430 = xten_nn.subgraph (%arg5 = %423: tensor<1x20x45x80xbf16>, %arg6 = %429: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_414", OutputName = "Add_414"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + %431 = xten_nn.subgraph (%arg5 = %416: tensor<1x20x45x80xbf16>, %arg6 = %430: tensor<1x20x45x80xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_415", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_415", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_415", + OutputName = "Concat_415", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc280) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc280) + } -> tensor<1x40x45x80xbf16> loc(#loc280) + %432 = xten_nn.subgraph (%arg5 = %431: tensor<1x40x45x80xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Resize_417", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Resize_417", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 40 : ui32, + config.dim_2 = 45 : ui32, + config.dim_3 = 80 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.output_H = 90 : ui32, + config.output_W = 160 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_417", + OutputName = "Resize_417", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x90x160xbf16> loc(#loc281) + xten_nn.output %461 : tensor<1x40x90x160xbf16> loc(#loc281) + } -> tensor<1x40x90x160xbf16> loc(#loc281) + %433 = xten_nn.subgraph (%arg5 = %432: tensor<1x40x90x160xbf16>, %arg6 = %175: tensor<1x16x90x160xbf16>, %arg7 = %391: tensor<1x3x90x160xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index, 2 : index], + LayerName = "Concat_418", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_418", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_418", + OutputName = "Concat_418", + axis = 1 : i32} : (tensor<1x40x90x160xbf16>, tensor<1x16x90x160xbf16>, tensor<1x3x90x160xbf16>) -> tensor<1x59x90x160xbf16> loc(#loc282) + xten_nn.output %461 : tensor<1x59x90x160xbf16> loc(#loc282) + } -> tensor<1x59x90x160xbf16> loc(#loc282) + %434 = xten_nn.subgraph (%arg5 = %433: tensor<1x59x90x160xbf16>, %arg6 = %12: tensor<32x59x3x3xbf16>, %arg7 = %11: tensor<32xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x59x90x160xbf16>, %arg9 = %arg6: tensor<32x59x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x59x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x59xbf16> loc(#loc351) + %465 = tosa.transpose %arg8, %463 : (tensor<1x59x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x59xbf16> loc(#loc351) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_419", + PartOfOutputName = "Conv_419", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x59xbf16>, tensor<32x3x3x59xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc283) + %467 = tosa.clamp %466 { + LayerName = "Relu_420", + OutputName = "Relu_420", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc284) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc351) + xten_nn.output %468 : tensor<1x32x90x160xbf16> loc(#loc284) + } -> tensor<1x32x90x160xbf16> loc(#loc351) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc351) + } -> tensor<1x32x90x160xbf16> loc(#loc351) + %435 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_421_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %436 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_421_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %437 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_422", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_422", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_422", + OutputName = "Concat_422", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc286) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc286) + } -> tensor<1x32x90x160xbf16> loc(#loc286) + %438 = xten_nn.subgraph (%arg5 = %437: tensor<1x32x90x160xbf16>, %arg6 = %10: tensor<32x32x3x3xbf16>, %arg7 = %9: tensor<32xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<32x32x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x32x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x32xbf16> loc(#loc287) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_423", + PartOfOutputName = "Conv_423", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<32x3x3x32xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %467 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + %439 = xten_nn.subgraph (%arg5 = %438: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_424", OutputName = "Sigmoid_424"} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %462 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + %440 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_425_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %441 = xten_nn.subgraph (%arg5 = %6: tensor<1x16x90x160xbf16>, %arg6 = %440: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [1 : index], + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_431", OutputName = "Sub_431"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + %442 = xten_nn.subgraph (%arg5 = %441: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_432", + OutputName = "Mul_432", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + %443 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_425_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %444 = xten_nn.subgraph (%arg5 = %443: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_426", + OutputName = "Mul_426", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + %445 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %444: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_427", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_427", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_427", + OutputName = "Concat_427", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc291) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc291) + } -> tensor<1x32x90x160xbf16> loc(#loc291) + %446 = xten_nn.subgraph (%arg5 = %445: tensor<1x32x90x160xbf16>, %arg6 = %8: tensor<16x32x3x3xbf16>, %arg7 = %7: tensor<16xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<16x32x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x32x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x32xbf16> loc(#loc292) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc292) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_428", + PartOfOutputName = "Conv_428", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<16x3x3x32xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc292) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc292) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc292) + } -> tensor<1x16x90x160xbf16> loc(#loc292) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc292) + } -> tensor<1x16x90x160xbf16> loc(#loc292) + %447 = xten_nn.subgraph (%arg5 = %446: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_429", OutputName = "Tanh_429"} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + %448 = xten_nn.subgraph (%arg5 = %440: tensor<1x16x90x160xbf16>, %arg6 = %447: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_433", + OutputName = "Mul_433", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + %449 = xten_nn.subgraph (%arg5 = %442: tensor<1x16x90x160xbf16>, %arg6 = %448: tensor<1x16x90x160xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_434", OutputName = "Add_434"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + %450 = xten_nn.subgraph (%arg5 = %435: tensor<1x16x90x160xbf16>, %arg6 = %449: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_435", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_435", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_435", + OutputName = "Concat_435", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc297) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc297) + } -> tensor<1x32x90x160xbf16> loc(#loc297) + %451 = xten_nn.subgraph (%arg5 = %450: tensor<1x32x90x160xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Resize_437", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Resize_437", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 32 : ui32, + config.dim_2 = 90 : ui32, + config.dim_3 = 160 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.output_H = 180 : ui32, + config.output_W = 320 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_437", + OutputName = "Resize_437", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x180x320xbf16> loc(#loc298) + xten_nn.output %461 : tensor<1x32x180x320xbf16> loc(#loc298) + } -> tensor<1x32x180x320xbf16> loc(#loc298) + %452 = xten_nn.subgraph (%arg5 = %451: tensor<1x32x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + Axis = 1 : i32, + IfmOperands = [0 : index, 1 : index], + LayerName = "Concat_438", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Concat_438", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_438", + OutputName = "Concat_438", + axis = 1 : i32} : (tensor<1x32x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x35x180x320xbf16> loc(#loc299) + xten_nn.output %461 : tensor<1x35x180x320xbf16> loc(#loc299) + } -> tensor<1x35x180x320xbf16> loc(#loc299) + %453 = xten_nn.subgraph (%arg5 = %452: tensor<1x35x180x320xbf16>, %arg6 = %5: tensor<16x35x3x3xbf16>, %arg7 = %4: tensor<16xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x35x180x320xbf16>, %arg9 = %arg6: tensor<16x35x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x35x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x35xbf16> loc(#loc352) + %465 = tosa.transpose %arg8, %463 : (tensor<1x35x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x35xbf16> loc(#loc352) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_439", + PartOfOutputName = "Conv_439", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x35xbf16>, tensor<16x3x3x35xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc300) + %467 = tosa.clamp %466 { + LayerName = "Relu_440", + OutputName = "Relu_440", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc301) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc352) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc301) + } -> tensor<1x16x180x320xbf16> loc(#loc352) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc352) + } -> tensor<1x16x180x320xbf16> loc(#loc352) + %454 = xten_nn.subgraph (%arg5 = %453: tensor<1x16x180x320xbf16>, %arg6 = %3: tensor<16x16x3x3xbf16>, %arg7 = %2: tensor<16xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<16x16x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x16x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x16xbf16> loc(#loc353) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc353) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_441", + PartOfOutputName = "Conv_441", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<16x3x3x16xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc302) + %467 = tosa.clamp %466 { + LayerName = "Relu_442", + OutputName = "Relu_442", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc303) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc353) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc303) + } -> tensor<1x16x180x320xbf16> loc(#loc353) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc353) + } -> tensor<1x16x180x320xbf16> loc(#loc353) + %455 = xten_nn.subgraph (%arg5 = %454: tensor<1x16x180x320xbf16>, %arg6 = %1: tensor<4x16x1x1xbf16>, %arg7 = %0: tensor<4xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<4x16x1x1xbf16>, %arg10 = %arg7: tensor<4xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc304) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<4x16x1x1xbf16>) -> tensor<4x1x1x16xbf16> loc(#loc304) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc304) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_443", + PartOfOutputName = "Conv_443", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<4x1x1x16xbf16>, tensor<4xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc304) + %467 = tosa.transpose %466, %462 : (tensor<1x180x320x4xbf16>, tensor<4xi32>) -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %467 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %461 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + %456 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_444_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 4 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 3 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc305) + } -> tensor<1x1x180x320xbf16> loc(#loc305) + %457 = xten_nn.subgraph (%arg5 = %456: tensor<1x1x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_447", + OutputName = "Clip_447", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %462 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + %458 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Split_444_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.num_ifm_shim_ch = 2 : ui32, + config.num_ofm_shim_ch = 2 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc305) + } -> tensor<1x3x180x320xbf16> loc(#loc305) + %459 = xten_nn.subgraph (%arg5 = %458: tensor<1x3x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index, 1 : index], + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_445", OutputName = "Add_445"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + %460 = xten_nn.subgraph (%arg5 = %459: tensor<1x3x180x320xbf16>) attributes { + IfmOperands = [0 : index], + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_446", + OutputName = "Clip_446", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + return %449, %430, %410, %387, %460, %457 : tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16> loc(#loc308) + } loc(#loc308) +} loc(#loc) +#loc1 = loc("Div_2") +#loc2 = loc("Sub_431") +#loc3 = loc("Sub_411") +#loc4 = loc("Sub_385") +#loc5 = loc("Sub_359") +#loc6 = loc("Div_16") +#loc7 = loc("Sub_14") +#loc8 = loc("Initializer_398") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("Conv_17") +#loc14 = loc("Add_19") +#loc15 = loc("Clip_22") +#loc16 = loc("Div_24") +#loc17 = loc("Mul_25") +#loc18 = loc("Conv_26") +#loc19 = loc("Relu_27") +#loc20 = loc("Conv_28") +#loc21 = loc("Add_29") +#loc22 = loc("Conv_30") +#loc23 = loc("Relu_31") +#loc24 = loc("Conv_32") +#loc25 = loc("Relu_33") +#loc26 = loc("Conv_34") +#loc27 = loc("Conv_35") +#loc28 = loc("Relu_36") +#loc29 = loc("Conv_37") +#loc30 = loc("Relu_38") +#loc31 = loc("Conv_39") +#loc32 = loc("Add_40") +#loc33 = loc("Conv_41") +#loc34 = loc("Relu_42") +#loc35 = loc("Conv_43") +#loc36 = loc("Relu_44") +#loc37 = loc("GlobalAveragePool_45") +#loc38 = loc("Conv_46") +#loc39 = loc("Relu_47") +#loc40 = loc("Conv_48") +#loc41 = loc("Add_50") +#loc42 = loc("Clip_53") +#loc43 = loc("Div_55") +#loc44 = loc("Mul_56") +#loc45 = loc("Conv_57") +#loc46 = loc("Conv_58") +#loc47 = loc("Relu_59") +#loc48 = loc("Conv_60") +#loc49 = loc("Relu_61") +#loc50 = loc("GlobalAveragePool_62") +#loc51 = loc("Conv_63") +#loc52 = loc("Relu_64") +#loc53 = loc("Conv_65") +#loc54 = loc("Add_67") +#loc55 = loc("Clip_70") +#loc56 = loc("Div_72") +#loc57 = loc("Mul_73") +#loc58 = loc("Conv_74") +#loc59 = loc("Add_75") +#loc60 = loc("Conv_76") +#loc61 = loc("Relu_77") +#loc62 = loc("Conv_78") +#loc63 = loc("Relu_79") +#loc64 = loc("GlobalAveragePool_80") +#loc65 = loc("Conv_81") +#loc66 = loc("Relu_82") +#loc67 = loc("Conv_83") +#loc68 = loc("Add_85") +#loc69 = loc("Clip_88") +#loc70 = loc("Div_90") +#loc71 = loc("Mul_91") +#loc72 = loc("Conv_92") +#loc73 = loc("Add_93") +#loc74 = loc("Conv_94") +#loc75 = loc("Add_96") +#loc76 = loc("Clip_99") +#loc77 = loc("Div_101") +#loc78 = loc("Mul_102") +#loc79 = loc("Conv_103") +#loc80 = loc("Add_105") +#loc81 = loc("Clip_108") +#loc82 = loc("Div_110") +#loc83 = loc("Mul_111") +#loc84 = loc("Conv_112") +#loc85 = loc("Conv_113") +#loc86 = loc("Add_115") +#loc87 = loc("Clip_118") +#loc88 = loc("Div_120") +#loc89 = loc("Mul_121") +#loc90 = loc("Conv_122") +#loc91 = loc("Add_124") +#loc92 = loc("Clip_127") +#loc93 = loc("Div_129") +#loc94 = loc("Mul_130") +#loc95 = loc("Conv_131") +#loc96 = loc("Add_132") +#loc97 = loc("Conv_133") +#loc98 = loc("Add_135") +#loc99 = loc("Clip_138") +#loc100 = loc("Div_140") +#loc101 = loc("Mul_141") +#loc102 = loc("Conv_142") +#loc103 = loc("Add_144") +#loc104 = loc("Clip_147") +#loc105 = loc("Div_149") +#loc106 = loc("Mul_150") +#loc107 = loc("Conv_151") +#loc108 = loc("Add_152") +#loc109 = loc("Conv_153") +#loc110 = loc("Add_155") +#loc111 = loc("Clip_158") +#loc112 = loc("Div_160") +#loc113 = loc("Mul_161") +#loc114 = loc("Conv_162") +#loc115 = loc("Add_164") +#loc116 = loc("Clip_167") +#loc117 = loc("Div_169") +#loc118 = loc("Mul_170") +#loc119 = loc("Conv_171") +#loc120 = loc("Add_172") +#loc121 = loc("Conv_173") +#loc122 = loc("Add_175") +#loc123 = loc("Clip_178") +#loc124 = loc("Div_180") +#loc125 = loc("Mul_181") +#loc126 = loc("Conv_182") +#loc127 = loc("Add_184") +#loc128 = loc("Clip_187") +#loc129 = loc("Div_189") +#loc130 = loc("Mul_190") +#loc131 = loc("GlobalAveragePool_191") +#loc132 = loc("Conv_192") +#loc133 = loc("Relu_193") +#loc134 = loc("Conv_194") +#loc135 = loc("Add_196") +#loc136 = loc("Clip_199") +#loc137 = loc("Div_201") +#loc138 = loc("Mul_202") +#loc139 = loc("Conv_203") +#loc140 = loc("Conv_204") +#loc141 = loc("Add_206") +#loc142 = loc("Clip_209") +#loc143 = loc("Div_211") +#loc144 = loc("Mul_212") +#loc145 = loc("Conv_213") +#loc146 = loc("Add_215") +#loc147 = loc("Clip_218") +#loc148 = loc("Div_220") +#loc149 = loc("Mul_221") +#loc150 = loc("GlobalAveragePool_222") +#loc151 = loc("Conv_223") +#loc152 = loc("Relu_224") +#loc153 = loc("Conv_225") +#loc154 = loc("Add_227") +#loc155 = loc("Clip_230") +#loc156 = loc("Div_232") +#loc157 = loc("Mul_233") +#loc158 = loc("Conv_234") +#loc159 = loc("Add_235") +#loc160 = loc("Conv_236") +#loc161 = loc("Add_238") +#loc162 = loc("Clip_241") +#loc163 = loc("Div_243") +#loc164 = loc("Mul_244") +#loc165 = loc("Conv_245") +#loc166 = loc("Add_247") +#loc167 = loc("Clip_250") +#loc168 = loc("Div_252") +#loc169 = loc("Mul_253") +#loc170 = loc("GlobalAveragePool_254") +#loc171 = loc("Conv_255") +#loc172 = loc("Relu_256") +#loc173 = loc("Conv_257") +#loc174 = loc("Add_259") +#loc175 = loc("Clip_262") +#loc176 = loc("Div_264") +#loc177 = loc("Mul_265") +#loc178 = loc("Conv_266") +#loc179 = loc("Conv_267") +#loc180 = loc("Add_269") +#loc181 = loc("Clip_272") +#loc182 = loc("Div_274") +#loc183 = loc("Mul_275") +#loc184 = loc("Conv_276") +#loc185 = loc("Add_278") +#loc186 = loc("Clip_281") +#loc187 = loc("Div_283") +#loc188 = loc("Mul_284") +#loc189 = loc("GlobalAveragePool_285") +#loc190 = loc("Conv_286") +#loc191 = loc("Relu_287") +#loc192 = loc("Conv_288") +#loc193 = loc("Add_290") +#loc194 = loc("Clip_293") +#loc195 = loc("Div_295") +#loc196 = loc("Mul_296") +#loc197 = loc("Conv_297") +#loc198 = loc("Add_298") +#loc199 = loc("Conv_299") +#loc200 = loc("Add_301") +#loc201 = loc("Clip_304") +#loc202 = loc("Div_306") +#loc203 = loc("Mul_307") +#loc204 = loc("Conv_308") +#loc205 = loc("Add_310") +#loc206 = loc("Clip_313") +#loc207 = loc("Div_315") +#loc208 = loc("Mul_316") +#loc209 = loc("GlobalAveragePool_317") +#loc210 = loc("Conv_318") +#loc211 = loc("Relu_319") +#loc212 = loc("Conv_320") +#loc213 = loc("Add_322") +#loc214 = loc("Clip_325") +#loc215 = loc("Div_327") +#loc216 = loc("Mul_328") +#loc217 = loc("Conv_329") +#loc218 = loc("Add_330") +#loc219 = loc("Conv_331") +#loc220 = loc("Add_333") +#loc221 = loc("Clip_336") +#loc222 = loc("Div_338") +#loc223 = loc("Mul_339") +#loc224 = loc("GlobalAveragePool_342") +#loc225 = loc("Conv_343") +#loc226 = loc("Sigmoid_344") +#loc227 = loc("Mul_345") +#loc228 = loc("Conv_340") +#loc229 = loc("Relu_341") +#loc230 = loc("Split_349") +#loc231 = loc("Concat_350") +#loc232 = loc("Conv_351") +#loc233 = loc("Sigmoid_352") +#loc234 = loc("Split_353") +#loc235 = loc("Mul_354") +#loc236 = loc("Concat_355") +#loc237 = loc("Conv_356") +#loc238 = loc("Tanh_357") +#loc239 = loc("Mul_361") +#loc240 = loc("Mul_360") +#loc241 = loc("Add_362") +#loc242 = loc("Concat_363") +#loc243 = loc("Resize_365") +#loc244 = loc("Slice_371") +#loc245 = loc("AveragePool_347") +#loc246 = loc("AveragePool_348") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_380") +#loc256 = loc("Concat_381") +#loc257 = loc("Conv_382") +#loc258 = loc("Tanh_383") +#loc259 = loc("Mul_387") +#loc260 = loc("Mul_386") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_406") +#loc274 = loc("Concat_407") +#loc275 = loc("Conv_408") +#loc276 = loc("Tanh_409") +#loc277 = loc("Mul_413") +#loc278 = loc("Mul_412") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_426") +#loc291 = loc("Concat_427") +#loc292 = loc("Conv_428") +#loc293 = loc("Tanh_429") +#loc294 = loc("Mul_433") +#loc295 = loc("Mul_432") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_446") +#loc307 = loc("Clip_447") +#loc308 = loc(fused[#loc1, #loc2, #loc3, #loc4, #loc5, #loc6, #loc7, #loc8, #loc9, #loc10, #loc11, #loc12, #loc13, #loc14, #loc15, #loc16, #loc17, #loc18, #loc19, #loc20, #loc21, #loc22, #loc23, #loc24, #loc25, #loc26, #loc27, #loc28, #loc29, #loc30, #loc31, #loc32, #loc33, #loc34, #loc35, #loc36, #loc37, #loc38, #loc39, #loc40, #loc41, #loc42, #loc43, #loc44, #loc45, #loc46, #loc47, #loc48, #loc49, #loc50, #loc51, #loc52, #loc53, #loc54, #loc55, #loc56, #loc57, #loc58, #loc59, #loc60, #loc61, #loc62, #loc63, #loc64, #loc65, #loc66, #loc67, #loc68, #loc69, #loc70, #loc71, #loc72, #loc73, #loc74, #loc75, #loc76, #loc77, #loc78, #loc79, #loc80, #loc81, #loc82, #loc83, #loc84, #loc85, #loc86, #loc87, #loc88, #loc89, #loc90, #loc91, #loc92, #loc93, #loc94, #loc95, #loc96, #loc97, #loc98, #loc99, #loc100, #loc101, #loc102, #loc103, #loc104, #loc105, #loc106, #loc107, #loc108, #loc109, #loc110, #loc111, #loc112, #loc113, #loc114, #loc115, #loc116, #loc117, #loc118, #loc119, #loc120, #loc121, #loc122, #loc123, #loc124, #loc125, #loc126, #loc127, #loc128, #loc129, #loc130, #loc131, #loc132, #loc133, #loc134, #loc135, #loc136, #loc137, #loc138, #loc139, #loc140, #loc141, #loc142, #loc143, #loc144, #loc145, #loc146, #loc147, #loc148, #loc149, #loc150, #loc151, #loc152, #loc153, #loc154, #loc155, #loc156, #loc157, #loc158, #loc159, #loc160, #loc161, #loc162, #loc163, #loc164, #loc165, #loc166, #loc167, #loc168, #loc169, #loc170, #loc171, #loc172, #loc173, #loc174, #loc175, #loc176, #loc177, #loc178, #loc179, #loc180, #loc181, #loc182, #loc183, #loc184, #loc185, #loc186, #loc187, #loc188, #loc189, #loc190, #loc191, #loc192, #loc193, #loc194, #loc195, #loc196, #loc197, #loc198, #loc199, #loc200, #loc201, #loc202, #loc203, #loc204, #loc205, #loc206, #loc207, #loc208, #loc209, #loc210, #loc211, #loc212, #loc213, #loc214, #loc215, #loc216, #loc217, #loc218, #loc219, #loc220, #loc221, #loc222, #loc223, #loc224, #loc225, #loc226, #loc227, #loc228, #loc229, #loc230, #loc231, #loc232, #loc233, #loc234, #loc235, #loc236, #loc237, #loc238, #loc239, #loc240, #loc241, #loc242, #loc243, #loc244, #loc245, #loc246, #loc247, #loc248, #loc249, #loc250, #loc251, #loc252, #loc253, #loc254, #loc255, #loc256, #loc257, #loc258, #loc259, #loc260, #loc261, #loc262, #loc263, #loc264, #loc265, #loc266, #loc267, #loc268, #loc269, #loc270, #loc271, #loc272, #loc273, #loc274, #loc275, #loc276, #loc277, #loc278, #loc279, #loc280, #loc281, #loc282, #loc283, #loc284, #loc285, #loc286, #loc287, #loc288, #loc289, #loc290, #loc291, #loc292, #loc293, #loc294, #loc295, #loc296, #loc297, #loc298, #loc299, #loc300, #loc301, #loc302, #loc303, #loc304, #loc305, #loc306, #loc307]) +#loc309 = loc(fused[#loc7, #loc8]) +#loc310 = loc(fused[#loc11, #loc9, #loc12]) +#loc311 = loc(fused[#loc9, #loc12, #loc11]) +#loc312 = loc(fused[#loc18, #loc19]) +#loc313 = loc(fused[#loc20, #loc21]) +#loc314 = loc(fused[#loc22, #loc23]) +#loc315 = loc(fused[#loc24, #loc25]) +#loc316 = loc(fused[#loc27, #loc28]) +#loc317 = loc(fused[#loc29, #loc30]) +#loc318 = loc(fused[#loc31, #loc32]) +#loc319 = loc(fused[#loc33, #loc34]) +#loc320 = loc(fused[#loc35, #loc36]) +#loc321 = loc(fused[#loc38, #loc39]) +#loc322 = loc(fused[#loc46, #loc47]) +#loc323 = loc(fused[#loc48, #loc49]) +#loc324 = loc(fused[#loc51, #loc52]) +#loc325 = loc(fused[#loc58, #loc59]) +#loc326 = loc(fused[#loc60, #loc61]) +#loc327 = loc(fused[#loc62, #loc63]) +#loc328 = loc(fused[#loc65, #loc66]) +#loc329 = loc(fused[#loc72, #loc73]) +#loc330 = loc(fused[#loc95, #loc96]) +#loc331 = loc(fused[#loc107, #loc108]) +#loc332 = loc(fused[#loc119, #loc120]) +#loc333 = loc(fused[#loc130, #loc131]) +#loc334 = loc(fused[#loc132, #loc133]) +#loc335 = loc(fused[#loc149, #loc150]) +#loc336 = loc(fused[#loc151, #loc152]) +#loc337 = loc(fused[#loc158, #loc159]) +#loc338 = loc(fused[#loc169, #loc170]) +#loc339 = loc(fused[#loc171, #loc172]) +#loc340 = loc(fused[#loc188, #loc189]) +#loc341 = loc(fused[#loc190, #loc191]) +#loc342 = loc(fused[#loc197, #loc198]) +#loc343 = loc(fused[#loc208, #loc209]) +#loc344 = loc(fused[#loc210, #loc211]) +#loc345 = loc(fused[#loc217, #loc218]) +#loc346 = loc(fused[#loc223, #loc224]) +#loc347 = loc(fused[#loc228, #loc229, #loc227]) +#loc348 = loc(fused[#loc228, #loc229]) +#loc349 = loc(fused[#loc248, #loc249]) +#loc350 = loc(fused[#loc266, #loc267]) +#loc351 = loc(fused[#loc283, #loc284]) +#loc352 = loc(fused[#loc300, #loc301]) +#loc353 = loc(fused[#loc302, #loc303]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/tensor_dims.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/tensor_dims.json new file mode 100644 index 0000000000000000000000000000000000000000..628b3bd20767c40fcdaad39a69c4001527983a71 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/tensor_dims.json @@ -0,0 +1,32071 @@ +{ + "__VERSION__": 1, + "Div_2": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 20, + "ifmsv_size": 1920, + "ifmsv_width": 4, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.00390625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 180, + 320, + 4 + ], + "dims": [ + 184, + 320, + 4 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 180, + 320, + 4 + ], + "dims": [ + 192, + 320, + 4 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Input.0", + "type": "external" + } + ], + "layer_order": 1 + }, + "Slice_7": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "W", + "dim_c": 184, + "dim_h": 320, + "dim_w": 4, + "dtype": "bfloat16", + "end": 3, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 180, + 320, + 4 + ], + "dims": [ + 184, + 320, + 4 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 180, + 320, + 3 + ], + "dims": [ + 184, + 320, + 3 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Div_2", + "type": "internal" + } + ], + "layer_order": 2 + }, + "Generated-#0": { + "BufferPadAdf": { + "aie_arch": "aie2p", + "dim_0": 320, + "dim_0_padded": 320, + "dim_1": 23, + "dim_1_padded": 23, + "dim_2": 3, + "dim_2_padded": 8, + "dim_3": 8, + "dim_3_padded": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 180, + 320, + 3 + ], + "dims": [ + 184, + 320, + 3 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 180, + 320, + 3 + ], + "dims": [ + 184, + 320, + 8 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Slice_7", + "type": "internal" + } + ], + "layer_order": 3 + }, + "Generated-#2": { + "Transpose4dAdf": { + "aie_arch": "aie2p", + "dim_0": 320, + "dim_1": 23, + "dim_2": 8, + "dim_3": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "perm": 10 + }, + "dims_and_bounds": [ + { + "bounds": [ + 180, + 320, + 3 + ], + "dims": [ + 184, + 320, + 8 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 184, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#0", + "type": "internal" + } + ], + "layer_order": 4 + }, + "Generated-#4": { + "BufferUnpadAdf": { + "aie_arch": "aie2p", + "dim_0": 184, + "dim_0_unpadded": 180, + "dim_1": 1, + "dim_1_unpadded": 1, + "dim_2": 320, + "dim_2_unpadded": 320, + "dim_3": 8, + "dim_3_unpadded": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 184, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#2", + "type": "internal" + } + ], + "layer_order": 5 + }, + "Sub_14": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 5, + "ifmsv_size": 640, + "ifmsv_width": 16, + "num_elems": 57600, + "num_kernel_iters": 180, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "constants_reader": { + "Const.0": { + "bytes_to_skip": 20764800, + "channel_stride": 8, + "depth": 1, + "dtype": "bfloat16", + "height": 180, + "nibbles": 4, + "num_reads": 86400, + "rank": 4, + "reads_per_line": 3, + "width": 320 + } + }, + "dims_and_bounds": [ + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#4", + "type": "internal" + }, + { + "name": "Const.0", + "type": "constant" + } + ], + "layer_order": 6 + }, + "Div_16": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 5, + "ifmsv_size": 640, + "ifmsv_width": 16, + "num_elems": 57600, + "num_kernel_iters": 180, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "constants_reader": { + "Const.1": { + "bytes_to_skip": 20764800, + "channel_stride": 8, + "depth": 1, + "dtype": "bfloat16", + "height": 180, + "nibbles": 4, + "num_reads": 86400, + "rank": 4, + "reads_per_line": 3, + "width": 320 + } + }, + "dims_and_bounds": [ + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sub_14", + "type": "internal" + }, + { + "name": "Const.1", + "type": "constant" + } + ], + "layer_order": 7 + }, + "Conv_17": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 8, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 180, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_height": 10, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 66, + "ifm_width": 320, + "ifm_x_iter": 5, + "ifm_y_iter": 6, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 6, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 16, + "ofm_depth_iter": 1, + "ofm_height": 90, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 8, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 160, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 1, + "pad_right": 0, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 10, + "padding_sv_width": 66, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 2, + "stride_height": 2, + "stride_log2.stride_log2_height": 1, + "stride_log2.stride_log2_width": 1, + "stride_w": 2, + "stride_width": 2, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Div_16", + "type": "internal" + } + ], + "layer_order": 8, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 1, + "ofm_depth_padded": 32, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 0, + "wts_no_of_reads": 864 + } + }, + "Add_19": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 23, + "ifmsv_size": 1472, + "ifmsv_width": 8, + "num_kernel_iters": 20, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 32, + 92, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_17", + "type": "internal" + } + ], + "layer_order": 9 + }, + "Clip_22": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 12, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 16, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_19", + "type": "internal" + } + ], + "layer_order": 10 + }, + "Div_24": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 23, + "ifmsv_size": 1472, + "ifmsv_width": 8, + "num_kernel_iters": 20, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 32, + 92, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_22", + "type": "internal" + } + ], + "layer_order": 11 + }, + "Mul_25": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 8, + "ifmsv_size": 1024, + "ifmsv_width": 16, + "num_elems": 14400, + "num_kernel_iters": 30, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_17", + "type": "internal" + }, + { + "name": "Div_24", + "type": "internal" + } + ], + "layer_order": 12 + }, + "Conv_26": { + "Conv2DBf16": { + "act": 1, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 16, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 90, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 8, + "ifm_sv_height.size": 8, + "ifm_sv_width": 18, + "ifm_sv_width.size": 18, + "ifm_width": 160, + "ifm_x_iter": 10, + "ifm_y_iter": 4, + "ifmsv_size": 1280, + "kernel_height": 3, + "kernel_width": 3, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 32, + "ofm_depth_iter": 1, + "ofm_height": 90, + "ofm_len": 40, + "ofm_sv_depth": 8, + "ofm_sv_height": 6, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 17, + "ofm_width": 159, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 32, + 96, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_25", + "type": "internal" + } + ], + "layer_order": 13, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 4, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 1, + "ofm_depth_padded": 32, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 10944, + "wts_no_of_reads": 144 + } + }, + "Conv_28": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 2, + "ifmsv_size": 512, + "ifmsv_width": 32, + "num_elems": 14400, + "num_kernel_iters": 60, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 11, + "ifm_depth": 16, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 90, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 2, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 160, + "ifm_x_iter": 5, + "ifm_y_iter": 12, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 12, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 16, + "ofm_depth_iter": 1, + "ofm_height": 90, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 8, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 160, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 2, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_26", + "type": "internal" + }, + { + "name": "Mul_25", + "type": "internal" + } + ], + "layer_order": 14, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 1, + "ofm_depth_padded": 32, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 13248, + "wts_no_of_reads": 768 + } + }, + "Conv_30": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 16, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 90, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 9, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 160, + "ifm_x_iter": 5, + "ifm_y_iter": 10, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 10, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 64, + "ofm_depth_iter": 1, + "ofm_height": 90, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 9, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 160, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 9, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 90, + 160 + ], + "dims": [ + 64, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_28", + "type": "internal" + } + ], + "layer_order": 15, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 23040, + "wts_no_of_reads": 1536 + } + }, + "Conv_32": { + "Conv2DBf16": { + "act": 1, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 64, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 160, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 6, + "ifm_sv_height.size": 6, + "ifm_sv_width": 26, + "ifm_sv_width.size": 26, + "ifm_width": 90, + "ifm_x_iter": 7, + "ifm_y_iter": 6, + "ifmsv_size": 1344, + "kernel_height": 3, + "kernel_width": 3, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 64, + "ofm_depth_iter": 2, + "ofm_height": 80, + "ofm_len": 84, + "ofm_sv_depth": 8, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 13, + "ofm_width": 44, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 1, + "pad_right": 0, + "pad_top": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 2, + "stride_height": 2, + "stride_width": 2, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 90, + 160 + ], + "dims": [ + 64, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_30", + "type": "internal" + } + ], + "layer_order": 16, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 4, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 2, + "ofm_depth_padded": 64, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 42624, + "wts_no_of_reads": 144 + } + }, + "Generated-#60": { + "BufferUnpadAdf": { + "aie_arch": "aie2p", + "dim_0": 48, + "dim_0_unpadded": 45, + "dim_1": 8, + "dim_1_unpadded": 8, + "dim_2": 80, + "dim_2_unpadded": 80, + "dim_3": 8, + "dim_3_unpadded": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 45, + 80 + ], + "dims": [ + 64, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_32", + "type": "internal" + } + ], + "layer_order": 17 + }, + "Conv_34": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 64, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 45, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 4, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 16, + "ifm_width": 80, + "ifm_x_iter": 5, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 24, + "ofm_depth_iter": 1, + "ofm_height": 45, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 80, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 4, + "padding_sv_width": 16, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 45, + 80 + ], + "dims": [ + 64, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 24, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#60", + "type": "internal" + } + ], + "layer_order": 18, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 47232, + "wts_no_of_reads": 1536 + } + }, + "Conv_35": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 24, + "ifm_depth_concat_extend": 40, + "ifm_depth_iter": 1, + "ifm_height": 45, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 2, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 80, + "ifm_x_iter": 3, + "ifm_y_iter": 6, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 6, + "num_ifm_width_iters": 3, + "num_ofm_depth_iters": 1, + "ofm_depth": 72, + "ofm_depth_iter": 1, + "ofm_height": 45, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 24, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 80, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 2, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 24, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 45, + 80 + ], + "dims": [ + 96, + 48, + 96 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_34", + "type": "internal" + } + ], + "layer_order": 19, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 432, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 3, + "ofm_depth_iter": 1, + "ofm_depth_padded": 96, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 66816, + "wts_no_of_reads": 2304 + } + }, + "Conv_37": { + "Conv2DBf16": { + "act": 1, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 72, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 80, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 8, + "ifm_sv_height.size": 8, + "ifm_sv_width": 18, + "ifm_sv_width.size": 18, + "ifm_width": 45, + "ifm_x_iter": 5, + "ifm_y_iter": 2, + "ifmsv_size": 1280, + "kernel_height": 3, + "kernel_width": 3, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 72, + "ofm_depth_iter": 3, + "ofm_height": 80, + "ofm_len": 30, + "ofm_sv_depth": 8, + "ofm_sv_height": 6, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 17, + "ofm_width": 44, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 45, + 80 + ], + "dims": [ + 72, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 45, + 80 + ], + "dims": [ + 72, + 46, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_35", + "type": "internal" + } + ], + "layer_order": 20, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 4, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 3, + "ofm_depth_padded": 96, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 96192, + "wts_no_of_reads": 144 + } + }, + "Generated-#62": { + "BufferUnpadAdf": { + "aie_arch": "aie2p", + "dim_0": 46, + "dim_0_unpadded": 45, + "dim_1": 9, + "dim_1_unpadded": 9, + "dim_2": 80, + "dim_2_unpadded": 80, + "dim_3": 8, + "dim_3_unpadded": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 45, + 80 + ], + "dims": [ + 72, + 46, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 45, + 80 + ], + "dims": [ + 72, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_37", + "type": "internal" + } + ], + "layer_order": 21 + }, + "Conv_39": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 4, + "ifmsv_size": 1024, + "ifmsv_width": 16, + "num_elems": 3600, + "num_kernel_iters": 15, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 11, + "ifm_depth": 72, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 45, + "ifm_sign": 0, + "ifm_sv_depth": 72, + "ifm_sv_height": 4, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 16, + "ifm_width": 80, + "ifm_x_iter": 5, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 24, + "ofm_depth_iter": 1, + "ofm_height": 45, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 80, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 4, + "padding_sv_width": 16, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 45, + 80 + ], + "dims": [ + 72, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 24, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 24, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#62", + "type": "internal" + }, + { + "name": "Conv_34", + "type": "internal" + } + ], + "layer_order": 22, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 72, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 103104, + "wts_no_of_reads": 1728 + } + }, + "Conv_41": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 24, + "ifm_depth_concat_extend": 40, + "ifm_depth_iter": 1, + "ifm_height": 45, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 2, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 80, + "ifm_x_iter": 3, + "ifm_y_iter": 6, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 6, + "num_ifm_width_iters": 3, + "num_ofm_depth_iters": 1, + "ofm_depth": 72, + "ofm_depth_iter": 1, + "ofm_height": 45, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 24, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 80, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 2, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 24, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 45, + 80 + ], + "dims": [ + 96, + 48, + 96 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_39", + "type": "internal" + } + ], + "layer_order": 23, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 432, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 3, + "ofm_depth_iter": 1, + "ofm_depth_padded": 96, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 124992, + "wts_no_of_reads": 2304 + } + }, + "Conv_43": { + "Conv2DBf16": { + "act": 1, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 96, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 48, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 8, + "ifm_sv_height.size": 8, + "ifm_sv_width": 20, + "ifm_sv_width.size": 20, + "ifm_width": 96, + "ifm_x_iter": 5, + "ifm_y_iter": 3, + "ifmsv_size": 1280, + "kernel_height": 5, + "kernel_width": 5, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 96, + "ofm_depth_iter": 3, + "ofm_height": 24, + "ofm_len": 45, + "ofm_sv_depth": 8, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 7, + "ofm_width": 46, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 2, + "pad_right": 2, + "pad_top": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 2, + "stride_height": 2, + "stride_width": 2, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 45, + 80 + ], + "dims": [ + 96, + 48, + 96 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 23, + 40 + ], + "dims": [ + 96, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_41", + "type": "internal" + } + ], + "layer_order": 24, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 5, + "kernel_width": 8, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 3, + "ofm_depth_padded": 96, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 154368, + "wts_no_of_reads": 480 + } + }, + "Generated-#6": { + "Transpose4dAdf": { + "aie_arch": "aie2p", + "dim_0": 23, + "dim_1": 9, + "dim_2": 40, + "dim_3": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "perm": 6 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 23, + 40 + ], + "dims": [ + 72, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 1, + 920 + ], + "dims": [ + 72, + 1, + 920 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_43", + "type": "internal" + } + ], + "layer_order": 25 + }, + "Generated-#8": { + "ReduceMeanC8Bf16": { + "L3_depth": 72, + "L3_height": 1, + "L3_width": 920, + "aie_arch": "aie2p", + "compiler": "chess", + "depth_iter": 1, + "full_channel": 72, + "full_height": 1, + "full_width": 920, + "height_iter": 1, + "loop_num_channel": 3, + "loop_num_height": 1, + "loop_num_width": 15, + "num_iters": 15, + "pad_value": 0, + "reduce_axis": 2, + "reduce_dim": "W", + "scale": 0.0010833740234375, + "scale_shift": 0, + "sv_channel": 32, + "sv_height": 1, + "sv_width": 64, + "width_iter": 15 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 1, + 920 + ], + "dims": [ + 72, + 1, + 920 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 1, + 1 + ], + "dims": [ + 128, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#6", + "type": "internal" + } + ], + "layer_order": 26 + }, + "Conv_46": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 72, + "ifm_depth_concat_extend": 56, + "ifm_depth_iter": 1, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 72, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 16, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 24, + "ofm_depth_iter": 1, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 16, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 1, + 1 + ], + "dims": [ + 128, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 24, + 1, + 1 + ], + "dims": [ + 64, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#8", + "type": "internal" + } + ], + "layer_order": 27, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 72, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 173376, + "wts_no_of_reads": 1728 + } + }, + "Conv_48": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 24, + "ifm_depth_concat_extend": 40, + "ifm_depth_iter": 1, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 72, + "ofm_depth_iter": 1, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 24, + 1, + 1 + ], + "dims": [ + 64, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 1, + 1 + ], + "dims": [ + 128, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_46", + "type": "internal" + } + ], + "layer_order": 28, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 1, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 195264, + "wts_no_of_reads": 3072 + } + }, + "Add_50": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 16, + "ifmsv_size": 384, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 1, + 1 + ], + "dims": [ + 128, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 1, + 1 + ], + "dims": [ + 96, + 64, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_48", + "type": "internal" + } + ], + "layer_order": 29 + }, + "Clip_53": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 8, + "ifmsv_size": 256, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 1, + 1 + ], + "dims": [ + 96, + 64, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 1, + 1 + ], + "dims": [ + 128, + 32, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_50", + "type": "internal" + } + ], + "layer_order": 30 + }, + "Div_55": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 16, + "ifmsv_size": 512, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 1, + 1 + ], + "dims": [ + 128, + 32, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 1, + 1 + ], + "dims": [ + 128, + 64, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_53", + "type": "internal" + } + ], + "layer_order": 31 + }, + "Generated-#10": { + "TileAdf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "i_dim_c": 72, + "i_dim_h": 1, + "i_dim_n": 1, + "i_dim_w": 1, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "rep_dim_c": 1, + "rep_dim_h": 23, + "rep_dim_w": 40 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 1, + 1 + ], + "dims": [ + 72, + 1, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 23, + 40 + ], + "dims": [ + 72, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Div_55", + "type": "internal" + } + ], + "layer_order": 32 + }, + "Mul_56": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 6, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 920, + "num_kernel_iters": 6, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 23, + 40 + ], + "dims": [ + 72, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 23, + 40 + ], + "dims": [ + 96, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 72, + 23, + 40 + ], + "dims": [ + 96, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#10", + "type": "internal" + }, + { + "name": "Conv_43", + "type": "internal" + } + ], + "layer_order": 33 + }, + "Conv_57": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 72, + "ifm_depth_concat_extend": 24, + "ifm_depth_iter": 1, + "ifm_height": 23, + "ifm_sign": 0, + "ifm_sv_depth": 72, + "ifm_sv_height": 6, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 40, + "ifm_x_iter": 5, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 40, + "ofm_depth_iter": 1, + "ofm_height": 23, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 6, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 40, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 6, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 72, + 23, + 40 + ], + "dims": [ + 96, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_56", + "type": "internal" + } + ], + "layer_order": 34, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 72, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 234432, + "wts_no_of_reads": 1728 + } + }, + "Conv_58": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 40, + "ifm_depth_concat_extend": 24, + "ifm_depth_iter": 1, + "ifm_height": 23, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 2, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 40, + "ifm_x_iter": 2, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 2, + "num_ofm_depth_iters": 1, + "ofm_depth": 120, + "ofm_depth_iter": 1, + "ofm_height": 23, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 40, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 2, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_57", + "type": "internal" + } + ], + "layer_order": 35, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 1, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 256320, + "wts_no_of_reads": 3072 + } + }, + "Conv_60": { + "Conv2DBf16": { + "act": 1, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 128, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 24, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 10, + "ifm_sv_height.size": 10, + "ifm_sv_width": 12, + "ifm_sv_width.size": 12, + "ifm_width": 64, + "ifm_x_iter": 5, + "ifm_y_iter": 1, + "ifmsv_size": 960, + "kernel_height": 5, + "kernel_width": 5, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 128, + "ofm_depth_iter": 4, + "ofm_height": 24, + "ofm_len": 20, + "ofm_sv_depth": 8, + "ofm_sv_height": 6, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 5, + "ofm_width": 61, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 2, + "pad_left": 2, + "pad_right": 2, + "pad_top": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_58", + "type": "internal" + } + ], + "layer_order": 36, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 5, + "kernel_width": 8, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 4, + "ofm_depth_padded": 128, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 295488, + "wts_no_of_reads": 480 + } + }, + "Generated-#12": { + "Transpose4dAdf": { + "aie_arch": "aie2p", + "dim_0": 23, + "dim_1": 15, + "dim_2": 40, + "dim_3": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "perm": 6 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 120, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 920 + ], + "dims": [ + 120, + 1, + 920 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_60", + "type": "internal" + } + ], + "layer_order": 37 + }, + "Generated-#14": { + "ReduceMeanC8Bf16": { + "L3_depth": 120, + "L3_height": 1, + "L3_width": 920, + "aie_arch": "aie2p", + "compiler": "chess", + "depth_iter": 1, + "full_channel": 120, + "full_height": 1, + "full_width": 920, + "height_iter": 1, + "loop_num_channel": 4, + "loop_num_height": 1, + "loop_num_width": 15, + "num_iters": 15, + "pad_value": 0, + "reduce_axis": 2, + "reduce_dim": "W", + "scale": 0.0010833740234375, + "scale_shift": 0, + "sv_channel": 32, + "sv_height": 1, + "sv_width": 64, + "width_iter": 15 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 920 + ], + "dims": [ + 120, + 1, + 920 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#12", + "type": "internal" + } + ], + "layer_order": 38 + }, + "Conv_63": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 120, + "ifm_depth_concat_extend": 8, + "ifm_depth_iter": 1, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 120, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 16, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 32, + "ofm_depth_iter": 1, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 16, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 32, + 1, + 1 + ], + "dims": [ + 64, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#14", + "type": "internal" + } + ], + "layer_order": 39, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 120, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 320832, + "wts_no_of_reads": 2880 + } + }, + "Conv_65": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 32, + "ifm_depth_concat_extend": 32, + "ifm_depth_iter": 1, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 120, + "ofm_depth_iter": 1, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 1, + 1 + ], + "dims": [ + 64, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_63", + "type": "internal" + } + ], + "layer_order": 40, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 1, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 356544, + "wts_no_of_reads": 3072 + } + }, + "Add_67": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 12, + "ifmsv_size": 384, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 48, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_65", + "type": "internal" + } + ], + "layer_order": 41 + }, + "Clip_70": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 8, + "ifmsv_size": 256, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 48, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 32, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_67", + "type": "internal" + } + ], + "layer_order": 42 + }, + "Div_72": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 16, + "ifmsv_size": 512, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 32, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 64, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_70", + "type": "internal" + } + ], + "layer_order": 43 + }, + "Generated-#16": { + "TileAdf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "i_dim_c": 120, + "i_dim_h": 1, + "i_dim_n": 1, + "i_dim_w": 1, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "rep_dim_c": 1, + "rep_dim_h": 23, + "rep_dim_w": 40 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 120, + 1, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 120, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Div_72", + "type": "internal" + } + ], + "layer_order": 44 + }, + "Mul_73": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 920, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 120, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#16", + "type": "internal" + }, + { + "name": "Conv_60", + "type": "internal" + } + ], + "layer_order": 45 + }, + "Conv_74": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 1, + "ifmsv_size": 640, + "ifmsv_width": 40, + "num_elems": 920, + "num_kernel_iters": 6, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 40, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 11, + "ifm_depth": 120, + "ifm_depth_concat_extend": 8, + "ifm_depth_iter": 2, + "ifm_height": 23, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 40, + "ifm_width": 40, + "ifm_x_iter": 1, + "ifm_y_iter": 6, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 6, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 40, + "ofm_depth_iter": 1, + "ofm_height": 23, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 40, + "ofm_width": 40, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 40, + "psum_buff_offset_scaled": 5, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_73", + "type": "internal" + }, + { + "name": "Conv_57", + "type": "internal" + } + ], + "layer_order": 46, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 128, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 395712, + "wts_no_of_reads": 1536 + } + }, + "Conv_76": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 40, + "ifm_depth_concat_extend": 24, + "ifm_depth_iter": 1, + "ifm_height": 23, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 2, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 40, + "ifm_x_iter": 2, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 2, + "num_ofm_depth_iters": 1, + "ofm_depth": 120, + "ofm_depth_iter": 1, + "ofm_height": 23, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 40, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 2, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_74", + "type": "internal" + } + ], + "layer_order": 47, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 1, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 434880, + "wts_no_of_reads": 3072 + } + }, + "Conv_78": { + "Conv2DBf16": { + "act": 1, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 128, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 24, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 10, + "ifm_sv_height.size": 10, + "ifm_sv_width": 12, + "ifm_sv_width.size": 12, + "ifm_width": 64, + "ifm_x_iter": 5, + "ifm_y_iter": 1, + "ifmsv_size": 960, + "kernel_height": 5, + "kernel_width": 5, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 128, + "ofm_depth_iter": 4, + "ofm_height": 24, + "ofm_len": 20, + "ofm_sv_depth": 8, + "ofm_sv_height": 6, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 5, + "ofm_width": 61, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 2, + "pad_left": 2, + "pad_right": 2, + "pad_top": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_76", + "type": "internal" + } + ], + "layer_order": 48, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 5, + "kernel_width": 8, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 4, + "ofm_depth_padded": 128, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 474048, + "wts_no_of_reads": 480 + } + }, + "Generated-#18": { + "Transpose4dAdf": { + "aie_arch": "aie2p", + "dim_0": 23, + "dim_1": 15, + "dim_2": 40, + "dim_3": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "perm": 6 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 120, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 920 + ], + "dims": [ + 120, + 1, + 920 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_78", + "type": "internal" + } + ], + "layer_order": 49 + }, + "Generated-#20": { + "ReduceMeanC8Bf16": { + "L3_depth": 120, + "L3_height": 1, + "L3_width": 920, + "aie_arch": "aie2p", + "compiler": "chess", + "depth_iter": 1, + "full_channel": 120, + "full_height": 1, + "full_width": 920, + "height_iter": 1, + "loop_num_channel": 4, + "loop_num_height": 1, + "loop_num_width": 15, + "num_iters": 15, + "pad_value": 0, + "reduce_axis": 2, + "reduce_dim": "W", + "scale": 0.0010833740234375, + "scale_shift": 0, + "sv_channel": 32, + "sv_height": 1, + "sv_width": 64, + "width_iter": 15 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 920 + ], + "dims": [ + 120, + 1, + 920 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#18", + "type": "internal" + } + ], + "layer_order": 50 + }, + "Conv_81": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 120, + "ifm_depth_concat_extend": 8, + "ifm_depth_iter": 1, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 120, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 16, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 32, + "ofm_depth_iter": 1, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 16, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 32, + 1, + 1 + ], + "dims": [ + 64, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#20", + "type": "internal" + } + ], + "layer_order": 51, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 120, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 499392, + "wts_no_of_reads": 2880 + } + }, + "Conv_83": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 32, + "ifm_depth_concat_extend": 32, + "ifm_depth_iter": 1, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 120, + "ofm_depth_iter": 1, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 1, + 1 + ], + "dims": [ + 64, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_81", + "type": "internal" + } + ], + "layer_order": 52, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 1, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 535104, + "wts_no_of_reads": 3072 + } + }, + "Add_85": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 12, + "ifmsv_size": 384, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 48, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_83", + "type": "internal" + } + ], + "layer_order": 53 + }, + "Clip_88": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 8, + "ifmsv_size": 256, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 48, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 32, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_85", + "type": "internal" + } + ], + "layer_order": 54 + }, + "Div_90": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 16, + "ifmsv_size": 512, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 32, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 64, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_88", + "type": "internal" + } + ], + "layer_order": 55 + }, + "Generated-#22": { + "TileAdf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "i_dim_c": 120, + "i_dim_h": 1, + "i_dim_n": 1, + "i_dim_w": 1, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "rep_dim_c": 1, + "rep_dim_h": 23, + "rep_dim_w": 40 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 120, + 1, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 120, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Div_90", + "type": "internal" + } + ], + "layer_order": 56 + }, + "Mul_91": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 920, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 120, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#22", + "type": "internal" + }, + { + "name": "Conv_78", + "type": "internal" + } + ], + "layer_order": 57 + }, + "Conv_92": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 1, + "ifmsv_size": 640, + "ifmsv_width": 40, + "num_elems": 920, + "num_kernel_iters": 6, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 40, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 11, + "ifm_depth": 120, + "ifm_depth_concat_extend": 8, + "ifm_depth_iter": 2, + "ifm_height": 23, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 40, + "ifm_width": 40, + "ifm_x_iter": 1, + "ifm_y_iter": 6, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 6, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 40, + "ofm_depth_iter": 1, + "ofm_height": 23, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 40, + "ofm_width": 40, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 40, + "psum_buff_offset_scaled": 5, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 23, + 40 + ], + "dims": [ + 128, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_91", + "type": "internal" + }, + { + "name": "Conv_74", + "type": "internal" + } + ], + "layer_order": 58, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 128, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 574272, + "wts_no_of_reads": 1536 + } + }, + "Conv_94": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 40, + "ifm_depth_concat_extend": 24, + "ifm_depth_iter": 1, + "ifm_height": 23, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 2, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 40, + "ifm_x_iter": 2, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 2, + "num_ofm_depth_iters": 2, + "ofm_depth": 240, + "ofm_depth_iter": 2, + "ofm_height": 23, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 40, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 2, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 256, + 24, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_92", + "type": "internal" + } + ], + "layer_order": 59, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 2, + "ofm_depth_padded": 256, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 613440, + "wts_no_of_reads": 3072 + } + }, + "Add_96": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 256, + 24, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 256, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_94", + "type": "internal" + } + ], + "layer_order": 60 + }, + "Clip_99": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 256, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 256, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_96", + "type": "internal" + } + ], + "layer_order": 61 + }, + "Div_101": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 256, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 256, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_99", + "type": "internal" + } + ], + "layer_order": 62 + }, + "Mul_102": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 920, + "num_kernel_iters": 16, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 240, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 240, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 240, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_94", + "type": "internal" + }, + { + "name": "Div_101", + "type": "internal" + } + ], + "layer_order": 63 + }, + "Generated-#64": { + "BufferUnpadAdf": { + "aie_arch": "aie2p", + "dim_0": 24, + "dim_0_unpadded": 23, + "dim_1": 30, + "dim_1_unpadded": 30, + "dim_2": 40, + "dim_2_unpadded": 40, + "dim_3": 8, + "dim_3_unpadded": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 240, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 240, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_102", + "type": "internal" + } + ], + "layer_order": 64 + }, + "Conv_103": { + "Conv2DBf16": { + "act": 0, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 240, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 23, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 10, + "ifm_sv_height.size": 10, + "ifm_sv_width": 10, + "ifm_sv_width.size": 10, + "ifm_width": 40, + "ifm_x_iter": 5, + "ifm_y_iter": 1, + "ifmsv_size": 960, + "kernel_height": 3, + "kernel_width": 3, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 256, + "ofm_depth_iter": 8, + "ofm_height": 12, + "ofm_len": 40, + "ofm_sv_depth": 8, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 5, + "ofm_width": 19, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 2, + "stride_height": 2, + "stride_width": 2, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 23, + 40 + ], + "dims": [ + 240, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 16, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#64", + "type": "internal" + } + ], + "layer_order": 65, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 4, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 8, + "ofm_depth_padded": 256, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 691776, + "wts_no_of_reads": 144 + } + }, + "Add_105": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 2, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 16, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_103", + "type": "internal" + } + ], + "layer_order": 66 + }, + "Clip_108": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 2, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_105", + "type": "internal" + } + ], + "layer_order": 67 + }, + "Div_110": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 2, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_108", + "type": "internal" + } + ], + "layer_order": 68 + }, + "Mul_111": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 16, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_103", + "type": "internal" + }, + { + "name": "Div_110", + "type": "internal" + } + ], + "layer_order": 69 + }, + "Conv_112": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 240, + "ifm_depth_concat_extend": 16, + "ifm_depth_iter": 3, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 3, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 80, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 24, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_111", + "type": "internal" + } + ], + "layer_order": 70, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 432, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 3, + "ifm_depth_padded": 240, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 3, + "ofm_depth_iter": 1, + "ofm_depth_padded": 96, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 710208, + "wts_no_of_reads": 2880 + } + }, + "Conv_113": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 80, + "ifm_depth_concat_extend": 16, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 2, + "ofm_depth": 200, + "ofm_depth_iter": 2, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_112", + "type": "internal" + } + ], + "layer_order": 71, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 80, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 2, + "ofm_depth_padded": 256, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 819072, + "wts_no_of_reads": 3840 + } + }, + "Add_115": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 2, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_113", + "type": "internal" + } + ], + "layer_order": 72 + }, + "Clip_118": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 2, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_115", + "type": "internal" + } + ], + "layer_order": 73 + }, + "Div_120": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 2, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_118", + "type": "internal" + } + ], + "layer_order": 74 + }, + "Mul_121": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_113", + "type": "internal" + }, + { + "name": "Div_120", + "type": "internal" + } + ], + "layer_order": 75 + }, + "Conv_122": { + "Conv2DBf16": { + "act": 0, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 256, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 6, + "ifm_sv_height.size": 6, + "ifm_sv_width": 26, + "ifm_sv_width.size": 26, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "ifmsv_size": 1344, + "kernel_height": 3, + "kernel_width": 3, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 224, + "ofm_depth_iter": 7, + "ofm_height": 12, + "ofm_len": 7, + "ofm_sv_depth": 8, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 25, + "ofm_width": 19, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 224, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_121", + "type": "internal" + } + ], + "layer_order": 76, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 4, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 7, + "ofm_depth_padded": 224, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 915840, + "wts_no_of_reads": 144 + } + }, + "Add_124": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 2, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 224, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_122", + "type": "internal" + } + ], + "layer_order": 77 + }, + "Clip_127": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 2, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_124", + "type": "internal" + } + ], + "layer_order": 78 + }, + "Div_129": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 2, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_127", + "type": "internal" + } + ], + "layer_order": 79 + }, + "Mul_130": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 224, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_122", + "type": "internal" + }, + { + "name": "Div_129", + "type": "internal" + } + ], + "layer_order": 80 + }, + "Conv_131": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 1, + "ifmsv_size": 768, + "ifmsv_width": 32, + "num_elems": 240, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 11, + "ifm_depth": 200, + "ifm_depth_concat_extend": 56, + "ifm_depth_iter": 2, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 104, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 80, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 24, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 200, + 12, + 20 + ], + "dims": [ + 256, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_130", + "type": "internal" + }, + { + "name": "Conv_112", + "type": "internal" + } + ], + "layer_order": 81, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 432, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 208, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 3, + "ofm_depth_iter": 1, + "ofm_depth_padded": 96, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 931968, + "wts_no_of_reads": 3744 + } + }, + "Conv_133": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 80, + "ifm_depth_concat_extend": 16, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 3, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 20, + "ifm_x_iter": 3, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 3, + "num_ofm_depth_iters": 1, + "ofm_depth": 184, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 48, + "ofm_sv_height": 3, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 3, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_131", + "type": "internal" + } + ], + "layer_order": 82, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 864, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 80, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 6, + "ofm_depth_iter": 1, + "ofm_depth_padded": 192, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 1025280, + "wts_no_of_reads": 5760 + } + }, + "Add_135": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_133", + "type": "internal" + } + ], + "layer_order": 83 + }, + "Clip_138": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_135", + "type": "internal" + } + ], + "layer_order": 84 + }, + "Div_140": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_138", + "type": "internal" + } + ], + "layer_order": 85 + }, + "Mul_141": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_133", + "type": "internal" + }, + { + "name": "Div_140", + "type": "internal" + } + ], + "layer_order": 86 + }, + "Conv_142": { + "Conv2DBf16": { + "act": 0, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 192, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 6, + "ifm_sv_height.size": 6, + "ifm_sv_width": 26, + "ifm_sv_width.size": 26, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "ifmsv_size": 1344, + "kernel_height": 3, + "kernel_width": 3, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 192, + "ofm_depth_iter": 6, + "ofm_height": 12, + "ofm_len": 6, + "ofm_sv_depth": 8, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 25, + "ofm_width": 19, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_141", + "type": "internal" + } + ], + "layer_order": 87, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 4, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 6, + "ofm_depth_padded": 192, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 1097856, + "wts_no_of_reads": 144 + } + }, + "Add_144": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_142", + "type": "internal" + } + ], + "layer_order": 88 + }, + "Clip_147": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_144", + "type": "internal" + } + ], + "layer_order": 89 + }, + "Div_149": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_147", + "type": "internal" + } + ], + "layer_order": 90 + }, + "Mul_150": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_142", + "type": "internal" + }, + { + "name": "Div_149", + "type": "internal" + } + ], + "layer_order": 91 + }, + "Conv_151": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 1, + "ifmsv_size": 768, + "ifmsv_width": 32, + "num_elems": 240, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 11, + "ifm_depth": 184, + "ifm_depth_concat_extend": 8, + "ifm_depth_iter": 2, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 96, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 80, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 24, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_150", + "type": "internal" + }, + { + "name": "Conv_131", + "type": "internal" + } + ], + "layer_order": 92, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 432, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 192, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 3, + "ofm_depth_iter": 1, + "ofm_depth_padded": 96, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 1111680, + "wts_no_of_reads": 3456 + } + }, + "Conv_153": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 80, + "ifm_depth_concat_extend": 16, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 3, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 20, + "ifm_x_iter": 3, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 3, + "num_ofm_depth_iters": 1, + "ofm_depth": 184, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 48, + "ofm_sv_height": 3, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 3, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_151", + "type": "internal" + } + ], + "layer_order": 93, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 864, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 80, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 6, + "ofm_depth_iter": 1, + "ofm_depth_padded": 192, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 1198080, + "wts_no_of_reads": 5760 + } + }, + "Add_155": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_153", + "type": "internal" + } + ], + "layer_order": 94 + }, + "Clip_158": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_155", + "type": "internal" + } + ], + "layer_order": 95 + }, + "Div_160": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_158", + "type": "internal" + } + ], + "layer_order": 96 + }, + "Mul_161": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_153", + "type": "internal" + }, + { + "name": "Div_160", + "type": "internal" + } + ], + "layer_order": 97 + }, + "Conv_162": { + "Conv2DBf16": { + "act": 0, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 192, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 6, + "ifm_sv_height.size": 6, + "ifm_sv_width": 26, + "ifm_sv_width.size": 26, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "ifmsv_size": 1344, + "kernel_height": 3, + "kernel_width": 3, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 192, + "ofm_depth_iter": 6, + "ofm_height": 12, + "ofm_len": 6, + "ofm_sv_depth": 8, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 25, + "ofm_width": 19, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_161", + "type": "internal" + } + ], + "layer_order": 98, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 4, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 6, + "ofm_depth_padded": 192, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 1270656, + "wts_no_of_reads": 144 + } + }, + "Add_164": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_162", + "type": "internal" + } + ], + "layer_order": 99 + }, + "Clip_167": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_164", + "type": "internal" + } + ], + "layer_order": 100 + }, + "Div_169": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_167", + "type": "internal" + } + ], + "layer_order": 101 + }, + "Mul_170": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_162", + "type": "internal" + }, + { + "name": "Div_169", + "type": "internal" + } + ], + "layer_order": 102 + }, + "Conv_171": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 1, + "ifmsv_size": 768, + "ifmsv_width": 32, + "num_elems": 240, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 11, + "ifm_depth": 184, + "ifm_depth_concat_extend": 8, + "ifm_depth_iter": 2, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 96, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 80, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 24, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 184, + 12, + 20 + ], + "dims": [ + 192, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_170", + "type": "internal" + }, + { + "name": "Conv_151", + "type": "internal" + } + ], + "layer_order": 103, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 432, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 192, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 3, + "ofm_depth_iter": 1, + "ofm_depth_padded": 96, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 1284480, + "wts_no_of_reads": 3456 + } + }, + "Conv_173": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 80, + "ifm_depth_concat_extend": 16, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 3, + "ofm_depth": 480, + "ofm_depth_iter": 3, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 40, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 12, + 20 + ], + "dims": [ + 96, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 480, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_171", + "type": "internal" + } + ], + "layer_order": 104, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 720, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 80, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 5, + "ofm_depth_iter": 3, + "ofm_depth_padded": 480, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 1370880, + "wts_no_of_reads": 4800 + } + }, + "Add_175": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 480, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_173", + "type": "internal" + } + ], + "layer_order": 105 + }, + "Clip_178": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_175", + "type": "internal" + } + ], + "layer_order": 106 + }, + "Div_180": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_178", + "type": "internal" + } + ], + "layer_order": 107 + }, + "Mul_181": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 480, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_173", + "type": "internal" + }, + { + "name": "Div_180", + "type": "internal" + } + ], + "layer_order": 108 + }, + "Conv_182": { + "Conv2DBf16": { + "act": 0, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 512, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 6, + "ifm_sv_height.size": 6, + "ifm_sv_width": 26, + "ifm_sv_width.size": 26, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "ifmsv_size": 1344, + "kernel_height": 3, + "kernel_width": 3, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 480, + "ofm_depth_iter": 15, + "ofm_height": 12, + "ofm_len": 15, + "ofm_sv_depth": 8, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 25, + "ofm_width": 19, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 480, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_181", + "type": "internal" + } + ], + "layer_order": 109, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 4, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 15, + "ofm_depth_padded": 480, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 1552320, + "wts_no_of_reads": 144 + } + }, + "Add_184": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 480, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_182", + "type": "internal" + } + ], + "layer_order": 110 + }, + "Clip_187": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_184", + "type": "internal" + } + ], + "layer_order": 111 + }, + "Div_189": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_187", + "type": "internal" + } + ], + "layer_order": 112 + }, + "Mul_190": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 480, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_182", + "type": "internal" + }, + { + "name": "Div_189", + "type": "internal" + } + ], + "layer_order": 113 + }, + "Generated-#24": { + "Transpose4dAdf": { + "aie_arch": "aie2p", + "dim_0": 12, + "dim_1": 60, + "dim_2": 20, + "dim_3": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "perm": 6 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 480, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 1, + 240 + ], + "dims": [ + 480, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_190", + "type": "internal" + } + ], + "layer_order": 114 + }, + "Generated-#26": { + "ReduceMeanC8Bf16": { + "L3_depth": 480, + "L3_height": 1, + "L3_width": 240, + "aie_arch": "aie2p", + "compiler": "chess", + "depth_iter": 3, + "full_channel": 480, + "full_height": 1, + "full_width": 240, + "height_iter": 1, + "loop_num_channel": 12, + "loop_num_height": 1, + "loop_num_width": 5, + "num_iters": 15, + "pad_value": 0, + "reduce_axis": 2, + "reduce_dim": "W", + "scale": 0.004180908203125, + "scale_shift": 0, + "sv_channel": 40, + "sv_height": 1, + "sv_width": 48, + "width_iter": 5 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 1, + 240 + ], + "dims": [ + 480, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 1, + 1 + ], + "dims": [ + 480, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#24", + "type": "internal" + } + ], + "layer_order": 115 + }, + "Conv_192": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 480, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 4, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 120, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 4, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 120, + "ofm_depth_iter": 1, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 1, + 1 + ], + "dims": [ + 480, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#26", + "type": "internal" + } + ], + "layer_order": 116, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 4, + "ifm_depth_padded": 480, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 1, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 1586880, + "wts_no_of_reads": 5760 + } + }, + "Conv_194": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 120, + "ifm_depth_concat_extend": 8, + "ifm_depth_iter": 1, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 120, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 4, + "ofm_depth": 480, + "ofm_depth_iter": 4, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 120, + 1, + 1 + ], + "dims": [ + 128, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 1, + 1 + ], + "dims": [ + 512, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_192", + "type": "internal" + } + ], + "layer_order": 117, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 120, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 4, + "ofm_depth_padded": 512, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 1872576, + "wts_no_of_reads": 5760 + } + }, + "Add_196": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 128, + "ifmsv_height": 3, + "ifmsv_size": 384, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 1, + 1 + ], + "dims": [ + 512, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 1, + 1 + ], + "dims": [ + 512, + 12, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_194", + "type": "internal" + } + ], + "layer_order": 118 + }, + "Clip_199": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 128, + "ifmsv_height": 2, + "ifmsv_size": 256, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 1, + 1 + ], + "dims": [ + 512, + 12, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 1, + 1 + ], + "dims": [ + 512, + 8, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_196", + "type": "internal" + } + ], + "layer_order": 119 + }, + "Div_201": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 128, + "ifmsv_height": 4, + "ifmsv_size": 512, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 1, + 1 + ], + "dims": [ + 512, + 8, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 1, + 1 + ], + "dims": [ + 512, + 16, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_199", + "type": "internal" + } + ], + "layer_order": 120 + }, + "Generated-#28": { + "TileAdf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "i_dim_c": 480, + "i_dim_h": 1, + "i_dim_n": 1, + "i_dim_w": 1, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "rep_dim_c": 1, + "rep_dim_h": 12, + "rep_dim_w": 20 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 1, + 1 + ], + "dims": [ + 480, + 1, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 480, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Div_201", + "type": "internal" + } + ], + "layer_order": 121 + }, + "Mul_202": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 480, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#28", + "type": "internal" + }, + { + "name": "Mul_190", + "type": "internal" + } + ], + "layer_order": 122 + }, + "Conv_203": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 480, + "ifm_depth_concat_extend": 32, + "ifm_depth_iter": 5, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 96, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 5, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 112, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 480, + 12, + 20 + ], + "dims": [ + 512, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 112, + 12, + 20 + ], + "dims": [ + 128, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_202", + "type": "internal" + } + ], + "layer_order": 123, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 5, + "ifm_depth_padded": 480, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 1, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 2158272, + "wts_no_of_reads": 4608 + } + }, + "Conv_204": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 112, + "ifm_depth_concat_extend": 16, + "ifm_depth_iter": 2, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 3, + "ofm_depth": 672, + "ofm_depth_iter": 3, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 56, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 112, + 12, + 20 + ], + "dims": [ + 128, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_203", + "type": "internal" + } + ], + "layer_order": 124, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 1008, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 128, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 7, + "ofm_depth_iter": 3, + "ofm_depth_padded": 672, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 2446272, + "wts_no_of_reads": 5376 + } + }, + "Add_206": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_204", + "type": "internal" + } + ], + "layer_order": 125 + }, + "Clip_209": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_206", + "type": "internal" + } + ], + "layer_order": 126 + }, + "Div_211": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_209", + "type": "internal" + } + ], + "layer_order": 127 + }, + "Mul_212": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_204", + "type": "internal" + }, + { + "name": "Div_211", + "type": "internal" + } + ], + "layer_order": 128 + }, + "Conv_213": { + "Conv2DBf16": { + "act": 0, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 704, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 6, + "ifm_sv_height.size": 6, + "ifm_sv_width": 26, + "ifm_sv_width.size": 26, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "ifmsv_size": 1344, + "kernel_height": 3, + "kernel_width": 3, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 672, + "ofm_depth_iter": 21, + "ofm_height": 12, + "ofm_len": 21, + "ofm_sv_depth": 8, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 25, + "ofm_width": 19, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_212", + "type": "internal" + } + ], + "layer_order": 129, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 4, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 21, + "ofm_depth_padded": 672, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 2857536, + "wts_no_of_reads": 144 + } + }, + "Add_215": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_213", + "type": "internal" + } + ], + "layer_order": 130 + }, + "Clip_218": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_215", + "type": "internal" + } + ], + "layer_order": 131 + }, + "Div_220": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_218", + "type": "internal" + } + ], + "layer_order": 132 + }, + "Mul_221": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_213", + "type": "internal" + }, + { + "name": "Div_220", + "type": "internal" + } + ], + "layer_order": 133 + }, + "Generated-#30": { + "Transpose4dAdf": { + "aie_arch": "aie2p", + "dim_0": 12, + "dim_1": 84, + "dim_2": 20, + "dim_3": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "perm": 6 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 240 + ], + "dims": [ + 672, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_221", + "type": "internal" + } + ], + "layer_order": 134 + }, + "Generated-#32": { + "ReduceMeanC8Bf16": { + "L3_depth": 672, + "L3_height": 1, + "L3_width": 240, + "aie_arch": "aie2p", + "compiler": "chess", + "depth_iter": 3, + "full_channel": 672, + "full_height": 1, + "full_width": 240, + "height_iter": 1, + "loop_num_channel": 12, + "loop_num_height": 1, + "loop_num_width": 8, + "num_iters": 24, + "pad_value": 0, + "reduce_axis": 2, + "reduce_dim": "W", + "scale": 0.004180908203125, + "scale_shift": 0, + "sv_channel": 56, + "sv_height": 1, + "sv_width": 32, + "width_iter": 8 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 240 + ], + "dims": [ + 672, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 672, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#30", + "type": "internal" + } + ], + "layer_order": 135 + }, + "Conv_223": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 672, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 7, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 96, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 7, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 168, + "ofm_depth_iter": 1, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 48, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 672, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 168, + 1, + 1 + ], + "dims": [ + 192, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#32", + "type": "internal" + } + ], + "layer_order": 136, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 864, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 7, + "ifm_depth_padded": 672, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 6, + "ofm_depth_iter": 1, + "ofm_depth_padded": 192, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 2905920, + "wts_no_of_reads": 6912 + } + }, + "Conv_225": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 168, + "ifm_depth_concat_extend": 24, + "ifm_depth_iter": 1, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 168, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 16, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 11, + "ofm_depth": 672, + "ofm_depth_iter": 11, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 16, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 168, + 1, + 1 + ], + "dims": [ + 192, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 704, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_223", + "type": "internal" + } + ], + "layer_order": 137, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 168, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 11, + "ofm_depth_padded": 704, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 3510720, + "wts_no_of_reads": 4032 + } + }, + "Add_227": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 192, + "ifmsv_height": 2, + "ifmsv_size": 384, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 704, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 768, + 8, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_225", + "type": "internal" + } + ], + "layer_order": 138 + }, + "Clip_230": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 256, + "ifmsv_height": 1, + "ifmsv_size": 256, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 768, + 8, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 1024, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_227", + "type": "internal" + } + ], + "layer_order": 139 + }, + "Div_232": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 256, + "ifmsv_height": 2, + "ifmsv_size": 512, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 1024, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 1024, + 8, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_230", + "type": "internal" + } + ], + "layer_order": 140 + }, + "Generated-#34": { + "TileAdf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "i_dim_c": 672, + "i_dim_h": 1, + "i_dim_n": 1, + "i_dim_w": 1, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "rep_dim_c": 1, + "rep_dim_h": 12, + "rep_dim_w": 20 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 672, + 1, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Div_232", + "type": "internal" + } + ], + "layer_order": 141 + }, + "Mul_233": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#34", + "type": "internal" + }, + { + "name": "Mul_221", + "type": "internal" + } + ], + "layer_order": 142 + }, + "Conv_234": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 1, + "ifmsv_size": 1024, + "ifmsv_width": 32, + "num_elems": 240, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 11, + "ifm_depth": 672, + "ifm_depth_concat_extend": 32, + "ifm_depth_iter": 7, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 96, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 7, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 112, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 112, + 12, + 20 + ], + "dims": [ + 128, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 112, + 12, + 20 + ], + "dims": [ + 128, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_233", + "type": "internal" + }, + { + "name": "Conv_203", + "type": "internal" + } + ], + "layer_order": 143, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 7, + "ifm_depth_padded": 672, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 1, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 4055616, + "wts_no_of_reads": 4608 + } + }, + "Conv_236": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 112, + "ifm_depth_concat_extend": 16, + "ifm_depth_iter": 2, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 3, + "ofm_depth": 672, + "ofm_depth_iter": 3, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 56, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 112, + 12, + 20 + ], + "dims": [ + 128, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_234", + "type": "internal" + } + ], + "layer_order": 144, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 1008, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 128, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 7, + "ofm_depth_iter": 3, + "ofm_depth_padded": 672, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 4458816, + "wts_no_of_reads": 5376 + } + }, + "Add_238": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_236", + "type": "internal" + } + ], + "layer_order": 145 + }, + "Clip_241": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_238", + "type": "internal" + } + ], + "layer_order": 146 + }, + "Div_243": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_241", + "type": "internal" + } + ], + "layer_order": 147 + }, + "Mul_244": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_236", + "type": "internal" + }, + { + "name": "Div_243", + "type": "internal" + } + ], + "layer_order": 148 + }, + "Conv_245": { + "Conv2DBf16": { + "act": 0, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 704, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 10, + "ifm_sv_height.size": 10, + "ifm_sv_width": 16, + "ifm_sv_width.size": 16, + "ifm_width": 20, + "ifm_x_iter": 3, + "ifm_y_iter": 2, + "ifmsv_size": 1280, + "kernel_height": 9, + "kernel_width": 9, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 672, + "ofm_depth_iter": 21, + "ofm_height": 12, + "ofm_len": 126, + "ofm_sv_depth": 8, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 5, + "ofm_width": 17, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 4, + "pad_left": 4, + "pad_right": 4, + "pad_top": 4, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_244", + "type": "internal" + } + ], + "layer_order": 149, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 9, + "kernel_width": 12, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 21, + "ofm_depth_padded": 672, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 4870080, + "wts_no_of_reads": 1296 + } + }, + "Add_247": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_245", + "type": "internal" + } + ], + "layer_order": 150 + }, + "Clip_250": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_247", + "type": "internal" + } + ], + "layer_order": 151 + }, + "Div_252": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_250", + "type": "internal" + } + ], + "layer_order": 152 + }, + "Mul_253": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_245", + "type": "internal" + }, + { + "name": "Div_252", + "type": "internal" + } + ], + "layer_order": 153 + }, + "Generated-#36": { + "Transpose4dAdf": { + "aie_arch": "aie2p", + "dim_0": 12, + "dim_1": 84, + "dim_2": 20, + "dim_3": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "perm": 6 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 240 + ], + "dims": [ + 672, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_253", + "type": "internal" + } + ], + "layer_order": 154 + }, + "Generated-#38": { + "ReduceMeanC8Bf16": { + "L3_depth": 672, + "L3_height": 1, + "L3_width": 240, + "aie_arch": "aie2p", + "compiler": "chess", + "depth_iter": 3, + "full_channel": 672, + "full_height": 1, + "full_width": 240, + "height_iter": 1, + "loop_num_channel": 12, + "loop_num_height": 1, + "loop_num_width": 8, + "num_iters": 24, + "pad_value": 0, + "reduce_axis": 2, + "reduce_dim": "W", + "scale": 0.004180908203125, + "scale_shift": 0, + "sv_channel": 56, + "sv_height": 1, + "sv_width": 32, + "width_iter": 8 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 240 + ], + "dims": [ + 672, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 672, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#36", + "type": "internal" + } + ], + "layer_order": 155 + }, + "Conv_255": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 672, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 7, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 96, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 7, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 168, + "ofm_depth_iter": 1, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 48, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 672, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 168, + 1, + 1 + ], + "dims": [ + 192, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#38", + "type": "internal" + } + ], + "layer_order": 156, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 864, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 7, + "ifm_depth_padded": 672, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 6, + "ofm_depth_iter": 1, + "ofm_depth_padded": 192, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 5208768, + "wts_no_of_reads": 6912 + } + }, + "Conv_257": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 168, + "ifm_depth_concat_extend": 24, + "ifm_depth_iter": 1, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 168, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 16, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 11, + "ofm_depth": 672, + "ofm_depth_iter": 11, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 16, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 168, + 1, + 1 + ], + "dims": [ + 192, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 704, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_255", + "type": "internal" + } + ], + "layer_order": 157, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 168, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 11, + "ofm_depth_padded": 704, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 5813568, + "wts_no_of_reads": 4032 + } + }, + "Add_259": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 192, + "ifmsv_height": 2, + "ifmsv_size": 384, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 704, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 768, + 8, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_257", + "type": "internal" + } + ], + "layer_order": 158 + }, + "Clip_262": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 256, + "ifmsv_height": 1, + "ifmsv_size": 256, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 768, + 8, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 1024, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_259", + "type": "internal" + } + ], + "layer_order": 159 + }, + "Div_264": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 256, + "ifmsv_height": 2, + "ifmsv_size": 512, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 1024, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 1024, + 8, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_262", + "type": "internal" + } + ], + "layer_order": 160 + }, + "Generated-#40": { + "TileAdf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "i_dim_c": 672, + "i_dim_h": 1, + "i_dim_n": 1, + "i_dim_w": 1, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "rep_dim_c": 1, + "rep_dim_h": 12, + "rep_dim_w": 20 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 1, + 1 + ], + "dims": [ + 672, + 1, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Div_264", + "type": "internal" + } + ], + "layer_order": 161 + }, + "Mul_265": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 11, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 672, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#40", + "type": "internal" + }, + { + "name": "Mul_253", + "type": "internal" + } + ], + "layer_order": 162 + }, + "Conv_266": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 672, + "ifm_depth_concat_extend": 32, + "ifm_depth_iter": 9, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 9, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 160, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 40, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 672, + 12, + 20 + ], + "dims": [ + 704, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 160, + 12, + 20 + ], + "dims": [ + 160, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_265", + "type": "internal" + } + ], + "layer_order": 163, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 720, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 9, + "ifm_depth_padded": 720, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 5, + "ofm_depth_iter": 1, + "ofm_depth_padded": 160, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 6358464, + "wts_no_of_reads": 4800 + } + }, + "Conv_267": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 160, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 2, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 6, + "ofm_depth": 960, + "ofm_depth_iter": 6, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 40, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 160, + 12, + 20 + ], + "dims": [ + 160, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_266", + "type": "internal" + } + ], + "layer_order": 164, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 720, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 160, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 5, + "ofm_depth_iter": 6, + "ofm_depth_padded": 960, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 6902784, + "wts_no_of_reads": 4800 + } + }, + "Add_269": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_267", + "type": "internal" + } + ], + "layer_order": 165 + }, + "Clip_272": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_269", + "type": "internal" + } + ], + "layer_order": 166 + }, + "Div_274": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_272", + "type": "internal" + } + ], + "layer_order": 167 + }, + "Mul_275": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 2, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 20, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_267", + "type": "internal" + }, + { + "name": "Div_274", + "type": "internal" + } + ], + "layer_order": 168 + }, + "Conv_276": { + "Conv2DBf16": { + "act": 0, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 960, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 10, + "ifm_sv_height.size": 10, + "ifm_sv_width": 16, + "ifm_sv_width.size": 16, + "ifm_width": 20, + "ifm_x_iter": 3, + "ifm_y_iter": 2, + "ifmsv_size": 1280, + "kernel_height": 9, + "kernel_width": 9, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 960, + "ofm_depth_iter": 30, + "ofm_height": 12, + "ofm_len": 180, + "ofm_sv_depth": 8, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 5, + "ofm_width": 17, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 4, + "pad_left": 4, + "pad_right": 4, + "pad_top": 4, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_275", + "type": "internal" + } + ], + "layer_order": 169, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 9, + "kernel_width": 12, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 30, + "ofm_depth_padded": 960, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 7628544, + "wts_no_of_reads": 1296 + } + }, + "Add_278": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_276", + "type": "internal" + } + ], + "layer_order": 170 + }, + "Clip_281": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_278", + "type": "internal" + } + ], + "layer_order": 171 + }, + "Div_283": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_281", + "type": "internal" + } + ], + "layer_order": 172 + }, + "Mul_284": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 2, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 20, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_276", + "type": "internal" + }, + { + "name": "Div_283", + "type": "internal" + } + ], + "layer_order": 173 + }, + "Generated-#42": { + "Transpose4dAdf": { + "aie_arch": "aie2p", + "dim_0": 12, + "dim_1": 120, + "dim_2": 20, + "dim_3": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "perm": 6 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 240 + ], + "dims": [ + 960, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_284", + "type": "internal" + } + ], + "layer_order": 174 + }, + "Generated-#44": { + "ReduceMeanC8Bf16": { + "L3_depth": 960, + "L3_height": 1, + "L3_width": 240, + "aie_arch": "aie2p", + "compiler": "chess", + "depth_iter": 6, + "full_channel": 960, + "full_height": 1, + "full_width": 240, + "height_iter": 1, + "loop_num_channel": 24, + "loop_num_height": 1, + "loop_num_width": 5, + "num_iters": 30, + "pad_value": 0, + "reduce_axis": 2, + "reduce_dim": "W", + "scale": 0.004180908203125, + "scale_shift": 0, + "sv_channel": 40, + "sv_height": 1, + "sv_width": 48, + "width_iter": 5 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 240 + ], + "dims": [ + 960, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#42", + "type": "internal" + } + ], + "layer_order": 175 + }, + "Conv_286": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 960, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 5, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 192, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 16, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 5, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 4, + "ofm_depth": 240, + "ofm_depth_iter": 4, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 16, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 1, + 1 + ], + "dims": [ + 256, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#44", + "type": "internal" + } + ], + "layer_order": 176, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 5, + "ifm_depth_padded": 960, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 4, + "ofm_depth_padded": 256, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 8112384, + "wts_no_of_reads": 4608 + } + }, + "Conv_288": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 240, + "ifm_depth_concat_extend": 16, + "ifm_depth_iter": 3, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 3, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 5, + "ofm_depth": 960, + "ofm_depth_iter": 5, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 48, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 1, + 1 + ], + "dims": [ + 256, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_286", + "type": "internal" + } + ], + "layer_order": 177, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 864, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 3, + "ifm_depth_padded": 240, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 6, + "ofm_depth_iter": 5, + "ofm_depth_padded": 960, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 9241344, + "wts_no_of_reads": 5760 + } + }, + "Add_290": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 384, + "ifmsv_height": 1, + "ifmsv_size": 384, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 1536, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_288", + "type": "internal" + } + ], + "layer_order": 178 + }, + "Clip_293": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 256, + "ifmsv_height": 1, + "ifmsv_size": 256, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 1536, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 1024, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_290", + "type": "internal" + } + ], + "layer_order": 179 + }, + "Div_295": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 256, + "ifmsv_height": 2, + "ifmsv_size": 512, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 1024, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 1024, + 8, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_293", + "type": "internal" + } + ], + "layer_order": 180 + }, + "Generated-#46": { + "TileAdf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "i_dim_c": 960, + "i_dim_h": 1, + "i_dim_n": 1, + "i_dim_w": 1, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "rep_dim_c": 1, + "rep_dim_h": 12, + "rep_dim_w": 20 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 1, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Div_295", + "type": "internal" + } + ], + "layer_order": 181 + }, + "Mul_296": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 15, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#46", + "type": "internal" + }, + { + "name": "Mul_284", + "type": "internal" + } + ], + "layer_order": 182 + }, + "Conv_297": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 1, + "ifmsv_size": 768, + "ifmsv_width": 32, + "num_elems": 240, + "num_kernel_iters": 6, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 11, + "ifm_depth": 960, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 9, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 112, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 9, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 2, + "ofm_depth": 160, + "ofm_depth_iter": 2, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 24, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 160, + 12, + 20 + ], + "dims": [ + 192, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 160, + 12, + 20 + ], + "dims": [ + 192, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_296", + "type": "internal" + }, + { + "name": "Conv_266", + "type": "internal" + } + ], + "layer_order": 183, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 432, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 9, + "ifm_depth_padded": 1008, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 3, + "ofm_depth_iter": 2, + "ofm_depth_padded": 192, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 10329984, + "wts_no_of_reads": 4032 + } + }, + "Conv_299": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 160, + "ifm_depth_concat_extend": 32, + "ifm_depth_iter": 2, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 6, + "ofm_depth": 960, + "ofm_depth_iter": 6, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 40, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 160, + 12, + 20 + ], + "dims": [ + 192, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_297", + "type": "internal" + } + ], + "layer_order": 184, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 720, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 160, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 5, + "ofm_depth_iter": 6, + "ofm_depth_padded": 960, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 11232000, + "wts_no_of_reads": 4800 + } + }, + "Add_301": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_299", + "type": "internal" + } + ], + "layer_order": 185 + }, + "Clip_304": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_301", + "type": "internal" + } + ], + "layer_order": 186 + }, + "Div_306": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_304", + "type": "internal" + } + ], + "layer_order": 187 + }, + "Mul_307": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 2, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 20, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_299", + "type": "internal" + }, + { + "name": "Div_306", + "type": "internal" + } + ], + "layer_order": 188 + }, + "Conv_308": { + "Conv2DBf16": { + "act": 0, + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 1, + "bias_shift": 0, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ifm.dimensions.batch.padding": 0, + "data_io.ifm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.batch.padding": 0, + "data_io.ofm.dimensions.batch.size": 1, + "data_io.ofm.dimensions.width.padding": 0, + "func_type": 10, + "ifm_depth": 960, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_depth.size": 8, + "ifm_sv_height": 10, + "ifm_sv_height.size": 10, + "ifm_sv_width": 16, + "ifm_sv_width.size": 16, + "ifm_width": 20, + "ifm_x_iter": 3, + "ifm_y_iter": 2, + "ifmsv_size": 1280, + "kernel_height": 9, + "kernel_width": 9, + "num_adf_cols": 4, + "num_adf_rows": 4, + "ofm_depth": 960, + "ofm_depth_iter": 30, + "ofm_height": 12, + "ofm_len": 180, + "ofm_sv_depth": 8, + "ofm_sv_height": 2, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 5, + "ofm_width": 17, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 4, + "pad_left": 4, + "pad_right": 4, + "pad_top": 4, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 1, + "stride": 1, + "stride_height": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_307", + "type": "internal" + } + ], + "layer_order": 189, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 8, + "is_dwc": true, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 9, + "kernel_width": 12, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 30, + "ofm_depth_padded": 960, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 11957760, + "wts_no_of_reads": 1296 + } + }, + "Add_310": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 16, + 24 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_308", + "type": "internal" + } + ], + "layer_order": 190 + }, + "Clip_313": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_310", + "type": "internal" + } + ], + "layer_order": 191 + }, + "Div_315": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_313", + "type": "internal" + } + ], + "layer_order": 192 + }, + "Mul_316": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 2, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 20, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_308", + "type": "internal" + }, + { + "name": "Div_315", + "type": "internal" + } + ], + "layer_order": 193 + }, + "Generated-#48": { + "Transpose4dAdf": { + "aie_arch": "aie2p", + "dim_0": 12, + "dim_1": 120, + "dim_2": 20, + "dim_3": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "perm": 6 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 240 + ], + "dims": [ + 960, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_316", + "type": "internal" + } + ], + "layer_order": 194 + }, + "Generated-#50": { + "ReduceMeanC8Bf16": { + "L3_depth": 960, + "L3_height": 1, + "L3_width": 240, + "aie_arch": "aie2p", + "compiler": "chess", + "depth_iter": 6, + "full_channel": 960, + "full_height": 1, + "full_width": 240, + "height_iter": 1, + "loop_num_channel": 24, + "loop_num_height": 1, + "loop_num_width": 5, + "num_iters": 30, + "pad_value": 0, + "reduce_axis": 2, + "reduce_dim": "W", + "scale": 0.004180908203125, + "scale_shift": 0, + "sv_channel": 40, + "sv_height": 1, + "sv_width": 48, + "width_iter": 5 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 240 + ], + "dims": [ + 960, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#48", + "type": "internal" + } + ], + "layer_order": 195 + }, + "Conv_318": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 960, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 5, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 192, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 16, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 5, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 4, + "ofm_depth": 240, + "ofm_depth_iter": 4, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 16, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 240, + 1, + 1 + ], + "dims": [ + 256, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#50", + "type": "internal" + } + ], + "layer_order": 196, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 5, + "ifm_depth_padded": 960, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 4, + "ofm_depth_padded": 256, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 12441600, + "wts_no_of_reads": 4608 + } + }, + "Conv_320": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 8, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 240, + "ifm_depth_concat_extend": 16, + "ifm_depth_iter": 3, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 8, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 3, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 5, + "ofm_depth": 960, + "ofm_depth_iter": 5, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 48, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 8, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 8, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 240, + 1, + 1 + ], + "dims": [ + 256, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_318", + "type": "internal" + } + ], + "layer_order": 197, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 864, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 3, + "ifm_depth_padded": 240, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 6, + "ofm_depth_iter": 5, + "ofm_depth_padded": 960, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 13570560, + "wts_no_of_reads": 5760 + } + }, + "Add_322": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 384, + "ifmsv_height": 1, + "ifmsv_size": 384, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 1, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 1536, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_320", + "type": "internal" + } + ], + "layer_order": 198 + }, + "Clip_325": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 256, + "ifmsv_height": 1, + "ifmsv_size": 256, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 1536, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 1024, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_322", + "type": "internal" + } + ], + "layer_order": 199 + }, + "Div_327": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 256, + "ifmsv_height": 2, + "ifmsv_size": 512, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 1024, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 1024, + 8, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_325", + "type": "internal" + } + ], + "layer_order": 200 + }, + "Generated-#52": { + "TileAdf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "i_dim_c": 960, + "i_dim_h": 1, + "i_dim_n": 1, + "i_dim_w": 1, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "rep_dim_c": 1, + "rep_dim_h": 12, + "rep_dim_w": 20 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 1, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Div_327", + "type": "internal" + } + ], + "layer_order": 201 + }, + "Mul_328": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 15, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#52", + "type": "internal" + }, + { + "name": "Mul_316", + "type": "internal" + } + ], + "layer_order": 202 + }, + "Conv_329": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 1, + "ifmsv_size": 768, + "ifmsv_width": 32, + "num_elems": 240, + "num_kernel_iters": 6, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 11, + "ifm_depth": 960, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 9, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 112, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 9, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 2, + "ofm_depth": 160, + "ofm_depth_iter": 2, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 24, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 160, + 12, + 20 + ], + "dims": [ + 192, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 160, + 12, + 20 + ], + "dims": [ + 192, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_328", + "type": "internal" + }, + { + "name": "Conv_297", + "type": "internal" + } + ], + "layer_order": 203, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 432, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 9, + "ifm_depth_padded": 1008, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 3, + "ofm_depth_iter": 2, + "ofm_depth_padded": 192, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 14659200, + "wts_no_of_reads": 4032 + } + }, + "Conv_331": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 160, + "ifm_depth_concat_extend": 32, + "ifm_depth_iter": 2, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 80, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 6, + "ofm_depth": 960, + "ofm_depth_iter": 6, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 40, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 160, + 12, + 20 + ], + "dims": [ + 192, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_329", + "type": "internal" + } + ], + "layer_order": 204, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 720, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 160, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 5, + "ofm_depth_iter": 6, + "ofm_depth_padded": 960, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 15561216, + "wts_no_of_reads": 4800 + } + }, + "Add_333": { + "AddAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 3, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_331", + "type": "internal" + } + ], + "layer_order": 205 + }, + "Clip_336": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 6, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Add_333", + "type": "internal" + } + ], + "layer_order": 206 + }, + "Div_338": { + "MulAttributeBroadcastingBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 80, + "ifmsv_height": 1, + "ifmsv_size": 1600, + "ifmsv_width": 20, + "num_kernel_iters": 9, + "ofm_shift": 0, + "ofm_shift_biased": 0, + "scalar": 0.166015625, + "scalar_position": 1, + "scalar_shift": 0, + "scalar_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Clip_336", + "type": "internal" + } + ], + "layer_order": 207 + }, + "Mul_339": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 2, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 20, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_331", + "type": "internal" + }, + { + "name": "Div_338", + "type": "internal" + } + ], + "layer_order": 208 + }, + "Generated-#54": { + "Transpose4dAdf": { + "aie_arch": "aie2p", + "dim_0": 12, + "dim_1": 120, + "dim_2": 20, + "dim_3": 8, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "perm": 6 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 240 + ], + "dims": [ + 960, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_339", + "type": "internal" + } + ], + "layer_order": 209 + }, + "Generated-#56": { + "ReduceMeanC8Bf16": { + "L3_depth": 960, + "L3_height": 1, + "L3_width": 240, + "aie_arch": "aie2p", + "compiler": "chess", + "depth_iter": 6, + "full_channel": 960, + "full_height": 1, + "full_width": 240, + "height_iter": 1, + "loop_num_channel": 24, + "loop_num_height": 1, + "loop_num_width": 5, + "num_iters": 30, + "pad_value": 0, + "reduce_axis": 2, + "reduce_dim": "W", + "scale": 0.004180908203125, + "scale_shift": 0, + "sv_channel": 40, + "sv_height": 1, + "sv_width": 48, + "width_iter": 5 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 240 + ], + "dims": [ + 960, + 1, + 240 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#54", + "type": "internal" + } + ], + "layer_order": 210 + }, + "Conv_343": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 12, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 960, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 5, + "ifm_height": 1, + "ifm_sign": 0, + "ifm_sv_depth": 192, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 16, + "ifm_width": 1, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 5, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 2, + "ofm_depth": 128, + "ofm_depth_iter": 2, + "ofm_height": 1, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 1, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 16, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 1, + 1 + ], + "dims": [ + 960, + 4, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 1, + 1 + ], + "dims": [ + 128, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#56", + "type": "internal" + } + ], + "layer_order": 211, + "split_width": true, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 5, + "ifm_depth_padded": 960, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 2, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 16286976, + "wts_no_of_reads": 4608 + } + }, + "Sigmoid_344": { + "SigmoidTemplatedBf16": { + "ENABLE_FP16_AS_BF16": 0, + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 16, + "ifmsv_size": 512, + "ifmsv_width": 1, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 1, + 1 + ], + "dims": [ + 128, + 1, + 64 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 1, + 1 + ], + "dims": [ + 128, + 64, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_343", + "type": "internal" + } + ], + "layer_order": 212 + }, + "Generated-#58": { + "TileAdf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "i_dim_c": 128, + "i_dim_h": 1, + "i_dim_n": 1, + "i_dim_w": 1, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "output_l3_format": "HCWN_C8", + "rep_dim_c": 1, + "rep_dim_h": 12, + "rep_dim_w": 20 + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 1, + 1 + ], + "dims": [ + 128, + 1, + 1 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sigmoid_344", + "type": "internal" + } + ], + "layer_order": 213 + }, + "Conv_340": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 16, + "ifm_depth": 960, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 10, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 96, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 32, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 3, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 10, + "num_ifm_height_iters": 3, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 128, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 32, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 32, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 1, + "ifmsv_size": 1024, + "ifmsv_width": 32, + "num_elems": 240, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 960, + 12, + 20 + ], + "dims": [ + 960, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Mul_339", + "type": "internal" + }, + { + "name": "Generated-#58", + "type": "internal" + } + ], + "layer_order": 214, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 576, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 10, + "ifm_depth_padded": 960, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 4, + "ofm_depth_iter": 1, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 16851456, + "wts_no_of_reads": 4608 + } + }, + "Split_349_Duplicated#0": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 128, + "dim_h": 12, + "dim_w": 20, + "dtype": "bfloat16", + "end": 64, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_340", + "type": "internal" + } + ], + "layer_order": 215 + }, + "Split_349_Duplicated#1": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 128, + "dim_h": 12, + "dim_w": 20, + "dtype": "bfloat16", + "end": 128, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 64, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_340", + "type": "internal" + } + ], + "layer_order": 216 + }, + "Concat_350": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Split_349_Duplicated#1", + "type": "internal" + }, + { + "name": "Input.4", + "type": "external" + } + ], + "layer_order": 217 + }, + "Conv_351": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 128, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 8, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 5, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 34, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 8, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 2, + "ofm_depth": 128, + "ofm_depth_iter": 2, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 3, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 5, + "padding_sv_width": 34, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_350", + "type": "internal" + } + ], + "layer_order": 218, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 8, + "ifm_depth_padded": 128, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 2, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 17427456, + "wts_no_of_reads": 3456 + } + }, + "Sigmoid_352": { + "SigmoidTemplatedBf16": { + "ENABLE_FP16_AS_BF16": 0, + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 32, + "ifmsv_height": 3, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_351", + "type": "internal" + } + ], + "layer_order": 219 + }, + "Split_353_Duplicated#1": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 128, + "dim_h": 12, + "dim_w": 20, + "dtype": "bfloat16", + "end": 128, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 64, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sigmoid_352", + "type": "internal" + } + ], + "layer_order": 220 + }, + "Sub_359": { + "SubBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "constants_reader": { + "Const.2": { + "bytes_to_skip": 20764800, + "channel_stride": 8, + "depth": 1, + "dtype": "bfloat16", + "height": 12, + "nibbles": 4, + "num_reads": 360, + "rank": 4, + "reads_per_line": 3, + "width": 20 + } + }, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Const.2", + "type": "constant" + }, + { + "name": "Split_353_Duplicated#1", + "type": "internal" + } + ], + "layer_order": 221 + }, + "Mul_360": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sub_359", + "type": "internal" + }, + { + "name": "Input.4", + "type": "external" + } + ], + "layer_order": 222 + }, + "Split_353_Duplicated#0": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 128, + "dim_h": 12, + "dim_w": 20, + "dtype": "bfloat16", + "end": 64, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sigmoid_352", + "type": "internal" + } + ], + "layer_order": 223 + }, + "Mul_354": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_353_Duplicated#0", + "type": "internal" + }, + { + "name": "Input.4", + "type": "external" + } + ], + "layer_order": 224 + }, + "Concat_355": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Split_349_Duplicated#1", + "type": "internal" + }, + { + "name": "Mul_354", + "type": "internal" + } + ], + "layer_order": 225 + }, + "Conv_356": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 128, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 8, + "ifm_height": 12, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 5, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 34, + "ifm_width": 20, + "ifm_x_iter": 1, + "ifm_y_iter": 1, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 8, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 64, + "ofm_depth_iter": 1, + "ofm_height": 12, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 3, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 20, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 5, + "padding_sv_width": 34, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_355", + "type": "internal" + } + ], + "layer_order": 226, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 8, + "ifm_depth_padded": 128, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 18109440, + "wts_no_of_reads": 3456 + } + }, + "Tanh_357": { + "TanhTemplatedBf16": { + "ENABLE_FP16_AS_BF16": 0, + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 32 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_356", + "type": "internal" + } + ], + "layer_order": 227 + }, + "Mul_361": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_353_Duplicated#1", + "type": "internal" + }, + { + "name": "Tanh_357", + "type": "internal" + } + ], + "layer_order": 228 + }, + "Add_362": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 3, + "ifmsv_size": 960, + "ifmsv_width": 20, + "num_elems": 240, + "num_kernel_iters": 1, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "external_output_layer": 3, + "input_layer_names": [ + { + "name": "Mul_360", + "type": "internal" + }, + { + "name": "Mul_361", + "type": "internal" + } + ], + "layer_order": 229 + }, + "Concat_363": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 64, + 12, + 20 + ], + "dims": [ + 64, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Split_349_Duplicated#0", + "type": "internal" + }, + { + "name": "Add_362", + "type": "internal" + } + ], + "layer_order": 230 + }, + "Resize_365": { + "ResizeAdf": { + "co_trans_mode": 1, + "dim_0": 1, + "dim_1": 128, + "dim_2": 12, + "dim_3": 20, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "mode": 1, + "nearest_mode": 0, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_H": 24, + "output_W": 40, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 12, + 20 + ], + "dims": [ + 128, + 12, + 20 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 24, + 40 + ], + "dims": [ + 128, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_363", + "type": "internal" + } + ], + "layer_order": 231 + }, + "Slice_371": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "H", + "dim_c": 128, + "dim_h": 24, + "dim_w": 40, + "dtype": "bfloat16", + "end": 23, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 24, + 40 + ], + "dims": [ + 128, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 128, + 23, + 40 + ], + "dims": [ + 128, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Resize_365", + "type": "internal" + } + ], + "layer_order": 232 + }, + "AveragePool_346": { + "AvgPool2DBf16": { + "aie_arch": "aie2p", + "avgpool_skip_width": 0, + "batch_size": 1, + "compiler": "chess", + "data_io.ofm.dimensions.width": 40, + "data_io.ofm.dimensions.width.padding": 0, + "div_factor": 0.25, + "div_shift": 0, + "dtype": "bfloat16", + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_height": 2, + "ifm_sv_height_eff": 2, + "ifm_sv_size": 1280, + "ifm_sv_width": 80, + "ifm_sv_width_eff": 80, + "kernel_height": 2, + "kernel_width": 2, + "ksize": 2, + "manual_ba_pad_val": 0, + "num_iter": 640, + "ofm_len": 96, + "ofm_sv_depth": 8, + "ofm_sv_height": 1, + "ofm_sv_width": 40, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "stride_height": 2, + "stride_log2": 1, + "stride_width": 2 + }, + "Conv2D": { + "act": 0, + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 0, + "conv2d_ofm_pad": 0, + "conv_type": 0, + "func_type": 6, + "ifm_depth": 8, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 0, + "ifm_height": 0, + "ifm_sign": 0, + "ifm_sv_depth": 0, + "ifm_sv_height": 0, + "ifm_sv_width": 0, + "ifm_width": 0, + "ifm_x_iter": 0, + "ifm_y_iter": 0, + "kernel_height": 0, + "kernel_width": 0, + "num_adf_cols": 0, + "num_adf_rows": 0, + "ofm_depth": 8, + "ofm_depth_iter": 0, + "ofm_height": 0, + "ofm_sv_depth": 0, + "ofm_sv_height": 0, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 0, + "ofm_width": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "relu_int8_enable": 1, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_height": 0, + "stride_width": 0, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 90, + 160 + ], + "dims": [ + 8, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Generated-#4", + "type": "internal" + } + ], + "layer_order": 233, + "split_width": false + }, + "AveragePool_347": { + "AvgPool2DBf16": { + "aie_arch": "aie2p", + "avgpool_skip_width": 0, + "batch_size": 1, + "compiler": "chess", + "data_io.ofm.dimensions.width": 8, + "data_io.ofm.dimensions.width.padding": 0, + "div_factor": 0.25, + "div_shift": 0, + "dtype": "bfloat16", + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_height": 12, + "ifm_sv_height_eff": 12, + "ifm_sv_size": 1536, + "ifm_sv_width": 16, + "ifm_sv_width_eff": 16, + "kernel_height": 2, + "kernel_width": 2, + "ksize": 2, + "manual_ba_pad_val": 0, + "num_iter": 768, + "ofm_len": 20, + "ofm_sv_depth": 8, + "ofm_sv_height": 6, + "ofm_sv_width": 8, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "stride_height": 2, + "stride_log2": 1, + "stride_width": 2 + }, + "Conv2D": { + "act": 0, + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 0, + "conv2d_ofm_pad": 0, + "conv_type": 0, + "func_type": 6, + "ifm_depth": 8, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 0, + "ifm_height": 0, + "ifm_sign": 0, + "ifm_sv_depth": 0, + "ifm_sv_height": 0, + "ifm_sv_width": 0, + "ifm_width": 0, + "ifm_x_iter": 0, + "ifm_y_iter": 0, + "kernel_height": 0, + "kernel_width": 0, + "num_adf_cols": 0, + "num_adf_rows": 0, + "ofm_depth": 8, + "ofm_depth_iter": 0, + "ofm_height": 0, + "ofm_sv_depth": 0, + "ofm_sv_height": 0, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 0, + "ofm_width": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "relu_int8_enable": 1, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_height": 0, + "stride_width": 0, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 3, + 90, + 160 + ], + "dims": [ + 8, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "AveragePool_346", + "type": "internal" + } + ], + "layer_order": 234, + "split_width": false + }, + "AveragePool_348": { + "AvgPool2DBf16": { + "aie_arch": "aie2p", + "avgpool_skip_width": 0, + "batch_size": 1, + "compiler": "chess", + "data_io.ofm.dimensions.width": 8, + "data_io.ofm.dimensions.width.padding": 0, + "div_factor": 0.25, + "div_shift": 0, + "dtype": "bfloat16", + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_height": 12, + "ifm_sv_height_eff": 12, + "ifm_sv_size": 1536, + "ifm_sv_width": 16, + "ifm_sv_width_eff": 16, + "kernel_height": 2, + "kernel_width": 2, + "ksize": 2, + "manual_ba_pad_val": 0, + "num_iter": 768, + "ofm_len": 5, + "ofm_sv_depth": 8, + "ofm_sv_height": 6, + "ofm_sv_width": 8, + "pad_bottom": 1, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "stride_height": 2, + "stride_log2": 1, + "stride_width": 2 + }, + "Conv2D": { + "act": 0, + "alpha": 0, + "alpha_scaled": 0, + "batch_size": 0, + "conv2d_ofm_pad": 0, + "conv_type": 0, + "func_type": 6, + "ifm_depth": 8, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 0, + "ifm_height": 0, + "ifm_sign": 0, + "ifm_sv_depth": 0, + "ifm_sv_height": 0, + "ifm_sv_width": 0, + "ifm_width": 0, + "ifm_x_iter": 0, + "ifm_y_iter": 0, + "kernel_height": 0, + "kernel_width": 0, + "num_adf_cols": 0, + "num_adf_rows": 0, + "ofm_depth": 8, + "ofm_depth_iter": 0, + "ofm_height": 0, + "ofm_sv_depth": 0, + "ofm_sv_height": 0, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 0, + "ofm_width": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "relu_int8_enable": 1, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_height": 0, + "stride_width": 0, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 3, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 23, + 40 + ], + "dims": [ + 32, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "AveragePool_347", + "type": "internal" + } + ], + "layer_order": 235, + "split_width": false + }, + "Concat_372": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 128, + 23, + 40 + ], + "dims": [ + 128, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 23, + 40 + ], + "dims": [ + 8, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 171, + 23, + 40 + ], + "dims": [ + 176, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Slice_371", + "type": "internal" + }, + { + "name": "Conv_92", + "type": "internal" + }, + { + "name": "AveragePool_348", + "type": "internal" + } + ], + "layer_order": 236 + }, + "Conv_373": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 48, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 176, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 11, + "ifm_height": 23, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 5, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 50, + "ifm_width": 40, + "ifm_x_iter": 1, + "ifm_y_iter": 2, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 11, + "num_ifm_height_iters": 2, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 2, + "ofm_depth": 80, + "ofm_depth_iter": 2, + "ofm_height": 23, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 3, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 48, + "ofm_width": 40, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 5, + "padding_sv_width": 50, + "psum_buff_offset_scaled": 3, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 171, + 23, + 40 + ], + "dims": [ + 176, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 128, + 24, + 48 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_372", + "type": "internal" + } + ], + "layer_order": 237, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 11, + "ifm_depth_padded": 176, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 2, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 18450432, + "wts_no_of_reads": 3456 + } + }, + "Split_375_Duplicated#0": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 80, + "dim_h": 23, + "dim_w": 40, + "dtype": "bfloat16", + "end": 40, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 80, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_373", + "type": "internal" + } + ], + "layer_order": 238 + }, + "Split_375_Duplicated#1": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 80, + "dim_h": 23, + "dim_w": 40, + "dtype": "bfloat16", + "end": 80, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 40, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 80, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_373", + "type": "internal" + } + ], + "layer_order": 239 + }, + "Concat_376": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 80, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Split_375_Duplicated#1", + "type": "internal" + }, + { + "name": "Input.3", + "type": "external" + } + ], + "layer_order": 240 + }, + "Conv_377": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 48, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 80, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 5, + "ifm_height": 23, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 5, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 50, + "ifm_width": 40, + "ifm_x_iter": 1, + "ifm_y_iter": 2, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 5, + "num_ifm_height_iters": 2, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 2, + "ofm_depth": 80, + "ofm_depth_iter": 2, + "ofm_height": 23, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 3, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 48, + "ofm_width": 40, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 5, + "padding_sv_width": 50, + "psum_buff_offset_scaled": 3, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 80, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 128, + 24, + 48 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_376", + "type": "internal" + } + ], + "layer_order": 241, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 5, + "ifm_depth_padded": 80, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 2, + "ofm_depth_padded": 128, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 19388160, + "wts_no_of_reads": 3456 + } + }, + "Sigmoid_378": { + "SigmoidTemplatedBf16": { + "ENABLE_FP16_AS_BF16": 0, + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 24, + "ifmsv_height": 2, + "ifmsv_size": 1920, + "ifmsv_width": 40, + "num_kernel_iters": 3, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 128, + 24, + 48 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 96, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_377", + "type": "internal" + } + ], + "layer_order": 242 + }, + "Split_379_Duplicated#1": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 80, + "dim_h": 23, + "dim_w": 40, + "dtype": "bfloat16", + "end": 80, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 40, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 80, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sigmoid_378", + "type": "internal" + } + ], + "layer_order": 243 + }, + "Sub_385": { + "SubBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 920, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "constants_reader": { + "Const.3": { + "bytes_to_skip": 20764800, + "channel_stride": 8, + "depth": 1, + "dtype": "bfloat16", + "height": 23, + "nibbles": 4, + "num_reads": 1380, + "rank": 4, + "reads_per_line": 3, + "width": 40 + } + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Const.3", + "type": "constant" + }, + { + "name": "Split_379_Duplicated#1", + "type": "internal" + } + ], + "layer_order": 244 + }, + "Mul_386": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 920, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sub_385", + "type": "internal" + }, + { + "name": "Input.3", + "type": "external" + } + ], + "layer_order": 245 + }, + "Split_379_Duplicated#0": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 80, + "dim_h": 23, + "dim_w": 40, + "dtype": "bfloat16", + "end": 40, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 80, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sigmoid_378", + "type": "internal" + } + ], + "layer_order": 246 + }, + "Mul_380": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 920, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_379_Duplicated#0", + "type": "internal" + }, + { + "name": "Input.3", + "type": "external" + } + ], + "layer_order": 247 + }, + "Concat_381": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 80, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Split_375_Duplicated#1", + "type": "internal" + }, + { + "name": "Mul_380", + "type": "internal" + } + ], + "layer_order": 248 + }, + "Conv_382": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 48, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 80, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 5, + "ifm_height": 23, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 5, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 50, + "ifm_width": 40, + "ifm_x_iter": 1, + "ifm_y_iter": 2, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 5, + "num_ifm_height_iters": 2, + "num_ifm_width_iters": 1, + "num_ofm_depth_iters": 1, + "ofm_depth": 40, + "ofm_depth_iter": 1, + "ofm_height": 23, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 3, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 48, + "ofm_width": 40, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 5, + "padding_sv_width": 50, + "psum_buff_offset_scaled": 3, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 80, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 48 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_381", + "type": "internal" + } + ], + "layer_order": 249, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 5, + "ifm_depth_padded": 80, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 19814400, + "wts_no_of_reads": 3456 + } + }, + "Tanh_383": { + "TanhTemplatedBf16": { + "ENABLE_FP16_AS_BF16": 0, + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 2, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 48 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_382", + "type": "internal" + } + ], + "layer_order": 250 + }, + "Mul_387": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 920, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_379_Duplicated#1", + "type": "internal" + }, + { + "name": "Tanh_383", + "type": "internal" + } + ], + "layer_order": 251 + }, + "Add_388": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 6, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 920, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 64, + 24, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "external_output_layer": 2, + "input_layer_names": [ + { + "name": "Mul_386", + "type": "internal" + }, + { + "name": "Mul_387", + "type": "internal" + } + ], + "layer_order": 252 + }, + "Concat_389": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 23, + 40 + ], + "dims": [ + 40, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 80, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Split_375_Duplicated#0", + "type": "internal" + }, + { + "name": "Add_388", + "type": "internal" + } + ], + "layer_order": 253 + }, + "Resize_391": { + "ResizeAdf": { + "co_trans_mode": 1, + "dim_0": 1, + "dim_1": 80, + "dim_2": 23, + "dim_3": 40, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "mode": 1, + "nearest_mode": 0, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_H": 46, + "output_W": 80, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 23, + 40 + ], + "dims": [ + 80, + 23, + 40 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 46, + 80 + ], + "dims": [ + 80, + 46, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_389", + "type": "internal" + } + ], + "layer_order": 254 + }, + "Slice_397": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "H", + "dim_c": 80, + "dim_h": 46, + "dim_w": 80, + "dtype": "bfloat16", + "end": 45, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 46, + 80 + ], + "dims": [ + 80, + 46, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 80, + 45, + 80 + ], + "dims": [ + 80, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Resize_391", + "type": "internal" + } + ], + "layer_order": 255 + }, + "Concat_398": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 80, + 45, + 80 + ], + "dims": [ + 80, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 24, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 45, + 80 + ], + "dims": [ + 8, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 107, + 45, + 80 + ], + "dims": [ + 112, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Slice_397", + "type": "internal" + }, + { + "name": "Conv_39", + "type": "internal" + }, + { + "name": "AveragePool_347", + "type": "internal" + } + ], + "layer_order": 256 + }, + "Conv_399": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 112, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 7, + "ifm_height": 45, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 14, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 18, + "ifm_width": 80, + "ifm_x_iter": 5, + "ifm_y_iter": 1, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 7, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 40, + "ofm_depth_iter": 1, + "ofm_height": 45, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 12, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 80, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 14, + "padding_sv_width": 18, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 107, + 45, + 80 + ], + "dims": [ + 112, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_398", + "type": "internal" + } + ], + "layer_order": 257, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 7, + "ifm_depth_padded": 112, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 20027520, + "wts_no_of_reads": 3456 + } + }, + "Split_401_Duplicated#0": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 40, + "dim_h": 45, + "dim_w": 80, + "dtype": "bfloat16", + "end": 20, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 40, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_399", + "type": "internal" + } + ], + "layer_order": 258 + }, + "Split_401_Duplicated#1": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 40, + "dim_h": 45, + "dim_w": 80, + "dtype": "bfloat16", + "end": 40, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 20, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 40, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_399", + "type": "internal" + } + ], + "layer_order": 259 + }, + "Concat_402": { + "ConcatC8Adf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "in1_dim_c": 24, + "in1_dim_h": 45, + "in1_dim_w": 80, + "in2_dim_c": 24, + "in2_dim_h": 45, + "in2_dim_w": 80, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_eff_concat_input0_size": 20, + "num_eff_concat_input0_start": 0, + "num_eff_concat_input1_size": 20, + "num_eff_concat_input1_start": 0, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 40, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_401_Duplicated#1", + "type": "internal" + }, + { + "name": "Input.2", + "type": "external" + } + ], + "layer_order": 260 + }, + "Conv_403": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 0, + "data_io.ofm.dimensions.width": 16, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 40, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 3, + "ifm_height": 45, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 14, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 18, + "ifm_width": 80, + "ifm_x_iter": 5, + "ifm_y_iter": 1, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 3, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 40, + "ofm_depth_iter": 1, + "ofm_height": 45, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 16, + "ofm_sv_height": 12, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 16, + "ofm_width": 80, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 14, + "padding_sv_width": 18, + "psum_buff_offset_scaled": 1, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 40, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_402", + "type": "internal" + } + ], + "layer_order": 261, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 288, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 3, + "ifm_depth_padded": 48, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 2, + "ofm_depth_iter": 1, + "ofm_depth_padded": 64, + "output_stride": 16, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 20325888, + "wts_no_of_reads": 3456 + } + }, + "Sigmoid_404": { + "SigmoidTemplatedBf16": { + "ENABLE_FP16_AS_BF16": 0, + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 16, + "ifmsv_height": 12, + "ifmsv_size": 1920, + "ifmsv_width": 10, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 64, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_403", + "type": "internal" + } + ], + "layer_order": 262 + }, + "Split_405_Duplicated#1": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 40, + "dim_h": 45, + "dim_w": 80, + "dtype": "bfloat16", + "end": 40, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 20, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 40, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sigmoid_404", + "type": "internal" + } + ], + "layer_order": 263 + }, + "Sub_411": { + "SubBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 12, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 3600, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "constants_reader": { + "Const.4": { + "bytes_to_skip": 20764800, + "channel_stride": 8, + "depth": 1, + "dtype": "bfloat16", + "height": 45, + "nibbles": 4, + "num_reads": 5400, + "rank": 4, + "reads_per_line": 3, + "width": 80 + } + }, + "dims_and_bounds": [ + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Const.4", + "type": "constant" + }, + { + "name": "Split_405_Duplicated#1", + "type": "internal" + } + ], + "layer_order": 264 + }, + "Mul_412": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 12, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 3600, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sub_411", + "type": "internal" + }, + { + "name": "Input.2", + "type": "external" + } + ], + "layer_order": 265 + }, + "Split_405_Duplicated#0": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 40, + "dim_h": 45, + "dim_w": 80, + "dtype": "bfloat16", + "end": 20, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 40, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sigmoid_404", + "type": "internal" + } + ], + "layer_order": 266 + }, + "Mul_406": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 12, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 3600, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_405_Duplicated#0", + "type": "internal" + }, + { + "name": "Input.2", + "type": "external" + } + ], + "layer_order": 267 + }, + "Concat_407": { + "ConcatC8Adf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "in1_dim_c": 24, + "in1_dim_h": 45, + "in1_dim_w": 80, + "in2_dim_c": 24, + "in2_dim_h": 45, + "in2_dim_w": 80, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_eff_concat_input0_size": 20, + "num_eff_concat_input0_start": 0, + "num_eff_concat_input1_size": 20, + "num_eff_concat_input1_start": 0, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 40, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_401_Duplicated#1", + "type": "internal" + }, + { + "name": "Mul_406", + "type": "internal" + } + ], + "layer_order": 268 + }, + "Conv_408": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 40, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 5, + "ifm_height": 45, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_height": 14, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 34, + "ifm_width": 80, + "ifm_x_iter": 3, + "ifm_y_iter": 1, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 5, + "num_ifm_height_iters": 1, + "num_ifm_width_iters": 3, + "num_ofm_depth_iters": 1, + "ofm_depth": 24, + "ofm_depth_iter": 1, + "ofm_height": 45, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 8, + "ofm_sv_height": 12, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 80, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 14, + "padding_sv_width": 34, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 40, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 96 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_407", + "type": "internal" + } + ], + "layer_order": 269, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 5, + "ifm_depth_padded": 40, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 1, + "ofm_depth_padded": 32, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 20453760, + "wts_no_of_reads": 864 + } + }, + "Tanh_409": { + "TanhTemplatedBf16": { + "ENABLE_FP16_AS_BF16": 0, + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 12, + "ifmsv_size": 1920, + "ifmsv_width": 20, + "num_kernel_iters": 4, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 96 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_408", + "type": "internal" + } + ], + "layer_order": 270 + }, + "Mul_413": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 12, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 3600, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_405_Duplicated#1", + "type": "internal" + }, + { + "name": "Tanh_409", + "type": "internal" + } + ], + "layer_order": 271 + }, + "Add_414": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 12, + "ifmsv_size": 960, + "ifmsv_width": 10, + "num_elems": 3600, + "num_kernel_iters": 8, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 32, + 48, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "external_output_layer": 1, + "input_layer_names": [ + { + "name": "Mul_412", + "type": "internal" + }, + { + "name": "Mul_413", + "type": "internal" + } + ], + "layer_order": 272 + }, + "Concat_415": { + "ConcatC8Adf": { + "aie_arch": "aie2p", + "dtype": "bfloat16", + "in1_dim_c": 24, + "in1_dim_h": 45, + "in1_dim_w": 80, + "in2_dim_c": 24, + "in2_dim_h": 45, + "in2_dim_w": 80, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_eff_concat_input0_size": 20, + "num_eff_concat_input0_start": 0, + "num_eff_concat_input1_size": 20, + "num_eff_concat_input1_start": 0, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 20, + 45, + 80 + ], + "dims": [ + 24, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 40, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_401_Duplicated#0", + "type": "internal" + }, + { + "name": "Add_414", + "type": "internal" + } + ], + "layer_order": 273 + }, + "Resize_417": { + "ResizeAdf": { + "co_trans_mode": 1, + "dim_0": 1, + "dim_1": 40, + "dim_2": 45, + "dim_3": 80, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "mode": 1, + "nearest_mode": 0, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_H": 90, + "output_W": 160, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 45, + 80 + ], + "dims": [ + 40, + 45, + 80 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 40, + 90, + 160 + ], + "dims": [ + 40, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_415", + "type": "internal" + } + ], + "layer_order": 274 + }, + "Concat_418": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 40, + 90, + 160 + ], + "dims": [ + 40, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 90, + 160 + ], + "dims": [ + 8, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 59, + 90, + 160 + ], + "dims": [ + 64, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Resize_417", + "type": "internal" + }, + { + "name": "Conv_28", + "type": "internal" + }, + { + "name": "AveragePool_346", + "type": "internal" + } + ], + "layer_order": 275 + }, + "Conv_419": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 64, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 4, + "ifm_height": 90, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 7, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 34, + "ifm_width": 160, + "ifm_x_iter": 5, + "ifm_y_iter": 5, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 4, + "num_ifm_height_iters": 5, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 32, + "ofm_depth_iter": 1, + "ofm_height": 90, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 8, + "ofm_sv_height": 5, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 160, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 7, + "padding_sv_width": 34, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 59, + 90, + 160 + ], + "dims": [ + 64, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_418", + "type": "internal" + } + ], + "layer_order": 276, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 4, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 1, + "ofm_depth_padded": 32, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 20508480, + "wts_no_of_reads": 1728 + } + }, + "Split_421_Duplicated#0": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 32, + "dim_h": 90, + "dim_w": 160, + "dtype": "bfloat16", + "end": 16, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_419", + "type": "internal" + } + ], + "layer_order": 277 + }, + "Split_421_Duplicated#1": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 32, + "dim_h": 90, + "dim_w": 160, + "dtype": "bfloat16", + "end": 32, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 16, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_419", + "type": "internal" + } + ], + "layer_order": 278 + }, + "Concat_422": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Split_421_Duplicated#1", + "type": "internal" + }, + { + "name": "Input.1", + "type": "external" + } + ], + "layer_order": 279 + }, + "Conv_423": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 32, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 2, + "ifm_height": 90, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 6, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 34, + "ifm_width": 160, + "ifm_x_iter": 5, + "ifm_y_iter": 6, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 6, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 32, + "ofm_depth_iter": 1, + "ofm_height": 90, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 8, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 160, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 6, + "padding_sv_width": 34, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_422", + "type": "internal" + } + ], + "layer_order": 280, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 32, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 1, + "ofm_depth_padded": 32, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 20593728, + "wts_no_of_reads": 1728 + } + }, + "Sigmoid_424": { + "SigmoidTemplatedBf16": { + "ENABLE_FP16_AS_BF16": 0, + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 8, + "ifmsv_size": 2048, + "ifmsv_width": 32, + "num_kernel_iters": 15, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_423", + "type": "internal" + } + ], + "layer_order": 281 + }, + "Split_425_Duplicated#1": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 32, + "dim_h": 90, + "dim_w": 160, + "dtype": "bfloat16", + "end": 32, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 16, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sigmoid_424", + "type": "internal" + } + ], + "layer_order": 282 + }, + "Sub_431": { + "SubBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 8, + "ifmsv_size": 1024, + "ifmsv_width": 16, + "num_elems": 14400, + "num_kernel_iters": 30, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "constants_reader": { + "Const.5": { + "bytes_to_skip": 20764800, + "channel_stride": 8, + "depth": 1, + "dtype": "bfloat16", + "height": 90, + "nibbles": 4, + "num_reads": 21600, + "rank": 4, + "reads_per_line": 3, + "width": 160 + } + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Const.5", + "type": "constant" + }, + { + "name": "Split_425_Duplicated#1", + "type": "internal" + } + ], + "layer_order": 283 + }, + "Mul_432": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 8, + "ifmsv_size": 1024, + "ifmsv_width": 16, + "num_elems": 14400, + "num_kernel_iters": 30, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sub_431", + "type": "internal" + }, + { + "name": "Input.1", + "type": "external" + } + ], + "layer_order": 284 + }, + "Split_425_Duplicated#0": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 32, + "dim_h": 90, + "dim_w": 160, + "dtype": "bfloat16", + "end": 16, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Sigmoid_424", + "type": "internal" + } + ], + "layer_order": 285 + }, + "Mul_426": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 8, + "ifmsv_size": 1024, + "ifmsv_width": 16, + "num_elems": 14400, + "num_kernel_iters": 30, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_425_Duplicated#0", + "type": "internal" + }, + { + "name": "Input.1", + "type": "external" + } + ], + "layer_order": 286 + }, + "Concat_427": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Split_421_Duplicated#1", + "type": "internal" + }, + { + "name": "Mul_426", + "type": "internal" + } + ], + "layer_order": 287 + }, + "Conv_428": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 32, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 2, + "ifm_height": 90, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 6, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 34, + "ifm_width": 160, + "ifm_x_iter": 5, + "ifm_y_iter": 6, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 2, + "num_ifm_height_iters": 6, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 16, + "ofm_depth_iter": 1, + "ofm_height": 90, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 8, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 160, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 6, + "padding_sv_width": 34, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_427", + "type": "internal" + } + ], + "layer_order": 288, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 2, + "ifm_depth_padded": 32, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 1, + "ofm_depth_padded": 32, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 20636352, + "wts_no_of_reads": 1728 + } + }, + "Tanh_429": { + "TanhTemplatedBf16": { + "ENABLE_FP16_AS_BF16": 0, + "aie_arch": "aie2p", + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 23, + "ifmsv_size": 1472, + "ifmsv_width": 8, + "num_kernel_iters": 20, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 32, + 92, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_428", + "type": "internal" + } + ], + "layer_order": 289 + }, + "Mul_433": { + "MulBf16": { + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 8, + "ifmsv_size": 1024, + "ifmsv_width": 16, + "num_elems": 14400, + "num_kernel_iters": 30, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_425_Duplicated#1", + "type": "internal" + }, + { + "name": "Tanh_429", + "type": "internal" + } + ], + "layer_order": 290 + }, + "Add_434": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 8, + "ifmsv_size": 1024, + "ifmsv_width": 16, + "num_elems": 14400, + "num_kernel_iters": 30, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "external_output_layer": 0, + "input_layer_names": [ + { + "name": "Mul_432", + "type": "internal" + }, + { + "name": "Mul_433", + "type": "internal" + } + ], + "layer_order": 291 + }, + "Concat_435": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 90, + 160 + ], + "dims": [ + 16, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Split_421_Duplicated#0", + "type": "internal" + }, + { + "name": "Add_434", + "type": "internal" + } + ], + "layer_order": 292 + }, + "Resize_437": { + "ResizeAdf": { + "co_trans_mode": 1, + "dim_0": 1, + "dim_1": 32, + "dim_2": 90, + "dim_3": 160, + "dtype": "bfloat16", + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "mode": 1, + "nearest_mode": 0, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_H": 180, + "output_W": 320, + "output_l3_format": "HCWN_C8" + }, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 90, + 160 + ], + "dims": [ + 32, + 90, + 160 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 32, + 180, + 320 + ], + "dims": [ + 32, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_435", + "type": "internal" + } + ], + "layer_order": 293 + }, + "Concat_438": { + "Concat": { + "axis": 1, + "dims_and_bounds": [ + { + "bounds": [ + 32, + 180, + 320 + ], + "dims": [ + 32, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 35, + 180, + 320 + ], + "dims": [ + 40, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "dtype": "bfloat16" + }, + "input_layer_names": [ + { + "name": "Resize_437", + "type": "internal" + }, + { + "name": "Generated-#4", + "type": "internal" + } + ], + "layer_order": 294 + }, + "Conv_439": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 64, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 40, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 5, + "ifm_height": 180, + "ifm_sign": 0, + "ifm_sv_depth": 8, + "ifm_sv_height": 6, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 66, + "ifm_width": 320, + "ifm_x_iter": 5, + "ifm_y_iter": 12, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 5, + "num_ifm_height_iters": 12, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 16, + "ofm_depth_iter": 1, + "ofm_height": 180, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 8, + "ofm_sv_height": 4, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 64, + "ofm_width": 320, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 6, + "padding_sv_width": 66, + "psum_buff_offset_scaled": 4, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 35, + 180, + 320 + ], + "dims": [ + 40, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 180, + 320 + ], + "dims": [ + 16, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Concat_438", + "type": "internal" + } + ], + "layer_order": 295, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 5, + "ifm_depth_padded": 40, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 1, + "ofm_depth_padded": 32, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 20678976, + "wts_no_of_reads": 864 + } + }, + "Conv_441": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 1, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 32, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 1, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 16, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 180, + "ifm_sign": 0, + "ifm_sv_depth": 16, + "ifm_sv_height": 7, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 34, + "ifm_width": 320, + "ifm_x_iter": 10, + "ifm_y_iter": 9, + "kernel_height": 3, + "kernel_width": 3, + "ksize.height": 3, + "ksize.width": 3, + "lrelu_alpha": 0, + "lrelu_alpha_kernel": 0, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 9, + "num_ifm_width_iters": 10, + "num_ofm_depth_iters": 1, + "ofm_depth": 16, + "ofm_depth_iter": 1, + "ofm_height": 180, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 8, + "ofm_sv_height": 5, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 32, + "ofm_width": 320, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 1, + "pad_left": 1, + "pad_right": 1, + "pad_top": 1, + "pad_value": 0, + "padding_sv_height": 7, + "padding_sv_width": 34, + "psum_buff_offset_scaled": 2, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 180, + 320 + ], + "dims": [ + 16, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 16, + 180, + 320 + ], + "dims": [ + 16, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_439", + "type": "internal" + } + ], + "layer_order": 296, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 16, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 3, + "kernel_width": 3, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 1, + "ofm_depth_padded": 32, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 20733696, + "wts_no_of_reads": 1728 + } + }, + "Conv_443": { + "Conv2DBf16": { + "AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16": 1, + "act": 0, + "act_type": "RELU", + "aie_arch": "aie2p", + "alpha": 0, + "alpha_scaled": 3277, + "batch_size": 1, + "compiler": "chess", + "conv2d_ofm_pad": 0, + "conv_type": 64, + "data_io.ofm.dimensions.width": 64, + "dtype_ifm": "bfloat16", + "dtype_ofm": "bfloat16", + "dtype_wts": "bfloat16", + "enable_padding": 0, + "force_och_gran_8": 1, + "func_type": 0, + "ifm_depth": 16, + "ifm_depth_concat_extend": 0, + "ifm_depth_iter": 1, + "ifm_height": 180, + "ifm_sign": 0, + "ifm_sv_depth": 64, + "ifm_sv_height": 1, + "ifm_sv_height.padding": 0, + "ifm_sv_width": 64, + "ifm_width": 320, + "ifm_x_iter": 5, + "ifm_y_iter": 45, + "kernel_height": 1, + "kernel_width": 1, + "ksize.height": 1, + "ksize.width": 1, + "lrelu_alpha": 1, + "lrelu_alpha_kernel": 1, + "num_adf_cols": 4, + "num_adf_rows": 4, + "num_ifm_depth_iters": 1, + "num_ifm_height_iters": 45, + "num_ifm_width_iters": 5, + "num_ofm_depth_iters": 1, + "ofm_depth": 8, + "ofm_depth_iter": 1, + "ofm_height": 180, + "ofm_offset_packed.left": 0, + "ofm_offset_packed.top": 0, + "ofm_sv_depth": 8, + "ofm_sv_height": 1, + "ofm_sv_overlap_height": 0, + "ofm_sv_overlap_width": 0, + "ofm_sv_width": 64, + "ofm_width": 320, + "ofm_width_pad": 0, + "op_type": 0, + "out_mode": 0, + "pad_bottom": 0, + "pad_left": 0, + "pad_right": 0, + "pad_top": 0, + "pad_value": 0, + "padding_sv_height": 1, + "padding_sv_width": 64, + "psum_buff_offset_scaled": 4, + "relu_int8_enable": 0, + "shift_bias_init": 0, + "shift_lrelu_alpha": 0, + "shift_lrelu_in": 0, + "shift_lrelu_out": 0, + "shift_out": 0, + "shift_psum": 0, + "shift_psum_in": 0, + "signed_value": 0, + "stride_h": 1, + "stride_height": 1, + "stride_log2.stride_log2_height": 0, + "stride_log2.stride_log2_width": 0, + "stride_w": 1, + "stride_width": 1, + "zp_en": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 16, + 180, + 320 + ], + "dims": [ + 16, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 4, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_441", + "type": "internal" + } + ], + "layer_order": 297, + "split_width": false, + "wts_reader": { + "BFP16MMUL": true, + "bias_bytes_to_skip": 144, + "bias_rep": 4, + "bias_sg_bytes_to_skip": 36, + "bias_sub_group_num_reads": 12, + "cstride": 8, + "ifm_depth_iter": 1, + "ifm_depth_padded": 64, + "is_dwc": false, + "is_zero_padded": false, + "istride": 8, + "kernel_height": 1, + "kernel_width": 1, + "krnl_w_div": 1, + "num_bias_sub_groups": 1, + "ofm_depth_iter": 1, + "ofm_depth_padded": 32, + "output_stride": 8, + "rtp_bytes_to_skip": 144, + "total_layer_bytes_to_skip": 20755008, + "wts_no_of_reads": 768 + } + }, + "Split_444_Duplicated#1": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 8, + "dim_h": 180, + "dim_w": 320, + "dtype": "bfloat16", + "end": 4, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 3, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 4, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 1, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_443", + "type": "internal" + } + ], + "layer_order": 298 + }, + "Clip_447": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 1, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 5, + "ifmsv_size": 1600, + "ifmsv_width": 40, + "num_kernel_iters": 72, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 1, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 1, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "external_output_layer": 5, + "input_layer_names": [ + { + "name": "Split_444_Duplicated#1", + "type": "internal" + } + ], + "layer_order": 299 + }, + "Split_444_Duplicated#0": { + "SliceHCWC8Adf": { + "aie_arch": "aie2p", + "axis_letter": "C", + "dim_c": 8, + "dim_h": 180, + "dim_w": 320, + "dtype": "bfloat16", + "end": 3, + "input_l3_format": "HCWN_C8", + "kernel_col": 4, + "num_ifm_shim_ch": 2, + "num_ofm_shim_ch": 2, + "output_l3_format": "HCWN_C8", + "start": 0, + "step": 1 + }, + "dims_and_bounds": [ + { + "bounds": [ + 4, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Conv_443", + "type": "internal" + } + ], + "layer_order": 300 + }, + "Add_445": { + "AddBf16": { + "act": 0, + "act_type": "LINEAR", + "aie_arch": "aie2p", + "compiler": "chess", + "dtype": "bfloat16", + "ifm1_shift": 0, + "ifm1_shift_biased": 0, + "ifm2_shift": 0, + "ifm2_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 5, + "ifmsv_size": 640, + "ifmsv_width": 16, + "num_elems": 57600, + "num_kernel_iters": 180, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "input_layer_names": [ + { + "name": "Split_444_Duplicated#0", + "type": "internal" + }, + { + "name": "Generated-#4", + "type": "internal" + } + ], + "layer_order": 301 + }, + "Clip_446": { + "ClipBf16": { + "aie_arch": "aie2p", + "clamp_max": 1, + "clamp_min": 0, + "compiler": "chess", + "ifm_shift": 0, + "ifm_shift_biased": 0, + "ifmsv_depth": 8, + "ifmsv_height": 5, + "ifmsv_size": 1600, + "ifmsv_width": 40, + "num_kernel_iters": 72, + "ofm_shift": 0, + "ofm_shift_biased": 0 + }, + "dims_and_bounds": [ + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + }, + { + "bounds": [ + 3, + 180, + 320 + ], + "dims": [ + 8, + 180, + 320 + ], + "dtype": "bfloat16", + "keys": [ + "depth", + "height", + "width" + ], + "padding_at_start": { + "cin": 0, + "cout": 0, + "height": 0, + "width": 0 + } + } + ], + "external_output_layer": 4, + "input_layer_names": [ + { + "name": "Add_445", + "type": "internal" + } + ], + "layer_order": 302 + } +} diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/unified-4x4.xclbin b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/unified-4x4.xclbin new file mode 100644 index 0000000000000000000000000000000000000000..99e47692029ea7d016e569c32eff828f7d049661 Binary files /dev/null and b/segmentation_1_4_0_fp32_combined/vaiml_par_0/0/unified-4x4.xclbin differ diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel-codegen-wrappers.c b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel-codegen-wrappers.c new file mode 100644 index 0000000000000000000000000000000000000000..c8f55228fb317bcbd1de8686e6871e7e7a02e7e4 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel-codegen-wrappers.c @@ -0,0 +1,659 @@ +#include +#include +#define MAX_VAIML_PARTITION_OUTPUTS 100 + +#ifdef _WIN32 + #define FLEXML_WRAPPER_IMPORT __declspec(dllimport) + #define FLEXML_WRAPPER_EXPORT __declspec(dllexport) +#else + #define FLEXML_WRAPPER_IMPORT + #define FLEXML_WRAPPER_EXPORT +#endif + +FLEXML_WRAPPER_IMPORT extern void *malloc(uint64_t); +FLEXML_WRAPPER_IMPORT extern void free(void *ptr); +FLEXML_WRAPPER_IMPORT extern void* memcpy(void *dest, const void *src, uint64_t count); +FLEXML_WRAPPER_IMPORT extern void flexml_rt_executeRunner(uint64_t modelId, unsigned clusterId, void *inbuf, unsigned long inbufNumElem, + void *outbuf, unsigned long outbufNumElem); +FLEXML_WRAPPER_IMPORT extern void flexml_rt_executeRunnerNew(uint64_t modelId, unsigned clusterId, const char *serializedArgs); +FLEXML_WRAPPER_IMPORT extern char *flexml_rt_paramLoader(uint64_t modelId, const char *file, const char *key, const char *transform); + + +int is_in_set(void *set[], int set_size, void *ptr) { + for (int i = 0; i < set_size; i++) { + if (set[i] == ptr) { + return 1; + } + } + return 0; +} + +void add_to_set(void *set[], int *set_size, void *ptr) { + if (!is_in_set(set, *set_size, ptr)) { + set[*set_size] = ptr; + (*set_size)++; + } +} + +void free_set(void *set[], int set_size) { + for (int i = 0; i < set_size; i++) { + free(set[i]); + } +} + +FLEXML_WRAPPER_EXPORT uint64_t OnnxModel_modelId = 0; + +typedef struct Flexml_4D_Tensor_ { + void *allocPtr; + void *dataPtr; + int64_t offset; + int64_t dimSizes[4]; + int64_t dimStrides[4]; +} Flexml_4D_Tensor; + +typedef struct Flexml_4D4D4D4D4D4D_Tensor_ { + Flexml_4D_Tensor val_0; + Flexml_4D_Tensor val_1; + Flexml_4D_Tensor val_2; + Flexml_4D_Tensor val_3; + Flexml_4D_Tensor val_4; + Flexml_4D_Tensor val_5; +} Flexml_4D4D4D4D4D4D_Tensor; + +extern Flexml_4D4D4D4D4D4D_Tensor forward(void *allocPtr_0, void *dataPtr_0, int64_t offset_0, int64_t dim0Size_0, int64_t dim1Size_0, int64_t dim2Size_0, int64_t dim3Size_0, int64_t dim0Stride_0, int64_t dim1Stride_0, int64_t dim2Stride_0, int64_t dim3Stride_0 +, void *allocPtr_1, void *dataPtr_1, int64_t offset_1, int64_t dim0Size_1, int64_t dim1Size_1, int64_t dim2Size_1, int64_t dim3Size_1, int64_t dim0Stride_1, int64_t dim1Stride_1, int64_t dim2Stride_1, int64_t dim3Stride_1 +, void *allocPtr_2, void *dataPtr_2, int64_t offset_2, int64_t dim0Size_2, int64_t dim1Size_2, int64_t dim2Size_2, int64_t dim3Size_2, int64_t dim0Stride_2, int64_t dim1Stride_2, int64_t dim2Stride_2, int64_t dim3Stride_2 +, void *allocPtr_3, void *dataPtr_3, int64_t offset_3, int64_t dim0Size_3, int64_t dim1Size_3, int64_t dim2Size_3, int64_t dim3Size_3, int64_t dim0Stride_3, int64_t dim1Stride_3, int64_t dim2Stride_3, int64_t dim3Stride_3 +, void *allocPtr_4, void *dataPtr_4, int64_t offset_4, int64_t dim0Size_4, int64_t dim1Size_4, int64_t dim2Size_4, int64_t dim3Size_4, int64_t dim0Stride_4, int64_t dim1Stride_4, int64_t dim2Stride_4, int64_t dim3Stride_4 +); + +FLEXML_WRAPPER_EXPORT void *flexml_OnnxModel_model(const float *ifm) { + return 0; +} + +FLEXML_WRAPPER_EXPORT int flexml_OnnxModel_modelNew(uint64_t modelId, const char *args) { +OnnxModel_modelId = modelId; + unsigned offset = 0; + if (*(unsigned long long *) (args + offset) != 5) + return 3; + offset += sizeof(unsigned long long); + void *inputPtrs[5]; + inputPtrs[0] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 180) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 320) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 4) + return 1; + offset += sizeof(unsigned long long); + inputPtrs[1] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 16) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 90) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 160) + return 1; + offset += sizeof(unsigned long long); + inputPtrs[2] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 20) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 45) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 80) + return 1; + offset += sizeof(unsigned long long); + inputPtrs[3] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 40) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 23) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 40) + return 1; + offset += sizeof(unsigned long long); + inputPtrs[4] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 64) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 12) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 20) + return 1; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 6) + return 4; + offset += sizeof(unsigned long long); + void *outputPtrs[6]; + outputPtrs[0] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 180) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 320) + return 2; + offset += sizeof(unsigned long long); + outputPtrs[1] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 16) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 90) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 160) + return 2; + offset += sizeof(unsigned long long); + outputPtrs[2] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 20) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 45) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 80) + return 2; + offset += sizeof(unsigned long long); + outputPtrs[3] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 40) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 23) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 40) + return 2; + offset += sizeof(unsigned long long); + outputPtrs[4] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 64) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 12) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 20) + return 2; + offset += sizeof(unsigned long long); + outputPtrs[5] = *(void **) (args + offset); + offset += sizeof(void *); + if (*(unsigned long long *) (args + offset) != 4) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 1) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 3) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 180) + return 2; + offset += sizeof(unsigned long long); + if (*(unsigned long long *) (args + offset) != 320) + return 2; + offset += sizeof(unsigned long long); + Flexml_4D4D4D4D4D4D_Tensor retVal = forward(inputPtrs[0], inputPtrs[0], 0, 1, 180, 320, 4, 230400, 1280, 4, 1, inputPtrs[1], inputPtrs[1], 0, 1, 16, 90, 160, 230400, 14400, 160, 1, inputPtrs[2], inputPtrs[2], 0, 1, 20, 45, 80, 72000, 3600, 80, 1, inputPtrs[3], inputPtrs[3], 0, 1, 40, 23, 40, 36800, 920, 40, 1, inputPtrs[4], inputPtrs[4], 0, 1, 64, 12, 20, 15360, 240, 20, 1); + + void *unique_ptrs[MAX_VAIML_PARTITION_OUTPUTS]; + int unique_ptrs_size = 0; + (void) memcpy(outputPtrs[0], retVal.val_0.dataPtr, 230400); + add_to_set(unique_ptrs, &unique_ptrs_size, retVal.val_0.allocPtr); + (void) memcpy(outputPtrs[1], retVal.val_1.dataPtr, 921600); + add_to_set(unique_ptrs, &unique_ptrs_size, retVal.val_1.allocPtr); + (void) memcpy(outputPtrs[2], retVal.val_2.dataPtr, 288000); + add_to_set(unique_ptrs, &unique_ptrs_size, retVal.val_2.allocPtr); + (void) memcpy(outputPtrs[3], retVal.val_3.dataPtr, 147200); + add_to_set(unique_ptrs, &unique_ptrs_size, retVal.val_3.allocPtr); + (void) memcpy(outputPtrs[4], retVal.val_4.dataPtr, 61440); + add_to_set(unique_ptrs, &unique_ptrs_size, retVal.val_4.allocPtr); + (void) memcpy(outputPtrs[5], retVal.val_5.dataPtr, 691200); + add_to_set(unique_ptrs, &unique_ptrs_size, retVal.val_5.allocPtr); + free_set(unique_ptrs, unique_ptrs_size); + return 0; +} + +Flexml_4D4D4D4D4D4D_Tensor +forward_outlined_part_0(void *allocPtr_0, void *dataPtr_0, int64_t offset_0, int64_t dim0Size_0, int64_t dim1Size_0, int64_t dim2Size_0, int64_t dim3Size_0, int64_t dim0Stride_0, int64_t dim1Stride_0, int64_t dim2Stride_0, int64_t dim3Stride_0 +, void *allocPtr_1, void *dataPtr_1, int64_t offset_1, int64_t dim0Size_1, int64_t dim1Size_1, int64_t dim2Size_1, int64_t dim3Size_1, int64_t dim0Stride_1, int64_t dim1Stride_1, int64_t dim2Stride_1, int64_t dim3Stride_1 +, void *allocPtr_2, void *dataPtr_2, int64_t offset_2, int64_t dim0Size_2, int64_t dim1Size_2, int64_t dim2Size_2, int64_t dim3Size_2, int64_t dim0Stride_2, int64_t dim1Stride_2, int64_t dim2Stride_2, int64_t dim3Stride_2 +, void *allocPtr_3, void *dataPtr_3, int64_t offset_3, int64_t dim0Size_3, int64_t dim1Size_3, int64_t dim2Size_3, int64_t dim3Size_3, int64_t dim0Stride_3, int64_t dim1Stride_3, int64_t dim2Stride_3, int64_t dim3Stride_3 +, void *allocPtr_4, void *dataPtr_4, int64_t offset_4, int64_t dim0Size_4, int64_t dim1Size_4, int64_t dim2Size_4, int64_t dim3Size_4, int64_t dim0Stride_4, int64_t dim1Stride_4, int64_t dim2Stride_4, int64_t dim3Stride_4 +) +{ + void *outbuf_0 = malloc(983040); + void *outbuf_1 = malloc(491520); + void *outbuf_2 = malloc(294912); + void *outbuf_3 = malloc(131072); + void *outbuf_4 = malloc(3932160); + void *outbuf_5 = malloc(245760); + char args[544]; + unsigned offset = 0; + *(unsigned long long *) (args + offset) = 5; + offset += sizeof(unsigned long long); + *(void **) (args + offset) = (void *) dataPtr_0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 180; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 320; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) dataPtr_1; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 16; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 90; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 160; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) dataPtr_2; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 20; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 45; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 80; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) dataPtr_3; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 40; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 23; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 40; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) dataPtr_4; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 64; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 12; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 20; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 6; + offset += sizeof(unsigned long long); + *(void **) (args + offset) = (void *) outbuf_0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 16; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 90; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 160; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) outbuf_1; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 20; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 45; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 80; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) outbuf_2; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 40; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 23; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 40; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) outbuf_3; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 64; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 12; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 20; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) outbuf_4; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 3; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 180; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 320; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) outbuf_5; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 180; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 320; + offset += sizeof(unsigned long long *); + flexml_rt_executeRunnerNew(/* modelId */ OnnxModel_modelId, /* clusterId */ 0, args); + Flexml_4D4D4D4D4D4D_Tensor result; + result.val_0.allocPtr = result.val_0.dataPtr = outbuf_0; + result.val_0.offset = 0; + result.val_0.dimSizes[0] = 1; + result.val_0.dimSizes[1] = 16; + result.val_0.dimSizes[2] = 90; + result.val_0.dimSizes[3] = 160; + result.val_0.dimStrides[0] = 230400; + result.val_0.dimStrides[1] = 14400; + result.val_0.dimStrides[2] = 160; + result.val_0.dimStrides[3] = 1; + result.val_1.allocPtr = result.val_1.dataPtr = outbuf_1; + result.val_1.offset = 0; + result.val_1.dimSizes[0] = 1; + result.val_1.dimSizes[1] = 20; + result.val_1.dimSizes[2] = 45; + result.val_1.dimSizes[3] = 80; + result.val_1.dimStrides[0] = 72000; + result.val_1.dimStrides[1] = 3600; + result.val_1.dimStrides[2] = 80; + result.val_1.dimStrides[3] = 1; + result.val_2.allocPtr = result.val_2.dataPtr = outbuf_2; + result.val_2.offset = 0; + result.val_2.dimSizes[0] = 1; + result.val_2.dimSizes[1] = 40; + result.val_2.dimSizes[2] = 23; + result.val_2.dimSizes[3] = 40; + result.val_2.dimStrides[0] = 36800; + result.val_2.dimStrides[1] = 920; + result.val_2.dimStrides[2] = 40; + result.val_2.dimStrides[3] = 1; + result.val_3.allocPtr = result.val_3.dataPtr = outbuf_3; + result.val_3.offset = 0; + result.val_3.dimSizes[0] = 1; + result.val_3.dimSizes[1] = 64; + result.val_3.dimSizes[2] = 12; + result.val_3.dimSizes[3] = 20; + result.val_3.dimStrides[0] = 15360; + result.val_3.dimStrides[1] = 240; + result.val_3.dimStrides[2] = 20; + result.val_3.dimStrides[3] = 1; + result.val_4.allocPtr = result.val_4.dataPtr = outbuf_4; + result.val_4.offset = 0; + result.val_4.dimSizes[0] = 1; + result.val_4.dimSizes[1] = 3; + result.val_4.dimSizes[2] = 180; + result.val_4.dimSizes[3] = 320; + result.val_4.dimStrides[0] = 172800; + result.val_4.dimStrides[1] = 57600; + result.val_4.dimStrides[2] = 320; + result.val_4.dimStrides[3] = 1; + result.val_5.allocPtr = result.val_5.dataPtr = outbuf_5; + result.val_5.offset = 0; + result.val_5.dimSizes[0] = 1; + result.val_5.dimSizes[1] = 1; + result.val_5.dimSizes[2] = 180; + result.val_5.dimSizes[3] = 320; + result.val_5.dimStrides[0] = 57600; + result.val_5.dimStrides[1] = 57600; + result.val_5.dimStrides[2] = 320; + result.val_5.dimStrides[3] = 1; + return result; +} + +FLEXML_WRAPPER_EXPORT unsigned long long flexml_part_0_getArgsSize(void) { + return 544; +} + +FLEXML_WRAPPER_EXPORT unsigned long long flexml_OnnxModel_getArgsSize(void) { + return 544; +} + +FLEXML_WRAPPER_EXPORT void flexml_OnnxModel_getArgs(char *args) { + unsigned offset = 0; + *(unsigned long long *) (args + offset) = 5; + offset += sizeof(unsigned long long); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 180; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 320; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 16; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 90; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 160; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 20; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 45; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 80; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 40; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 23; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 40; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 64; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 12; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 20; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 6; + offset += sizeof(unsigned long long); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 180; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 320; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 16; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 90; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 160; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 20; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 45; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 80; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 40; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 23; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 40; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 64; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 12; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 20; + offset += sizeof(unsigned long long *); + *(void **) (args + offset) = (void *) 0; + offset += sizeof(void *); + *(unsigned long long *) (args + offset) = 4; + offset += sizeof(unsigned long long); + *(unsigned long long *) (args + offset) = 1; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 3; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 180; + offset += sizeof(unsigned long long *); + *(unsigned long long *) (args + offset) = 320; + offset += sizeof(unsigned long long *); +} + +void flexml_GraphModule_useRuntimeLib() { + flexml_rt_executeRunner(0, 0, (void *)0, 0, (void *)0, 0); +} diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel-codegen-wrappers.o b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel-codegen-wrappers.o new file mode 100644 index 0000000000000000000000000000000000000000..405094acf5fc0a5cac6c3b946c410106e9743a81 Binary files /dev/null and b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel-codegen-wrappers.o differ diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.chained_kernel.tosa.mlir b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.chained_kernel.tosa.mlir new file mode 100644 index 0000000000000000000000000000000000000000..02d6839f09491af288313ecb3a9314fd3d684b23 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.chained_kernel.tosa.mlir @@ -0,0 +1,24899 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func @forward(%arg0: tensor<1x180x320x4xbf16> {onnx.name = "385"} loc(unknown), %arg1: tensor<1x16x90x160xbf16> {onnx.name = "394"} loc(unknown), %arg2: tensor<1x20x45x80xbf16> {onnx.name = "395"} loc(unknown), %arg3: tensor<1x40x23x40xbf16> {onnx.name = "396"} loc(unknown), %arg4: tensor<1x64x12x20xbf16> {onnx.name = "397"} loc(unknown)) -> (tensor<1x1x180x320xbf16> {onnx.name = "921"}, tensor<1x16x90x160xbf16> {onnx.name = "894"}, tensor<1x20x45x80xbf16> {onnx.name = "868"}, tensor<1x40x23x40xbf16> {onnx.name = "832"}, tensor<1x64x12x20xbf16> {onnx.name = "796"}, tensor<1x3x180x320xbf16> {onnx.name = "916"}) { + %0 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/biases"} -> tensor<4xbf16> loc(#loc) + %1 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/weights"} -> tensor<4x16x1x1xbf16> loc(#loc) + %2 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/biases"} -> tensor<16xbf16> loc(#loc) + %3 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/weights"} -> tensor<16x16x3x3xbf16> loc(#loc) + %4 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/biases"} -> tensor<16xbf16> loc(#loc) + %5 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/weights"} -> tensor<16x35x3x3xbf16> loc(#loc) + %6 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_431/Constant_0_0"} -> tensor<1x16x90x160xbf16> loc(#loc1) + %7 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/biases"} -> tensor<16xbf16> loc(#loc) + %8 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/weights"} -> tensor<16x32x3x3xbf16> loc(#loc) + %9 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/biases"} -> tensor<32xbf16> loc(#loc) + %10 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/weights"} -> tensor<32x32x3x3xbf16> loc(#loc) + %11 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/biases"} -> tensor<32xbf16> loc(#loc) + %12 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/weights"} -> tensor<32x59x3x3xbf16> loc(#loc) + %13 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_411/Constant_0_0"} -> tensor<1x20x45x80xbf16> loc(#loc2) + %14 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/biases"} -> tensor<20xbf16> loc(#loc) + %15 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/weights"} -> tensor<20x40x3x3xbf16> loc(#loc) + %16 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/biases"} -> tensor<40xbf16> loc(#loc) + %17 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/weights"} -> tensor<40x40x3x3xbf16> loc(#loc) + %18 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/biases"} -> tensor<40xbf16> loc(#loc) + %19 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/weights"} -> tensor<40x107x3x3xbf16> loc(#loc) + %20 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_385/Constant_0_0"} -> tensor<1x40x23x40xbf16> loc(#loc3) + %21 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/biases"} -> tensor<40xbf16> loc(#loc) + %22 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/weights"} -> tensor<40x80x3x3xbf16> loc(#loc) + %23 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/biases"} -> tensor<80xbf16> loc(#loc) + %24 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/weights"} -> tensor<80x80x3x3xbf16> loc(#loc) + %25 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/biases"} -> tensor<80xbf16> loc(#loc) + %26 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/weights"} -> tensor<80x171x3x3xbf16> loc(#loc) + %27 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_359/Constant_0_0"} -> tensor<1x64x12x20xbf16> loc(#loc4) + %28 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/biases"} -> tensor<64xbf16> loc(#loc) + %29 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/weights"} -> tensor<64x128x3x3xbf16> loc(#loc) + %30 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/biases"} -> tensor<128xbf16> loc(#loc) + %31 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/weights"} -> tensor<128x128x3x3xbf16> loc(#loc) + %32 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/biases"} -> tensor<128xbf16> loc(#loc) + %33 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %34 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/biases"} -> tensor<128xbf16> loc(#loc) + %35 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %36 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/biases"} -> tensor<960xbf16> loc(#loc) + %37 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %38 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/biases"} -> tensor<160xbf16> loc(#loc) + %39 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %40 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/biases"} -> tensor<960xbf16> loc(#loc) + %41 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %42 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/biases"} -> tensor<240xbf16> loc(#loc) + %43 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %44 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/biases"} -> tensor<960xbf16> loc(#loc) + %45 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %46 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/biases"} -> tensor<960xbf16> loc(#loc) + %47 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %48 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/biases"} -> tensor<160xbf16> loc(#loc) + %49 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %50 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/biases"} -> tensor<960xbf16> loc(#loc) + %51 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %52 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/biases"} -> tensor<240xbf16> loc(#loc) + %53 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %54 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/biases"} -> tensor<960xbf16> loc(#loc) + %55 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %56 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/biases"} -> tensor<960xbf16> loc(#loc) + %57 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %58 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/biases"} -> tensor<160xbf16> loc(#loc) + %59 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/weights"} -> tensor<160x672x1x1xbf16> loc(#loc) + %60 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/biases"} -> tensor<672xbf16> loc(#loc) + %61 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %62 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/biases"} -> tensor<168xbf16> loc(#loc) + %63 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %64 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/biases"} -> tensor<672xbf16> loc(#loc) + %65 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/weights"} -> tensor<672x1x9x9xbf16> loc(#loc) + %66 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/biases"} -> tensor<672xbf16> loc(#loc) + %67 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %68 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/biases"} -> tensor<112xbf16> loc(#loc) + %69 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/weights"} -> tensor<112x672x1x1xbf16> loc(#loc) + %70 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/biases"} -> tensor<672xbf16> loc(#loc) + %71 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %72 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/biases"} -> tensor<168xbf16> loc(#loc) + %73 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %74 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/biases"} -> tensor<672xbf16> loc(#loc) + %75 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/weights"} -> tensor<672x1x3x3xbf16> loc(#loc) + %76 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/biases"} -> tensor<672xbf16> loc(#loc) + %77 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %78 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/biases"} -> tensor<112xbf16> loc(#loc) + %79 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/weights"} -> tensor<112x480x1x1xbf16> loc(#loc) + %80 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/biases"} -> tensor<480xbf16> loc(#loc) + %81 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/weights"} -> tensor<480x120x1x1xbf16> loc(#loc) + %82 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/biases"} -> tensor<120xbf16> loc(#loc) + %83 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/weights"} -> tensor<120x480x1x1xbf16> loc(#loc) + %84 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/biases"} -> tensor<480xbf16> loc(#loc) + %85 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/weights"} -> tensor<480x1x3x3xbf16> loc(#loc) + %86 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/biases"} -> tensor<480xbf16> loc(#loc) + %87 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/weights"} -> tensor<480x80x1x1xbf16> loc(#loc) + %88 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/biases"} -> tensor<80xbf16> loc(#loc) + %89 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %90 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/biases"} -> tensor<184xbf16> loc(#loc) + %91 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %92 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/biases"} -> tensor<184xbf16> loc(#loc) + %93 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %94 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/biases"} -> tensor<80xbf16> loc(#loc) + %95 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %96 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/biases"} -> tensor<184xbf16> loc(#loc) + %97 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %98 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/biases"} -> tensor<184xbf16> loc(#loc) + %99 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %100 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/biases"} -> tensor<80xbf16> loc(#loc) + %101 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/weights"} -> tensor<80x200x1x1xbf16> loc(#loc) + %102 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/biases"} -> tensor<200xbf16> loc(#loc) + %103 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/weights"} -> tensor<200x1x3x3xbf16> loc(#loc) + %104 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/biases"} -> tensor<200xbf16> loc(#loc) + %105 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/weights"} -> tensor<200x80x1x1xbf16> loc(#loc) + %106 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/biases"} -> tensor<80xbf16> loc(#loc) + %107 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/weights"} -> tensor<80x240x1x1xbf16> loc(#loc) + %108 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/biases"} -> tensor<240xbf16> loc(#loc) + %109 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/weights"} -> tensor<240x1x3x3xbf16> loc(#loc) + %110 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/biases"} -> tensor<240xbf16> loc(#loc) + %111 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/weights"} -> tensor<240x40x1x1xbf16> loc(#loc) + %112 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/biases"} -> tensor<40xbf16> loc(#loc) + %113 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %114 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/biases"} -> tensor<120xbf16> loc(#loc) + %115 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %116 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/biases"} -> tensor<32xbf16> loc(#loc) + %117 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %118 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/biases"} -> tensor<120xbf16> loc(#loc) + %119 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %120 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/biases"} -> tensor<120xbf16> loc(#loc) + %121 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %122 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/biases"} -> tensor<40xbf16> loc(#loc) + %123 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %124 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/biases"} -> tensor<120xbf16> loc(#loc) + %125 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %126 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/biases"} -> tensor<32xbf16> loc(#loc) + %127 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %128 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/biases"} -> tensor<120xbf16> loc(#loc) + %129 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %130 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/biases"} -> tensor<120xbf16> loc(#loc) + %131 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %132 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/biases"} -> tensor<40xbf16> loc(#loc) + %133 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/weights"} -> tensor<40x72x1x1xbf16> loc(#loc) + %134 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/biases"} -> tensor<72xbf16> loc(#loc) + %135 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %136 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/biases"} -> tensor<24xbf16> loc(#loc) + %137 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %138 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/biases"} -> tensor<72xbf16> loc(#loc) + %139 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/weights"} -> tensor<72x1x5x5xbf16> loc(#loc) + %140 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/biases"} -> tensor<72xbf16> loc(#loc) + %141 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %142 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/biases"} -> tensor<24xbf16> loc(#loc) + %143 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %144 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/biases"} -> tensor<72xbf16> loc(#loc) + %145 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/weights"} -> tensor<72x1x3x3xbf16> loc(#loc) + %146 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/biases"} -> tensor<72xbf16> loc(#loc) + %147 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %148 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/biases"} -> tensor<24xbf16> loc(#loc) + %149 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/weights"} -> tensor<24x64x1x1xbf16> loc(#loc) + %150 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/biases"} -> tensor<64xbf16> loc(#loc) + %151 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/weights"} -> tensor<64x1x3x3xbf16> loc(#loc) + %152 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/biases"} -> tensor<64xbf16> loc(#loc) + %153 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/weights"} -> tensor<64x16x1x1xbf16> loc(#loc) + %154 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/biases"} -> tensor<16xbf16> loc(#loc) + %155 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/weights"} -> tensor<16x16x1x1xbf16> loc(#loc) + %156 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/biases"} -> tensor<16xbf16> loc(#loc) + %157 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/weights"} -> tensor<16x1x3x3xbf16> loc(#loc) + %158 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/biases"} -> tensor<16xbf16> loc(#loc) + %159 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/weights"} -> tensor<16x3x3x3xbf16> loc(#loc) + %160 = xten_nn.load_external_const {file = "constants.h5", key = "Div_16/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc5) + %161 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_14/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc308) + %162 = xten_nn.subgraph (%arg5 = %arg0: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.906250e-03 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.906250e-03> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc8) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_2", + OutputName = "Div_2", + shift = 0 : i8} : (tensor<1x180x320x4xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc8) + xten_nn.output %463 : tensor<1x180x320x4xbf16> loc(#loc8) + } -> tensor<1x180x320x4xbf16> loc(#loc8) + xten_nn.output %461 : tensor<1x180x320x4xbf16> loc(#loc8) + } -> tensor<1x180x320x4xbf16> loc(#loc8) + %163 = xten_nn.subgraph (%arg5 = %162: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Slice_7", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Slice_7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "W", + config.dim_c = 184 : ui32, + config.dim_h = 320 : ui32, + config.dim_w = 4 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_7", + OutputName = "Slice_7", + size = array, + start = array} : (tensor<1x180x320x4xbf16>) -> tensor<1x180x320x3xbf16> loc(#loc9) + xten_nn.output %461 : tensor<1x180x320x3xbf16> loc(#loc9) + } -> tensor<1x180x320x3xbf16> loc(#loc9) + %164 = xten_nn.subgraph (%arg5 = %163: tensor<1x180x320x3xbf16>) attributes { + LayerName = "Generated-#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Generated-#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "BufferPadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_0_padded = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_1_padded = 23 : ui32, + config.dim_2 = 3 : ui32, + config.dim_2_padded = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_padded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x180x320x3xbf16> loc(#loc10) + } -> tensor<1x180x320x3xbf16> loc(#loc10) + %165 = xten_nn.subgraph (%arg5 = %164: tensor<1x180x320x3xbf16>) attributes { + LayerName = "Generated-#2", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Generated-#3", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_2 = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 10 : ui32 + }} { + %461 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc11) + %462 = tosa.transpose %arg5, %461 : (tensor<1x180x320x3xbf16>, tensor<4xi32>) -> tensor<1x3x180x320xbf16> loc(#loc310) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc310) + } -> tensor<1x3x180x320xbf16> loc(#loc309) + %166 = xten_nn.subgraph (%arg5 = %165: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Generated-#4", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Generated-#5", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 184 : ui32, + config.dim_0_unpadded = 180 : ui32, + config.dim_1 = 1 : ui32, + config.dim_1_unpadded = 1 : ui32, + config.dim_2 = 320 : ui32, + config.dim_2_unpadded = 320 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x3x180x320xbf16> loc(#loc10) + } -> tensor<1x3x180x320xbf16> loc(#loc10) + %167 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>, %arg6 = %161: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Initializer_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Initializer_398", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Sub_14", OutputName = "Initializer_398"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc308) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc308) + } -> tensor<1x3x180x320xbf16> loc(#loc308) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc308) + } -> tensor<1x3x180x320xbf16> loc(#loc308) + %168 = xten_nn.subgraph (%arg5 = %167: tensor<1x3x180x320xbf16>, %arg6 = %160: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + OutputName = "Div_16", + PartOfLayerName = "Div_16", + shift = 0 : i8} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc5) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc5) + } -> tensor<1x3x180x320xbf16> loc(#loc5) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc5) + } -> tensor<1x3x180x320xbf16> loc(#loc5) + %169 = xten_nn.subgraph (%arg5 = %168: tensor<1x3x180x320xbf16>, %arg6 = %159: tensor<16x3x3x3xbf16>, %arg7 = %158: tensor<16xbf16>) attributes { + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x3x180x320xbf16>, %arg9 = %arg6: tensor<16x3x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 2 : ui8, + config.stride_w = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x3x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x3xbf16> loc(#loc13) + %465 = tosa.transpose %arg8, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc13) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_17", + PartOfOutputName = "Conv_17", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>, tensor<16x3x3x3xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc13) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc13) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc13) + } -> tensor<1x16x90x160xbf16> loc(#loc13) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc13) + } -> tensor<1x16x90x160xbf16> loc(#loc13) + %170 = xten_nn.subgraph (%arg5 = %169: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_19", OutputName = "Add_19"} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc14) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc14) + } -> tensor<1x16x90x160xbf16> loc(#loc14) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc14) + } -> tensor<1x16x90x160xbf16> loc(#loc14) + %171 = xten_nn.subgraph (%arg5 = %170: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_22", + OutputName = "Clip_22", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + %172 = xten_nn.subgraph (%arg5 = %171: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_24", + OutputName = "Div_24", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + %173 = xten_nn.subgraph (%arg5 = %169: tensor<1x16x90x160xbf16>, %arg6 = %172: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_25", + OutputName = "Mul_25", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + %174 = xten_nn.subgraph (%arg5 = %173: tensor<1x16x90x160xbf16>, %arg6 = %157: tensor<16x1x3x3xbf16>, %arg7 = %156: tensor<16xbf16>) attributes { + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<16x1x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc311) + %465 = tosa.transpose %arg9, %464 : (tensor<16x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x16x1xbf16> loc(#loc311) + %466 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc311) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_26", + PartOfOutputName = "Conv_26", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<3x3x16x1xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc18) + %468 = tosa.clamp %467 { + LayerName = "Relu_27", + OutputName = "Relu_27", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc19) + %469 = tosa.transpose %468, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc311) + xten_nn.output %469 : tensor<1x16x90x160xbf16> loc(#loc19) + } -> tensor<1x16x90x160xbf16> loc(#loc311) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc311) + } -> tensor<1x16x90x160xbf16> loc(#loc311) + %175 = xten_nn.subgraph (%arg5 = %174: tensor<1x16x90x160xbf16>, %arg6 = %155: tensor<16x16x1x1xbf16>, %arg7 = %154: tensor<16xbf16>, %arg8 = %173: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Conv_28", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x16x90x160xbf16>, %arg10 = %arg6: tensor<16x16x1x1xbf16>, %arg11 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_28", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_28", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc20) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<16x16x1x1xbf16>) -> tensor<16x1x1x16xbf16> loc(#loc20) + %466 = tosa.transpose %arg9, %464 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_28", + PartOfOutputName = "Conv_28", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<16x1x1x16xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %468 = tosa.transpose %467, %463 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc20) + xten_nn.output %468 : tensor<1x16x90x160xbf16> loc(#loc20) + } -> tensor<1x16x90x160xbf16> loc(#loc20) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x16x90x160xbf16>, %arg10 = %arg8: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_29", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_29", OutputName = "Add_29"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc312) + %176 = xten_nn.subgraph (%arg5 = %175: tensor<1x16x90x160xbf16>, %arg6 = %153: tensor<64x16x1x1xbf16>, %arg7 = %152: tensor<64xbf16>) attributes { + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<64x16x1x1xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc313) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<64x16x1x1xbf16>) -> tensor<64x1x1x16xbf16> loc(#loc313) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc313) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_30", + PartOfOutputName = "Conv_30", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<64x1x1x16xbf16>, tensor<64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc22) + %467 = tosa.clamp %466 { + LayerName = "Relu_31", + OutputName = "Relu_31", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc23) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x64xbf16>, tensor<4xi32>) -> tensor<1x64x90x160xbf16> loc(#loc313) + xten_nn.output %468 : tensor<1x64x90x160xbf16> loc(#loc23) + } -> tensor<1x64x90x160xbf16> loc(#loc313) + xten_nn.output %461 : tensor<1x64x90x160xbf16> loc(#loc313) + } -> tensor<1x64x90x160xbf16> loc(#loc313) + %177 = xten_nn.subgraph (%arg5 = %176: tensor<1x64x90x160xbf16>, %arg6 = %151: tensor<64x1x3x3xbf16>, %arg7 = %150: tensor<64xbf16>) attributes { + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x90x160xbf16>, %arg9 = %arg6: tensor<64x1x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc314) + %465 = tosa.transpose %arg9, %464 : (tensor<64x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x64x1xbf16> loc(#loc314) + %466 = tosa.transpose %arg8, %463 : (tensor<1x64x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x64xbf16> loc(#loc314) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_32", + PartOfOutputName = "Conv_32", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x64xbf16>, tensor<3x3x64x1xbf16>, tensor<64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc24) + %468 = tosa.clamp %467 { + LayerName = "Relu_33", + OutputName = "Relu_33", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc25) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x64xbf16>, tensor<4xi32>) -> tensor<1x64x45x80xbf16> loc(#loc314) + xten_nn.output %469 : tensor<1x64x45x80xbf16> loc(#loc25) + } -> tensor<1x64x45x80xbf16> loc(#loc314) + xten_nn.output %461 : tensor<1x64x45x80xbf16> loc(#loc314) + } -> tensor<1x64x45x80xbf16> loc(#loc314) + %178 = xten_nn.subgraph (%arg5 = %177: tensor<1x64x45x80xbf16>, %arg6 = %149: tensor<24x64x1x1xbf16>, %arg7 = %148: tensor<24xbf16>) attributes { + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x45x80xbf16>, %arg9 = %arg6: tensor<24x64x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc26) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x64x1x1xbf16>) -> tensor<24x1x1x64xbf16> loc(#loc26) + %465 = tosa.transpose %arg8, %463 : (tensor<1x64x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x64xbf16> loc(#loc26) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_34", + PartOfOutputName = "Conv_34", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x64xbf16>, tensor<24x1x1x64xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc26) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc26) + xten_nn.output %467 : tensor<1x24x45x80xbf16> loc(#loc26) + } -> tensor<1x24x45x80xbf16> loc(#loc26) + xten_nn.output %461 : tensor<1x24x45x80xbf16> loc(#loc26) + } -> tensor<1x24x45x80xbf16> loc(#loc26) + %179 = xten_nn.subgraph (%arg5 = %178: tensor<1x24x45x80xbf16>, %arg6 = %147: tensor<72x24x1x1xbf16>, %arg7 = %146: tensor<72xbf16>) attributes { + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc315) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc315) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc315) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_35", + PartOfOutputName = "Conv_35", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc27) + %467 = tosa.clamp %466 { + LayerName = "Relu_36", + OutputName = "Relu_36", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc28) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc315) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc28) + } -> tensor<1x72x45x80xbf16> loc(#loc315) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc315) + } -> tensor<1x72x45x80xbf16> loc(#loc315) + %180 = xten_nn.subgraph (%arg5 = %179: tensor<1x72x45x80xbf16>, %arg6 = %145: tensor<72x1x3x3xbf16>, %arg7 = %144: tensor<72xbf16>) attributes { + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x3x3xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc316) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x72x1xbf16> loc(#loc316) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc316) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_37", + PartOfOutputName = "Conv_37", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<3x3x72x1xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc29) + %468 = tosa.clamp %467 { + LayerName = "Relu_38", + OutputName = "Relu_38", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc30) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc316) + xten_nn.output %469 : tensor<1x72x45x80xbf16> loc(#loc30) + } -> tensor<1x72x45x80xbf16> loc(#loc316) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc316) + } -> tensor<1x72x45x80xbf16> loc(#loc316) + %181 = xten_nn.subgraph (%arg5 = %180: tensor<1x72x45x80xbf16>, %arg6 = %143: tensor<24x72x1x1xbf16>, %arg7 = %142: tensor<24xbf16>, %arg8 = %178: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Conv_39", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x72x45x80xbf16>, %arg10 = %arg6: tensor<24x72x1x1xbf16>, %arg11 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_39", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_39", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc31) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc31) + %466 = tosa.transpose %arg9, %464 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc31) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_39", + PartOfOutputName = "Conv_39", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc31) + %468 = tosa.transpose %467, %463 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc31) + xten_nn.output %468 : tensor<1x24x45x80xbf16> loc(#loc31) + } -> tensor<1x24x45x80xbf16> loc(#loc31) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x24x45x80xbf16>, %arg10 = %arg8: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Add_40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_40", OutputName = "Add_40"} : (tensor<1x24x45x80xbf16>, tensor<1x24x45x80xbf16>) -> tensor<1x24x45x80xbf16> loc(#loc32) + xten_nn.output %463 : tensor<1x24x45x80xbf16> loc(#loc32) + } -> tensor<1x24x45x80xbf16> loc(#loc32) + xten_nn.output %462 : tensor<1x24x45x80xbf16> loc(#loc32) + } -> tensor<1x24x45x80xbf16> loc(#loc317) + %182 = xten_nn.subgraph (%arg5 = %181: tensor<1x24x45x80xbf16>, %arg6 = %141: tensor<72x24x1x1xbf16>, %arg7 = %140: tensor<72xbf16>) attributes { + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc318) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc318) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc318) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_41", + PartOfOutputName = "Conv_41", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc33) + %467 = tosa.clamp %466 { + LayerName = "Relu_42", + OutputName = "Relu_42", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc34) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc318) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc34) + } -> tensor<1x72x45x80xbf16> loc(#loc318) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc318) + } -> tensor<1x72x45x80xbf16> loc(#loc318) + %183 = xten_nn.subgraph (%arg5 = %182: tensor<1x72x45x80xbf16>, %arg6 = %139: tensor<72x1x5x5xbf16>, %arg7 = %138: tensor<72xbf16>) attributes { + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x5x5xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 1]], + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc319) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x72x1xbf16> loc(#loc319) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc319) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_43", + PartOfOutputName = "Conv_43", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<5x5x72x1xbf16>, tensor<72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc35) + %468 = tosa.clamp %467 { + LayerName = "Relu_44", + OutputName = "Relu_44", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc36) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x72xbf16>, tensor<4xi32>) -> tensor<1x72x23x40xbf16> loc(#loc319) + xten_nn.output %469 : tensor<1x72x23x40xbf16> loc(#loc36) + } -> tensor<1x72x23x40xbf16> loc(#loc319) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc319) + } -> tensor<1x72x23x40xbf16> loc(#loc319) + %184 = xten_nn.subgraph (%arg5 = %183: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Generated-#6", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 9 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x72x23x40xbf16>) -> tensor<1x72x1x920xbf16> loc(#loc37) + xten_nn.output %461 : tensor<1x72x1x920xbf16> loc(#loc37) + } -> tensor<1x72x1x920xbf16> loc(#loc37) + %185 = xten_nn.subgraph (%arg5 = %184: tensor<1x72x1x920xbf16>) attributes { + LayerName = "Generated-#8", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#9", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x920xbf16>) attributes { + LayerName = "Generated-#8", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#9", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 72 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x72x1x920xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc37) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc37) + } -> tensor<1x72x1x1xbf16> loc(#loc37) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc37) + } -> tensor<1x72x1x1xbf16> loc(#loc37) + %186 = xten_nn.subgraph (%arg5 = %185: tensor<1x72x1x1xbf16>, %arg6 = %137: tensor<24x72x1x1xbf16>, %arg7 = %136: tensor<24xbf16>) attributes { + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x1x1xbf16>, %arg9 = %arg6: tensor<24x72x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc320) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc320) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_46", + PartOfOutputName = "Conv_46", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc38) + %465 = tosa.clamp %464 { + LayerName = "Relu_47", + OutputName = "Relu_47", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc39) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x24xbf16>) -> tensor<1x24x1x1xbf16> loc(#loc320) + xten_nn.output %466 : tensor<1x24x1x1xbf16> loc(#loc39) + } -> tensor<1x24x1x1xbf16> loc(#loc320) + xten_nn.output %461 : tensor<1x24x1x1xbf16> loc(#loc320) + } -> tensor<1x24x1x1xbf16> loc(#loc320) + %187 = xten_nn.subgraph (%arg5 = %186: tensor<1x24x1x1xbf16>, %arg6 = %135: tensor<72x24x1x1xbf16>, %arg7 = %134: tensor<72xbf16>) attributes { + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x1x1xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc40) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x24x1x1xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc40) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_48", + PartOfOutputName = "Conv_48", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc40) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x72xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc40) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc40) + } -> tensor<1x72x1x1xbf16> loc(#loc40) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc40) + } -> tensor<1x72x1x1xbf16> loc(#loc40) + %188 = xten_nn.subgraph (%arg5 = %187: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_50", OutputName = "Add_50"} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc41) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc41) + } -> tensor<1x72x1x1xbf16> loc(#loc41) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc41) + } -> tensor<1x72x1x1xbf16> loc(#loc41) + %189 = xten_nn.subgraph (%arg5 = %188: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_53", + OutputName = "Clip_53", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + %190 = xten_nn.subgraph (%arg5 = %189: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_55", + OutputName = "Div_55", + shift = 0 : i8} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + %191 = xten_nn.subgraph (%arg5 = %190: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Generated-#10", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#11", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 72 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + %192 = xten_nn.subgraph (%arg5 = %191: tensor<1x72x23x40xbf16>, %arg6 = %183: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x72x23x40xbf16>, %arg8 = %arg6: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_56", + OutputName = "Mul_56", + shift = 0 : i8} : (tensor<1x72x23x40xbf16>, tensor<1x72x23x40xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %462 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + %193 = xten_nn.subgraph (%arg5 = %192: tensor<1x72x23x40xbf16>, %arg6 = %133: tensor<40x72x1x1xbf16>, %arg7 = %132: tensor<40xbf16>) attributes { + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x23x40xbf16>, %arg9 = %arg6: tensor<40x72x1x1xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc45) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<40x72x1x1xbf16>) -> tensor<40x1x1x72xbf16> loc(#loc45) + %465 = tosa.transpose %arg8, %463 : (tensor<1x72x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x72xbf16> loc(#loc45) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_57", + PartOfOutputName = "Conv_57", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x72xbf16>, tensor<40x1x1x72xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc45) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc45) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc45) + } -> tensor<1x40x23x40xbf16> loc(#loc45) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc45) + } -> tensor<1x40x23x40xbf16> loc(#loc45) + %194 = xten_nn.subgraph (%arg5 = %193: tensor<1x40x23x40xbf16>, %arg6 = %131: tensor<120x40x1x1xbf16>, %arg7 = %130: tensor<120xbf16>) attributes { + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc321) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc321) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc321) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_58", + PartOfOutputName = "Conv_58", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc46) + %467 = tosa.clamp %466 { + LayerName = "Relu_59", + OutputName = "Relu_59", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc47) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc321) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc47) + } -> tensor<1x120x23x40xbf16> loc(#loc321) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc321) + } -> tensor<1x120x23x40xbf16> loc(#loc321) + %195 = xten_nn.subgraph (%arg5 = %194: tensor<1x120x23x40xbf16>, %arg6 = %129: tensor<120x1x5x5xbf16>, %arg7 = %128: tensor<120xbf16>) attributes { + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc322) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc322) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc322) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_60", + PartOfOutputName = "Conv_60", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc48) + %468 = tosa.clamp %467 { + LayerName = "Relu_61", + OutputName = "Relu_61", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc49) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc322) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc49) + } -> tensor<1x120x23x40xbf16> loc(#loc322) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc322) + } -> tensor<1x120x23x40xbf16> loc(#loc322) + %196 = xten_nn.subgraph (%arg5 = %195: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Generated-#12", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#13", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc50) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc50) + } -> tensor<1x120x1x920xbf16> loc(#loc50) + %197 = xten_nn.subgraph (%arg5 = %196: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#15", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#15", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc50) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc50) + } -> tensor<1x120x1x1xbf16> loc(#loc50) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc50) + } -> tensor<1x120x1x1xbf16> loc(#loc50) + %198 = xten_nn.subgraph (%arg5 = %197: tensor<1x120x1x1xbf16>, %arg6 = %127: tensor<32x120x1x1xbf16>, %arg7 = %126: tensor<32xbf16>) attributes { + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc323) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc323) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_63", + PartOfOutputName = "Conv_63", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc51) + %465 = tosa.clamp %464 { + LayerName = "Relu_64", + OutputName = "Relu_64", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc52) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc323) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc52) + } -> tensor<1x32x1x1xbf16> loc(#loc323) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc323) + } -> tensor<1x32x1x1xbf16> loc(#loc323) + %199 = xten_nn.subgraph (%arg5 = %198: tensor<1x32x1x1xbf16>, %arg6 = %125: tensor<120x32x1x1xbf16>, %arg7 = %124: tensor<120xbf16>) attributes { + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc53) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc53) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_65", + PartOfOutputName = "Conv_65", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc53) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc53) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc53) + } -> tensor<1x120x1x1xbf16> loc(#loc53) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc53) + } -> tensor<1x120x1x1xbf16> loc(#loc53) + %200 = xten_nn.subgraph (%arg5 = %199: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_67", OutputName = "Add_67"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc54) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc54) + } -> tensor<1x120x1x1xbf16> loc(#loc54) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc54) + } -> tensor<1x120x1x1xbf16> loc(#loc54) + %201 = xten_nn.subgraph (%arg5 = %200: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_70", + OutputName = "Clip_70", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + %202 = xten_nn.subgraph (%arg5 = %201: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_72", + OutputName = "Div_72", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + %203 = xten_nn.subgraph (%arg5 = %202: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Generated-#16", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#17", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + %204 = xten_nn.subgraph (%arg5 = %203: tensor<1x120x23x40xbf16>, %arg6 = %195: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_73", + OutputName = "Mul_73", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + %205 = xten_nn.subgraph (%arg5 = %204: tensor<1x120x23x40xbf16>, %arg6 = %123: tensor<40x120x1x1xbf16>, %arg7 = %122: tensor<40xbf16>, %arg8 = %193: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_74", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_74", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_74", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc58) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc58) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc58) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_74", + PartOfOutputName = "Conv_74", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc58) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc58) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc58) + } -> tensor<1x40x23x40xbf16> loc(#loc58) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_75", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_75", OutputName = "Add_75"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc59) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc59) + } -> tensor<1x40x23x40xbf16> loc(#loc59) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc59) + } -> tensor<1x40x23x40xbf16> loc(#loc324) + %206 = xten_nn.subgraph (%arg5 = %205: tensor<1x40x23x40xbf16>, %arg6 = %121: tensor<120x40x1x1xbf16>, %arg7 = %120: tensor<120xbf16>) attributes { + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc325) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc325) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc325) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_76", + PartOfOutputName = "Conv_76", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc60) + %467 = tosa.clamp %466 { + LayerName = "Relu_77", + OutputName = "Relu_77", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc61) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc325) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc61) + } -> tensor<1x120x23x40xbf16> loc(#loc325) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc325) + } -> tensor<1x120x23x40xbf16> loc(#loc325) + %207 = xten_nn.subgraph (%arg5 = %206: tensor<1x120x23x40xbf16>, %arg6 = %119: tensor<120x1x5x5xbf16>, %arg7 = %118: tensor<120xbf16>) attributes { + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc326) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc326) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc326) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_78", + PartOfOutputName = "Conv_78", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc62) + %468 = tosa.clamp %467 { + LayerName = "Relu_79", + OutputName = "Relu_79", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc63) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc326) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc63) + } -> tensor<1x120x23x40xbf16> loc(#loc326) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc326) + } -> tensor<1x120x23x40xbf16> loc(#loc326) + %208 = xten_nn.subgraph (%arg5 = %207: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Generated-#18", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#19", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc64) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc64) + } -> tensor<1x120x1x920xbf16> loc(#loc64) + %209 = xten_nn.subgraph (%arg5 = %208: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#20", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#21", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#20", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#21", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc64) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc64) + } -> tensor<1x120x1x1xbf16> loc(#loc64) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc64) + } -> tensor<1x120x1x1xbf16> loc(#loc64) + %210 = xten_nn.subgraph (%arg5 = %209: tensor<1x120x1x1xbf16>, %arg6 = %117: tensor<32x120x1x1xbf16>, %arg7 = %116: tensor<32xbf16>) attributes { + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc327) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc327) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_81", + PartOfOutputName = "Conv_81", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc65) + %465 = tosa.clamp %464 { + LayerName = "Relu_82", + OutputName = "Relu_82", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc66) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc327) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc66) + } -> tensor<1x32x1x1xbf16> loc(#loc327) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc327) + } -> tensor<1x32x1x1xbf16> loc(#loc327) + %211 = xten_nn.subgraph (%arg5 = %210: tensor<1x32x1x1xbf16>, %arg6 = %115: tensor<120x32x1x1xbf16>, %arg7 = %114: tensor<120xbf16>) attributes { + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc67) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc67) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_83", + PartOfOutputName = "Conv_83", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc67) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc67) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc67) + } -> tensor<1x120x1x1xbf16> loc(#loc67) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc67) + } -> tensor<1x120x1x1xbf16> loc(#loc67) + %212 = xten_nn.subgraph (%arg5 = %211: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_85", OutputName = "Add_85"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc68) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc68) + } -> tensor<1x120x1x1xbf16> loc(#loc68) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc68) + } -> tensor<1x120x1x1xbf16> loc(#loc68) + %213 = xten_nn.subgraph (%arg5 = %212: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_88", + OutputName = "Clip_88", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + %214 = xten_nn.subgraph (%arg5 = %213: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_90", + OutputName = "Div_90", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + %215 = xten_nn.subgraph (%arg5 = %214: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Generated-#22", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#23", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + %216 = xten_nn.subgraph (%arg5 = %215: tensor<1x120x23x40xbf16>, %arg6 = %207: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_91", + OutputName = "Mul_91", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + %217 = xten_nn.subgraph (%arg5 = %216: tensor<1x120x23x40xbf16>, %arg6 = %113: tensor<40x120x1x1xbf16>, %arg7 = %112: tensor<40xbf16>, %arg8 = %205: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_92", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_92", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_92", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc72) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc72) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc72) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_92", + PartOfOutputName = "Conv_92", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc72) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc72) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc72) + } -> tensor<1x40x23x40xbf16> loc(#loc72) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_93", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_93", OutputName = "Add_93"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc73) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc73) + } -> tensor<1x40x23x40xbf16> loc(#loc73) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc73) + } -> tensor<1x40x23x40xbf16> loc(#loc328) + %218 = xten_nn.subgraph (%arg5 = %217: tensor<1x40x23x40xbf16>, %arg6 = %111: tensor<240x40x1x1xbf16>, %arg7 = %110: tensor<240xbf16>) attributes { + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<240x40x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc74) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x40x1x1xbf16>) -> tensor<240x1x1x40xbf16> loc(#loc74) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc74) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_94", + PartOfOutputName = "Conv_94", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<240x1x1x40xbf16>, tensor<240xbf16>) -> tensor<1x23x40x240xbf16> loc(#loc74) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x240xbf16>, tensor<4xi32>) -> tensor<1x240x23x40xbf16> loc(#loc74) + xten_nn.output %467 : tensor<1x240x23x40xbf16> loc(#loc74) + } -> tensor<1x240x23x40xbf16> loc(#loc74) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc74) + } -> tensor<1x240x23x40xbf16> loc(#loc74) + %219 = xten_nn.subgraph (%arg5 = %218: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_96", OutputName = "Add_96"} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc75) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc75) + } -> tensor<1x240x23x40xbf16> loc(#loc75) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc75) + } -> tensor<1x240x23x40xbf16> loc(#loc75) + %220 = xten_nn.subgraph (%arg5 = %219: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_99", + OutputName = "Clip_99", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + %221 = xten_nn.subgraph (%arg5 = %220: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_101", + OutputName = "Div_101", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + %222 = xten_nn.subgraph (%arg5 = %218: tensor<1x240x23x40xbf16>, %arg6 = %221: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x23x40xbf16>, %arg8 = %arg6: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_102", + OutputName = "Mul_102", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + %223 = xten_nn.subgraph (%arg5 = %222: tensor<1x240x23x40xbf16>, %arg6 = %109: tensor<240x1x3x3xbf16>, %arg7 = %108: tensor<240xbf16>) attributes { + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x23x40xbf16>, %arg9 = %arg6: tensor<240x1x3x3xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 0]], + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc79) + %465 = tosa.transpose %arg9, %464 : (tensor<240x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x240x1xbf16> loc(#loc79) + %466 = tosa.transpose %arg8, %463 : (tensor<1x240x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x240xbf16> loc(#loc79) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_103", + PartOfOutputName = "Conv_103", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x240xbf16>, tensor<3x3x240x1xbf16>, tensor<240xbf16>) -> tensor<1x12x20x240xbf16> loc(#loc79) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x240xbf16>, tensor<4xi32>) -> tensor<1x240x12x20xbf16> loc(#loc79) + xten_nn.output %468 : tensor<1x240x12x20xbf16> loc(#loc79) + } -> tensor<1x240x12x20xbf16> loc(#loc79) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc79) + } -> tensor<1x240x12x20xbf16> loc(#loc79) + %224 = xten_nn.subgraph (%arg5 = %223: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_105", OutputName = "Add_105"} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc80) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc80) + } -> tensor<1x240x12x20xbf16> loc(#loc80) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc80) + } -> tensor<1x240x12x20xbf16> loc(#loc80) + %225 = xten_nn.subgraph (%arg5 = %224: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_108", + OutputName = "Clip_108", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + %226 = xten_nn.subgraph (%arg5 = %225: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_110", + OutputName = "Div_110", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + %227 = xten_nn.subgraph (%arg5 = %223: tensor<1x240x12x20xbf16>, %arg6 = %226: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x12x20xbf16>, %arg8 = %arg6: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_111", + OutputName = "Mul_111", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + %228 = xten_nn.subgraph (%arg5 = %227: tensor<1x240x12x20xbf16>, %arg6 = %107: tensor<80x240x1x1xbf16>, %arg7 = %106: tensor<80xbf16>) attributes { + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x12x20xbf16>, %arg9 = %arg6: tensor<80x240x1x1xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc84) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<80x240x1x1xbf16>) -> tensor<80x1x1x240xbf16> loc(#loc84) + %465 = tosa.transpose %arg8, %463 : (tensor<1x240x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x240xbf16> loc(#loc84) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_112", + PartOfOutputName = "Conv_112", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x240xbf16>, tensor<80x1x1x240xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc84) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc84) + xten_nn.output %467 : tensor<1x80x12x20xbf16> loc(#loc84) + } -> tensor<1x80x12x20xbf16> loc(#loc84) + xten_nn.output %461 : tensor<1x80x12x20xbf16> loc(#loc84) + } -> tensor<1x80x12x20xbf16> loc(#loc84) + %229 = xten_nn.subgraph (%arg5 = %228: tensor<1x80x12x20xbf16>, %arg6 = %105: tensor<200x80x1x1xbf16>, %arg7 = %104: tensor<200xbf16>) attributes { + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<200x80x1x1xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc85) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<200x80x1x1xbf16>) -> tensor<200x1x1x80xbf16> loc(#loc85) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc85) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_113", + PartOfOutputName = "Conv_113", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<200x1x1x80xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc85) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc85) + xten_nn.output %467 : tensor<1x200x12x20xbf16> loc(#loc85) + } -> tensor<1x200x12x20xbf16> loc(#loc85) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc85) + } -> tensor<1x200x12x20xbf16> loc(#loc85) + %230 = xten_nn.subgraph (%arg5 = %229: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_115", OutputName = "Add_115"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc86) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc86) + } -> tensor<1x200x12x20xbf16> loc(#loc86) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc86) + } -> tensor<1x200x12x20xbf16> loc(#loc86) + %231 = xten_nn.subgraph (%arg5 = %230: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_118", + OutputName = "Clip_118", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + %232 = xten_nn.subgraph (%arg5 = %231: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_120", + OutputName = "Div_120", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + %233 = xten_nn.subgraph (%arg5 = %229: tensor<1x200x12x20xbf16>, %arg6 = %232: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_121", + OutputName = "Mul_121", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + %234 = xten_nn.subgraph (%arg5 = %233: tensor<1x200x12x20xbf16>, %arg6 = %103: tensor<200x1x3x3xbf16>, %arg7 = %102: tensor<200xbf16>) attributes { + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x200x12x20xbf16>, %arg9 = %arg6: tensor<200x1x3x3xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc90) + %465 = tosa.transpose %arg9, %464 : (tensor<200x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x200x1xbf16> loc(#loc90) + %466 = tosa.transpose %arg8, %463 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc90) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_122", + PartOfOutputName = "Conv_122", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<3x3x200x1xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc90) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %468 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + %235 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_124", OutputName = "Add_124"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + %236 = xten_nn.subgraph (%arg5 = %235: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_127", + OutputName = "Clip_127", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + %237 = xten_nn.subgraph (%arg5 = %236: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_129", + OutputName = "Div_129", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + %238 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>, %arg6 = %237: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_130", + OutputName = "Mul_130", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + %239 = xten_nn.subgraph (%arg5 = %238: tensor<1x200x12x20xbf16>, %arg6 = %101: tensor<80x200x1x1xbf16>, %arg7 = %100: tensor<80xbf16>, %arg8 = %228: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_131", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x200x12x20xbf16>, %arg10 = %arg6: tensor<80x200x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_131", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_131", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc95) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x200x1x1xbf16>) -> tensor<80x1x1x200xbf16> loc(#loc95) + %466 = tosa.transpose %arg9, %464 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc95) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_131", + PartOfOutputName = "Conv_131", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<80x1x1x200xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc95) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc95) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc95) + } -> tensor<1x80x12x20xbf16> loc(#loc95) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_132", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_132", OutputName = "Add_132"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc96) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc96) + } -> tensor<1x80x12x20xbf16> loc(#loc96) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc96) + } -> tensor<1x80x12x20xbf16> loc(#loc329) + %240 = xten_nn.subgraph (%arg5 = %239: tensor<1x80x12x20xbf16>, %arg6 = %99: tensor<184x80x1x1xbf16>, %arg7 = %98: tensor<184xbf16>) attributes { + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc97) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc97) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc97) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_133", + PartOfOutputName = "Conv_133", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc97) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc97) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc97) + } -> tensor<1x184x12x20xbf16> loc(#loc97) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc97) + } -> tensor<1x184x12x20xbf16> loc(#loc97) + %241 = xten_nn.subgraph (%arg5 = %240: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_135", OutputName = "Add_135"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc98) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc98) + } -> tensor<1x184x12x20xbf16> loc(#loc98) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc98) + } -> tensor<1x184x12x20xbf16> loc(#loc98) + %242 = xten_nn.subgraph (%arg5 = %241: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_138", + OutputName = "Clip_138", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + %243 = xten_nn.subgraph (%arg5 = %242: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_140", + OutputName = "Div_140", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + %244 = xten_nn.subgraph (%arg5 = %240: tensor<1x184x12x20xbf16>, %arg6 = %243: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_141", + OutputName = "Mul_141", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + %245 = xten_nn.subgraph (%arg5 = %244: tensor<1x184x12x20xbf16>, %arg6 = %97: tensor<184x1x3x3xbf16>, %arg7 = %96: tensor<184xbf16>) attributes { + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc102) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc102) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc102) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_142", + PartOfOutputName = "Conv_142", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc102) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + %246 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_144", OutputName = "Add_144"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + %247 = xten_nn.subgraph (%arg5 = %246: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_147", + OutputName = "Clip_147", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + %248 = xten_nn.subgraph (%arg5 = %247: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_149", + OutputName = "Div_149", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + %249 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>, %arg6 = %248: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_150", + OutputName = "Mul_150", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + %250 = xten_nn.subgraph (%arg5 = %249: tensor<1x184x12x20xbf16>, %arg6 = %95: tensor<80x184x1x1xbf16>, %arg7 = %94: tensor<80xbf16>, %arg8 = %239: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_151", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_151", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_151", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc107) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc107) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc107) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_151", + PartOfOutputName = "Conv_151", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc107) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc107) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc107) + } -> tensor<1x80x12x20xbf16> loc(#loc107) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_152", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_152", OutputName = "Add_152"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc108) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc108) + } -> tensor<1x80x12x20xbf16> loc(#loc108) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc108) + } -> tensor<1x80x12x20xbf16> loc(#loc330) + %251 = xten_nn.subgraph (%arg5 = %250: tensor<1x80x12x20xbf16>, %arg6 = %93: tensor<184x80x1x1xbf16>, %arg7 = %92: tensor<184xbf16>) attributes { + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc109) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc109) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc109) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_153", + PartOfOutputName = "Conv_153", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc109) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + %252 = xten_nn.subgraph (%arg5 = %251: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_155", OutputName = "Add_155"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + %253 = xten_nn.subgraph (%arg5 = %252: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_158", + OutputName = "Clip_158", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + %254 = xten_nn.subgraph (%arg5 = %253: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_160", + OutputName = "Div_160", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + %255 = xten_nn.subgraph (%arg5 = %251: tensor<1x184x12x20xbf16>, %arg6 = %254: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_161", + OutputName = "Mul_161", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + %256 = xten_nn.subgraph (%arg5 = %255: tensor<1x184x12x20xbf16>, %arg6 = %91: tensor<184x1x3x3xbf16>, %arg7 = %90: tensor<184xbf16>) attributes { + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc114) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc114) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc114) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_162", + PartOfOutputName = "Conv_162", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc114) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + %257 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_164", OutputName = "Add_164"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + %258 = xten_nn.subgraph (%arg5 = %257: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_167", + OutputName = "Clip_167", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + %259 = xten_nn.subgraph (%arg5 = %258: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_169", + OutputName = "Div_169", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + %260 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>, %arg6 = %259: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_170", + OutputName = "Mul_170", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + %261 = xten_nn.subgraph (%arg5 = %260: tensor<1x184x12x20xbf16>, %arg6 = %89: tensor<80x184x1x1xbf16>, %arg7 = %88: tensor<80xbf16>, %arg8 = %250: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_171", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_171", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_171", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc119) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc119) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc119) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_171", + PartOfOutputName = "Conv_171", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc119) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc119) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc119) + } -> tensor<1x80x12x20xbf16> loc(#loc119) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_172", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_172", OutputName = "Add_172"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc120) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc120) + } -> tensor<1x80x12x20xbf16> loc(#loc120) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc120) + } -> tensor<1x80x12x20xbf16> loc(#loc331) + %262 = xten_nn.subgraph (%arg5 = %261: tensor<1x80x12x20xbf16>, %arg6 = %87: tensor<480x80x1x1xbf16>, %arg7 = %86: tensor<480xbf16>) attributes { + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<480x80x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc121) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x80x1x1xbf16>) -> tensor<480x1x1x80xbf16> loc(#loc121) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc121) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_173", + PartOfOutputName = "Conv_173", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<480x1x1x80xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc121) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc121) + xten_nn.output %467 : tensor<1x480x12x20xbf16> loc(#loc121) + } -> tensor<1x480x12x20xbf16> loc(#loc121) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc121) + } -> tensor<1x480x12x20xbf16> loc(#loc121) + %263 = xten_nn.subgraph (%arg5 = %262: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_175", OutputName = "Add_175"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc122) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc122) + } -> tensor<1x480x12x20xbf16> loc(#loc122) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc122) + } -> tensor<1x480x12x20xbf16> loc(#loc122) + %264 = xten_nn.subgraph (%arg5 = %263: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_178", + OutputName = "Clip_178", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + %265 = xten_nn.subgraph (%arg5 = %264: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_180", + OutputName = "Div_180", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + %266 = xten_nn.subgraph (%arg5 = %262: tensor<1x480x12x20xbf16>, %arg6 = %265: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_181", + OutputName = "Mul_181", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + %267 = xten_nn.subgraph (%arg5 = %266: tensor<1x480x12x20xbf16>, %arg6 = %85: tensor<480x1x3x3xbf16>, %arg7 = %84: tensor<480xbf16>) attributes { + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<480x1x3x3xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc126) + %465 = tosa.transpose %arg9, %464 : (tensor<480x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x480x1xbf16> loc(#loc126) + %466 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc126) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_182", + PartOfOutputName = "Conv_182", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<3x3x480x1xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc126) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %468 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + %268 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_184", OutputName = "Add_184"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + %269 = xten_nn.subgraph (%arg5 = %268: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_187", + OutputName = "Clip_187", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + %270 = xten_nn.subgraph (%arg5 = %269: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_189", + OutputName = "Div_189", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + %271 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>, %arg6 = %270: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_190", + OutputName = "Mul_190", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + %272 = xten_nn.subgraph (%arg5 = %271: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Generated-#24", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#25", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 60 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x1x240xbf16> loc(#loc332) + xten_nn.output %461 : tensor<1x480x1x240xbf16> loc(#loc332) + } -> tensor<1x480x1x240xbf16> loc(#loc332) + %273 = xten_nn.subgraph (%arg5 = %272: tensor<1x480x1x240xbf16>) attributes { + LayerName = "Generated-#26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x240xbf16>) attributes { + LayerName = "Generated-#26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#27", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 480 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x480x1x240xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc131) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc131) + } -> tensor<1x480x1x1xbf16> loc(#loc131) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc131) + } -> tensor<1x480x1x1xbf16> loc(#loc131) + %274 = xten_nn.subgraph (%arg5 = %273: tensor<1x480x1x1xbf16>, %arg6 = %83: tensor<120x480x1x1xbf16>, %arg7 = %82: tensor<120xbf16>) attributes { + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x1x1xbf16>, %arg9 = %arg6: tensor<120x480x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x480x1x1xbf16>) -> tensor<120x1x1x480xbf16> loc(#loc333) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc333) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_192", + PartOfOutputName = "Conv_192", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x480xbf16>, tensor<120x1x1x480xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc132) + %465 = tosa.clamp %464 { + LayerName = "Relu_193", + OutputName = "Relu_193", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc133) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc333) + xten_nn.output %466 : tensor<1x120x1x1xbf16> loc(#loc133) + } -> tensor<1x120x1x1xbf16> loc(#loc333) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc333) + } -> tensor<1x120x1x1xbf16> loc(#loc333) + %275 = xten_nn.subgraph (%arg5 = %274: tensor<1x120x1x1xbf16>, %arg6 = %81: tensor<480x120x1x1xbf16>, %arg7 = %80: tensor<480xbf16>) attributes { + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<480x120x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x120x1x1xbf16>) -> tensor<480x1x1x120xbf16> loc(#loc134) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc134) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_194", + PartOfOutputName = "Conv_194", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<480x1x1x120xbf16>, tensor<480xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc134) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x480xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc134) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc134) + } -> tensor<1x480x1x1xbf16> loc(#loc134) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc134) + } -> tensor<1x480x1x1xbf16> loc(#loc134) + %276 = xten_nn.subgraph (%arg5 = %275: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_196", OutputName = "Add_196"} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc135) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc135) + } -> tensor<1x480x1x1xbf16> loc(#loc135) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc135) + } -> tensor<1x480x1x1xbf16> loc(#loc135) + %277 = xten_nn.subgraph (%arg5 = %276: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_199", + OutputName = "Clip_199", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + %278 = xten_nn.subgraph (%arg5 = %277: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_201", + OutputName = "Div_201", + shift = 0 : i8} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + %279 = xten_nn.subgraph (%arg5 = %278: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Generated-#28", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#29", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 480 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + %280 = xten_nn.subgraph (%arg5 = %279: tensor<1x480x12x20xbf16>, %arg6 = %271: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_202", + OutputName = "Mul_202", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + %281 = xten_nn.subgraph (%arg5 = %280: tensor<1x480x12x20xbf16>, %arg6 = %79: tensor<112x480x1x1xbf16>, %arg7 = %78: tensor<112xbf16>) attributes { + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<112x480x1x1xbf16>, %arg10 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc139) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<112x480x1x1xbf16>) -> tensor<112x1x1x480xbf16> loc(#loc139) + %465 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc139) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_203", + PartOfOutputName = "Conv_203", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<112x1x1x480xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc139) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc139) + xten_nn.output %467 : tensor<1x112x12x20xbf16> loc(#loc139) + } -> tensor<1x112x12x20xbf16> loc(#loc139) + xten_nn.output %461 : tensor<1x112x12x20xbf16> loc(#loc139) + } -> tensor<1x112x12x20xbf16> loc(#loc139) + %282 = xten_nn.subgraph (%arg5 = %281: tensor<1x112x12x20xbf16>, %arg6 = %77: tensor<672x112x1x1xbf16>, %arg7 = %76: tensor<672xbf16>) attributes { + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc140) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc140) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc140) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_204", + PartOfOutputName = "Conv_204", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc140) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc140) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc140) + } -> tensor<1x672x12x20xbf16> loc(#loc140) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc140) + } -> tensor<1x672x12x20xbf16> loc(#loc140) + %283 = xten_nn.subgraph (%arg5 = %282: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_206", OutputName = "Add_206"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc141) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc141) + } -> tensor<1x672x12x20xbf16> loc(#loc141) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc141) + } -> tensor<1x672x12x20xbf16> loc(#loc141) + %284 = xten_nn.subgraph (%arg5 = %283: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_209", + OutputName = "Clip_209", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + %285 = xten_nn.subgraph (%arg5 = %284: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_211", + OutputName = "Div_211", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + %286 = xten_nn.subgraph (%arg5 = %282: tensor<1x672x12x20xbf16>, %arg6 = %285: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_212", + OutputName = "Mul_212", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + %287 = xten_nn.subgraph (%arg5 = %286: tensor<1x672x12x20xbf16>, %arg6 = %75: tensor<672x1x3x3xbf16>, %arg7 = %74: tensor<672xbf16>) attributes { + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x3x3xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc145) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x672x1xbf16> loc(#loc145) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc145) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_213", + PartOfOutputName = "Conv_213", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<3x3x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc145) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + %288 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_215", OutputName = "Add_215"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + %289 = xten_nn.subgraph (%arg5 = %288: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_218", + OutputName = "Clip_218", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + %290 = xten_nn.subgraph (%arg5 = %289: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_220", + OutputName = "Div_220", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + %291 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>, %arg6 = %290: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_221", + OutputName = "Mul_221", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + %292 = xten_nn.subgraph (%arg5 = %291: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Generated-#30", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#31", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc334) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc334) + } -> tensor<1x672x1x240xbf16> loc(#loc334) + %293 = xten_nn.subgraph (%arg5 = %292: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#33", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc150) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc150) + } -> tensor<1x672x1x1xbf16> loc(#loc150) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc150) + } -> tensor<1x672x1x1xbf16> loc(#loc150) + %294 = xten_nn.subgraph (%arg5 = %293: tensor<1x672x1x1xbf16>, %arg6 = %73: tensor<168x672x1x1xbf16>, %arg7 = %72: tensor<168xbf16>) attributes { + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc335) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc335) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_223", + PartOfOutputName = "Conv_223", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc151) + %465 = tosa.clamp %464 { + LayerName = "Relu_224", + OutputName = "Relu_224", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc152) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc335) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc152) + } -> tensor<1x168x1x1xbf16> loc(#loc335) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc335) + } -> tensor<1x168x1x1xbf16> loc(#loc335) + %295 = xten_nn.subgraph (%arg5 = %294: tensor<1x168x1x1xbf16>, %arg6 = %71: tensor<672x168x1x1xbf16>, %arg7 = %70: tensor<672xbf16>) attributes { + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc153) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc153) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_225", + PartOfOutputName = "Conv_225", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc153) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc153) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc153) + } -> tensor<1x672x1x1xbf16> loc(#loc153) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc153) + } -> tensor<1x672x1x1xbf16> loc(#loc153) + %296 = xten_nn.subgraph (%arg5 = %295: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_227", OutputName = "Add_227"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc154) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc154) + } -> tensor<1x672x1x1xbf16> loc(#loc154) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc154) + } -> tensor<1x672x1x1xbf16> loc(#loc154) + %297 = xten_nn.subgraph (%arg5 = %296: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_230", + OutputName = "Clip_230", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + %298 = xten_nn.subgraph (%arg5 = %297: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_232", + OutputName = "Div_232", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + %299 = xten_nn.subgraph (%arg5 = %298: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Generated-#34", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#35", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + %300 = xten_nn.subgraph (%arg5 = %299: tensor<1x672x12x20xbf16>, %arg6 = %291: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_233", + OutputName = "Mul_233", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + %301 = xten_nn.subgraph (%arg5 = %300: tensor<1x672x12x20xbf16>, %arg6 = %69: tensor<112x672x1x1xbf16>, %arg7 = %68: tensor<112xbf16>, %arg8 = %281: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Conv_234", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x672x12x20xbf16>, %arg10 = %arg6: tensor<112x672x1x1xbf16>, %arg11 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_234", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_234", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc158) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<112x672x1x1xbf16>) -> tensor<112x1x1x672xbf16> loc(#loc158) + %466 = tosa.transpose %arg9, %464 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc158) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_234", + PartOfOutputName = "Conv_234", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<112x1x1x672xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc158) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc158) + xten_nn.output %468 : tensor<1x112x12x20xbf16> loc(#loc158) + } -> tensor<1x112x12x20xbf16> loc(#loc158) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x112x12x20xbf16>, %arg10 = %arg8: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Add_235", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_235", OutputName = "Add_235"} : (tensor<1x112x12x20xbf16>, tensor<1x112x12x20xbf16>) -> tensor<1x112x12x20xbf16> loc(#loc159) + xten_nn.output %463 : tensor<1x112x12x20xbf16> loc(#loc159) + } -> tensor<1x112x12x20xbf16> loc(#loc159) + xten_nn.output %462 : tensor<1x112x12x20xbf16> loc(#loc159) + } -> tensor<1x112x12x20xbf16> loc(#loc336) + %302 = xten_nn.subgraph (%arg5 = %301: tensor<1x112x12x20xbf16>, %arg6 = %67: tensor<672x112x1x1xbf16>, %arg7 = %66: tensor<672xbf16>) attributes { + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc160) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc160) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc160) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_236", + PartOfOutputName = "Conv_236", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc160) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc160) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc160) + } -> tensor<1x672x12x20xbf16> loc(#loc160) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc160) + } -> tensor<1x672x12x20xbf16> loc(#loc160) + %303 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_238", OutputName = "Add_238"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc161) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc161) + } -> tensor<1x672x12x20xbf16> loc(#loc161) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc161) + } -> tensor<1x672x12x20xbf16> loc(#loc161) + %304 = xten_nn.subgraph (%arg5 = %303: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_241", + OutputName = "Clip_241", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + %305 = xten_nn.subgraph (%arg5 = %304: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_243", + OutputName = "Div_243", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + %306 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>, %arg6 = %305: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_244", + OutputName = "Mul_244", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + %307 = xten_nn.subgraph (%arg5 = %306: tensor<1x672x12x20xbf16>, %arg6 = %65: tensor<672x1x9x9xbf16>, %arg7 = %64: tensor<672xbf16>) attributes { + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x9x9xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc165) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x672x1xbf16> loc(#loc165) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc165) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_245", + PartOfOutputName = "Conv_245", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<9x9x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc165) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + %308 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_247", OutputName = "Add_247"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + %309 = xten_nn.subgraph (%arg5 = %308: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_250", + OutputName = "Clip_250", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + %310 = xten_nn.subgraph (%arg5 = %309: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_252", + OutputName = "Div_252", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + %311 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>, %arg6 = %310: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_253", + OutputName = "Mul_253", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + %312 = xten_nn.subgraph (%arg5 = %311: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Generated-#36", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#37", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc337) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc337) + } -> tensor<1x672x1x240xbf16> loc(#loc337) + %313 = xten_nn.subgraph (%arg5 = %312: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#38", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#39", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#38", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#39", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc170) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc170) + } -> tensor<1x672x1x1xbf16> loc(#loc170) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc170) + } -> tensor<1x672x1x1xbf16> loc(#loc170) + %314 = xten_nn.subgraph (%arg5 = %313: tensor<1x672x1x1xbf16>, %arg6 = %63: tensor<168x672x1x1xbf16>, %arg7 = %62: tensor<168xbf16>) attributes { + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc338) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc338) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_255", + PartOfOutputName = "Conv_255", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc171) + %465 = tosa.clamp %464 { + LayerName = "Relu_256", + OutputName = "Relu_256", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc172) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc338) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc172) + } -> tensor<1x168x1x1xbf16> loc(#loc338) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc338) + } -> tensor<1x168x1x1xbf16> loc(#loc338) + %315 = xten_nn.subgraph (%arg5 = %314: tensor<1x168x1x1xbf16>, %arg6 = %61: tensor<672x168x1x1xbf16>, %arg7 = %60: tensor<672xbf16>) attributes { + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc173) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc173) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_257", + PartOfOutputName = "Conv_257", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc173) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc173) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc173) + } -> tensor<1x672x1x1xbf16> loc(#loc173) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc173) + } -> tensor<1x672x1x1xbf16> loc(#loc173) + %316 = xten_nn.subgraph (%arg5 = %315: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_259", OutputName = "Add_259"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc174) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc174) + } -> tensor<1x672x1x1xbf16> loc(#loc174) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc174) + } -> tensor<1x672x1x1xbf16> loc(#loc174) + %317 = xten_nn.subgraph (%arg5 = %316: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_262", + OutputName = "Clip_262", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + %318 = xten_nn.subgraph (%arg5 = %317: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_264", + OutputName = "Div_264", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + %319 = xten_nn.subgraph (%arg5 = %318: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Generated-#40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#41", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + %320 = xten_nn.subgraph (%arg5 = %319: tensor<1x672x12x20xbf16>, %arg6 = %311: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_265", + OutputName = "Mul_265", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + %321 = xten_nn.subgraph (%arg5 = %320: tensor<1x672x12x20xbf16>, %arg6 = %59: tensor<160x672x1x1xbf16>, %arg7 = %58: tensor<160xbf16>) attributes { + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<160x672x1x1xbf16>, %arg10 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc178) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<160x672x1x1xbf16>) -> tensor<160x1x1x672xbf16> loc(#loc178) + %465 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc178) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_266", + PartOfOutputName = "Conv_266", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<160x1x1x672xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc178) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc178) + xten_nn.output %467 : tensor<1x160x12x20xbf16> loc(#loc178) + } -> tensor<1x160x12x20xbf16> loc(#loc178) + xten_nn.output %461 : tensor<1x160x12x20xbf16> loc(#loc178) + } -> tensor<1x160x12x20xbf16> loc(#loc178) + %322 = xten_nn.subgraph (%arg5 = %321: tensor<1x160x12x20xbf16>, %arg6 = %57: tensor<960x160x1x1xbf16>, %arg7 = %56: tensor<960xbf16>) attributes { + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc179) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc179) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc179) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_267", + PartOfOutputName = "Conv_267", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc179) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc179) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc179) + } -> tensor<1x960x12x20xbf16> loc(#loc179) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc179) + } -> tensor<1x960x12x20xbf16> loc(#loc179) + %323 = xten_nn.subgraph (%arg5 = %322: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_269", OutputName = "Add_269"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc180) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc180) + } -> tensor<1x960x12x20xbf16> loc(#loc180) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc180) + } -> tensor<1x960x12x20xbf16> loc(#loc180) + %324 = xten_nn.subgraph (%arg5 = %323: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_272", + OutputName = "Clip_272", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + %325 = xten_nn.subgraph (%arg5 = %324: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_274", + OutputName = "Div_274", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + %326 = xten_nn.subgraph (%arg5 = %322: tensor<1x960x12x20xbf16>, %arg6 = %325: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_275", + OutputName = "Mul_275", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + %327 = xten_nn.subgraph (%arg5 = %326: tensor<1x960x12x20xbf16>, %arg6 = %55: tensor<960x1x9x9xbf16>, %arg7 = %54: tensor<960xbf16>) attributes { + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc184) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc184) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc184) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_276", + PartOfOutputName = "Conv_276", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc184) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + %328 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_278", OutputName = "Add_278"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + %329 = xten_nn.subgraph (%arg5 = %328: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_281", + OutputName = "Clip_281", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + %330 = xten_nn.subgraph (%arg5 = %329: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_283", + OutputName = "Div_283", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + %331 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>, %arg6 = %330: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_284", + OutputName = "Mul_284", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + %332 = xten_nn.subgraph (%arg5 = %331: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Generated-#42", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#43", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc339) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc339) + } -> tensor<1x960x1x240xbf16> loc(#loc339) + %333 = xten_nn.subgraph (%arg5 = %332: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#44", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#45", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#44", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#45", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc189) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc189) + } -> tensor<1x960x1x1xbf16> loc(#loc189) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc189) + } -> tensor<1x960x1x1xbf16> loc(#loc189) + %334 = xten_nn.subgraph (%arg5 = %333: tensor<1x960x1x1xbf16>, %arg6 = %53: tensor<240x960x1x1xbf16>, %arg7 = %52: tensor<240xbf16>) attributes { + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc340) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc340) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_286", + PartOfOutputName = "Conv_286", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc190) + %465 = tosa.clamp %464 { + LayerName = "Relu_287", + OutputName = "Relu_287", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc191) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc340) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc191) + } -> tensor<1x240x1x1xbf16> loc(#loc340) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc340) + } -> tensor<1x240x1x1xbf16> loc(#loc340) + %335 = xten_nn.subgraph (%arg5 = %334: tensor<1x240x1x1xbf16>, %arg6 = %51: tensor<960x240x1x1xbf16>, %arg7 = %50: tensor<960xbf16>) attributes { + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc192) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc192) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_288", + PartOfOutputName = "Conv_288", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc192) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc192) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc192) + } -> tensor<1x960x1x1xbf16> loc(#loc192) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc192) + } -> tensor<1x960x1x1xbf16> loc(#loc192) + %336 = xten_nn.subgraph (%arg5 = %335: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_290", OutputName = "Add_290"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc193) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc193) + } -> tensor<1x960x1x1xbf16> loc(#loc193) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc193) + } -> tensor<1x960x1x1xbf16> loc(#loc193) + %337 = xten_nn.subgraph (%arg5 = %336: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_293", + OutputName = "Clip_293", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + %338 = xten_nn.subgraph (%arg5 = %337: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_295", + OutputName = "Div_295", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + %339 = xten_nn.subgraph (%arg5 = %338: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Generated-#46", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#47", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + %340 = xten_nn.subgraph (%arg5 = %339: tensor<1x960x12x20xbf16>, %arg6 = %331: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_296", + OutputName = "Mul_296", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + %341 = xten_nn.subgraph (%arg5 = %340: tensor<1x960x12x20xbf16>, %arg6 = %49: tensor<160x960x1x1xbf16>, %arg7 = %48: tensor<160xbf16>, %arg8 = %321: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_297", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_297", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_297", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc197) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc197) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc197) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_297", + PartOfOutputName = "Conv_297", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc197) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc197) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc197) + } -> tensor<1x160x12x20xbf16> loc(#loc197) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_298", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_298", OutputName = "Add_298"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc198) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc198) + } -> tensor<1x160x12x20xbf16> loc(#loc198) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc198) + } -> tensor<1x160x12x20xbf16> loc(#loc341) + %342 = xten_nn.subgraph (%arg5 = %341: tensor<1x160x12x20xbf16>, %arg6 = %47: tensor<960x160x1x1xbf16>, %arg7 = %46: tensor<960xbf16>) attributes { + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc199) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc199) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc199) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_299", + PartOfOutputName = "Conv_299", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc199) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc199) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc199) + } -> tensor<1x960x12x20xbf16> loc(#loc199) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc199) + } -> tensor<1x960x12x20xbf16> loc(#loc199) + %343 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_301", OutputName = "Add_301"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc200) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc200) + } -> tensor<1x960x12x20xbf16> loc(#loc200) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc200) + } -> tensor<1x960x12x20xbf16> loc(#loc200) + %344 = xten_nn.subgraph (%arg5 = %343: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_304", + OutputName = "Clip_304", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + %345 = xten_nn.subgraph (%arg5 = %344: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_306", + OutputName = "Div_306", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + %346 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>, %arg6 = %345: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_307", + OutputName = "Mul_307", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + %347 = xten_nn.subgraph (%arg5 = %346: tensor<1x960x12x20xbf16>, %arg6 = %45: tensor<960x1x9x9xbf16>, %arg7 = %44: tensor<960xbf16>) attributes { + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc204) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc204) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc204) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_308", + PartOfOutputName = "Conv_308", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc204) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + %348 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_310", OutputName = "Add_310"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + %349 = xten_nn.subgraph (%arg5 = %348: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_313", + OutputName = "Clip_313", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + %350 = xten_nn.subgraph (%arg5 = %349: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_315", + OutputName = "Div_315", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + %351 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>, %arg6 = %350: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_316", + OutputName = "Mul_316", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + %352 = xten_nn.subgraph (%arg5 = %351: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Generated-#48", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#49", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc342) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc342) + } -> tensor<1x960x1x240xbf16> loc(#loc342) + %353 = xten_nn.subgraph (%arg5 = %352: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#51", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#51", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc209) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc209) + } -> tensor<1x960x1x1xbf16> loc(#loc209) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc209) + } -> tensor<1x960x1x1xbf16> loc(#loc209) + %354 = xten_nn.subgraph (%arg5 = %353: tensor<1x960x1x1xbf16>, %arg6 = %43: tensor<240x960x1x1xbf16>, %arg7 = %42: tensor<240xbf16>) attributes { + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc343) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc343) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_318", + PartOfOutputName = "Conv_318", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc210) + %465 = tosa.clamp %464 { + LayerName = "Relu_319", + OutputName = "Relu_319", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc211) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc343) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc211) + } -> tensor<1x240x1x1xbf16> loc(#loc343) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc343) + } -> tensor<1x240x1x1xbf16> loc(#loc343) + %355 = xten_nn.subgraph (%arg5 = %354: tensor<1x240x1x1xbf16>, %arg6 = %41: tensor<960x240x1x1xbf16>, %arg7 = %40: tensor<960xbf16>) attributes { + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc212) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc212) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_320", + PartOfOutputName = "Conv_320", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc212) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc212) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc212) + } -> tensor<1x960x1x1xbf16> loc(#loc212) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc212) + } -> tensor<1x960x1x1xbf16> loc(#loc212) + %356 = xten_nn.subgraph (%arg5 = %355: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_322", OutputName = "Add_322"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc213) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc213) + } -> tensor<1x960x1x1xbf16> loc(#loc213) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc213) + } -> tensor<1x960x1x1xbf16> loc(#loc213) + %357 = xten_nn.subgraph (%arg5 = %356: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_325", + OutputName = "Clip_325", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + %358 = xten_nn.subgraph (%arg5 = %357: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_327", + OutputName = "Div_327", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + %359 = xten_nn.subgraph (%arg5 = %358: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Generated-#52", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#53", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + %360 = xten_nn.subgraph (%arg5 = %359: tensor<1x960x12x20xbf16>, %arg6 = %351: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_328", + OutputName = "Mul_328", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + %361 = xten_nn.subgraph (%arg5 = %360: tensor<1x960x12x20xbf16>, %arg6 = %39: tensor<160x960x1x1xbf16>, %arg7 = %38: tensor<160xbf16>, %arg8 = %341: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_329", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_329", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_329", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc217) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc217) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc217) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_329", + PartOfOutputName = "Conv_329", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc217) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc217) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc217) + } -> tensor<1x160x12x20xbf16> loc(#loc217) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_330", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_330", OutputName = "Add_330"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc218) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc218) + } -> tensor<1x160x12x20xbf16> loc(#loc218) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc218) + } -> tensor<1x160x12x20xbf16> loc(#loc344) + %362 = xten_nn.subgraph (%arg5 = %361: tensor<1x160x12x20xbf16>, %arg6 = %37: tensor<960x160x1x1xbf16>, %arg7 = %36: tensor<960xbf16>) attributes { + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc219) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc219) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc219) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_331", + PartOfOutputName = "Conv_331", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc219) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc219) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc219) + } -> tensor<1x960x12x20xbf16> loc(#loc219) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc219) + } -> tensor<1x960x12x20xbf16> loc(#loc219) + %363 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_333", OutputName = "Add_333"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc220) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc220) + } -> tensor<1x960x12x20xbf16> loc(#loc220) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc220) + } -> tensor<1x960x12x20xbf16> loc(#loc220) + %364 = xten_nn.subgraph (%arg5 = %363: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_336", + OutputName = "Clip_336", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + %365 = xten_nn.subgraph (%arg5 = %364: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_338", + OutputName = "Div_338", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + %366 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>, %arg6 = %365: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_339", + OutputName = "Mul_339", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + %367 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Generated-#54", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#55", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc345) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc345) + } -> tensor<1x960x1x240xbf16> loc(#loc345) + %368 = xten_nn.subgraph (%arg5 = %367: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#57", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc224) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc224) + } -> tensor<1x960x1x1xbf16> loc(#loc224) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc224) + } -> tensor<1x960x1x1xbf16> loc(#loc224) + %369 = xten_nn.subgraph (%arg5 = %368: tensor<1x960x1x1xbf16>, %arg6 = %35: tensor<128x960x1x1xbf16>, %arg7 = %34: tensor<128xbf16>) attributes { + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<128x960x1x1xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc225) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc225) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_343", + PartOfOutputName = "Conv_343", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x1x1x128xbf16> loc(#loc225) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x128xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc225) + xten_nn.output %465 : tensor<1x128x1x1xbf16> loc(#loc225) + } -> tensor<1x128x1x1xbf16> loc(#loc225) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc225) + } -> tensor<1x128x1x1xbf16> loc(#loc225) + %370 = xten_nn.subgraph (%arg5 = %369: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_344", OutputName = "Sigmoid_344"} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc226) + xten_nn.output %462 : tensor<1x128x1x1xbf16> loc(#loc226) + } -> tensor<1x128x1x1xbf16> loc(#loc226) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc226) + } -> tensor<1x128x1x1xbf16> loc(#loc226) + %371 = xten_nn.subgraph (%arg5 = %370: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Generated-#58", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#59", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 128 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc227) + %372 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>, %arg6 = %33: tensor<128x960x1x1xbf16>, %arg7 = %32: tensor<128xbf16>, %arg8 = %371: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Conv_340", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<128x960x1x1xbf16>, %arg11 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_340", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_341", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc347) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc347) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc347) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_340", + PartOfOutputName = "Conv_340", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc228) + %468 = tosa.clamp %467 { + LayerName = "Relu_341", + OutputName = "Relu_341", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x12x20x128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc229) + %469 = tosa.transpose %468, %463 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc347) + xten_nn.output %469 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc347) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x128x12x20xbf16>, %arg10 = %arg8: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Mul_345", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.mul %arg9, %arg10 { + LayerName = "Mul_345", + OutputName = "Mul_345", + shift = 0 : i8} : (tensor<1x128x12x20xbf16>, tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %463 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc346) + %373 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc230) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc230) + } -> tensor<1x64x12x20xbf16> loc(#loc230) + %374 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc230) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc230) + } -> tensor<1x64x12x20xbf16> loc(#loc230) + %375 = xten_nn.subgraph (%arg5 = %374: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_350", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_350", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_350", + OutputName = "Concat_350", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc231) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc231) + } -> tensor<1x128x12x20xbf16> loc(#loc231) + %376 = xten_nn.subgraph (%arg5 = %375: tensor<1x128x12x20xbf16>, %arg6 = %31: tensor<128x128x3x3xbf16>, %arg7 = %30: tensor<128xbf16>) attributes { + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<128x128x3x3xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<128x128x3x3xbf16>, tensor<4xi32>) -> tensor<128x3x3x128xbf16> loc(#loc232) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc232) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_351", + PartOfOutputName = "Conv_351", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<128x3x3x128xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc232) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc232) + xten_nn.output %467 : tensor<1x128x12x20xbf16> loc(#loc232) + } -> tensor<1x128x12x20xbf16> loc(#loc232) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc232) + } -> tensor<1x128x12x20xbf16> loc(#loc232) + %377 = xten_nn.subgraph (%arg5 = %376: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_352", OutputName = "Sigmoid_352"} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + %378 = xten_nn.subgraph (%arg5 = %377: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc234) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc234) + } -> tensor<1x64x12x20xbf16> loc(#loc234) + %379 = xten_nn.subgraph (%arg5 = %377: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc234) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc234) + } -> tensor<1x64x12x20xbf16> loc(#loc234) + %380 = xten_nn.subgraph (%arg5 = %378: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_354", + OutputName = "Mul_354", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc235) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc235) + } -> tensor<1x64x12x20xbf16> loc(#loc235) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc235) + } -> tensor<1x64x12x20xbf16> loc(#loc235) + %381 = xten_nn.subgraph (%arg5 = %374: tensor<1x64x12x20xbf16>, %arg6 = %380: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_355", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_355", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_355", + OutputName = "Concat_355", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc236) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc236) + } -> tensor<1x128x12x20xbf16> loc(#loc236) + %382 = xten_nn.subgraph (%arg5 = %381: tensor<1x128x12x20xbf16>, %arg6 = %29: tensor<64x128x3x3xbf16>, %arg7 = %28: tensor<64xbf16>) attributes { + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<64x128x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<64x128x3x3xbf16>, tensor<4xi32>) -> tensor<64x3x3x128xbf16> loc(#loc237) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc237) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_356", + PartOfOutputName = "Conv_356", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<64x3x3x128xbf16>, tensor<64xbf16>) -> tensor<1x12x20x64xbf16> loc(#loc237) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x64xbf16>, tensor<4xi32>) -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %467 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + %383 = xten_nn.subgraph (%arg5 = %382: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_357", OutputName = "Tanh_357"} : (tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + %384 = xten_nn.subgraph (%arg5 = %379: tensor<1x64x12x20xbf16>, %arg6 = %383: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_361", + OutputName = "Mul_361", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc239) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc239) + } -> tensor<1x64x12x20xbf16> loc(#loc239) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc239) + } -> tensor<1x64x12x20xbf16> loc(#loc239) + %385 = xten_nn.subgraph (%arg5 = %27: tensor<1x64x12x20xbf16>, %arg6 = %379: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_359", OutputName = "Sub_359"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc4) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc4) + } -> tensor<1x64x12x20xbf16> loc(#loc4) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc4) + } -> tensor<1x64x12x20xbf16> loc(#loc4) + %386 = xten_nn.subgraph (%arg5 = %385: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_360", + OutputName = "Mul_360", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + %387 = xten_nn.subgraph (%arg5 = %386: tensor<1x64x12x20xbf16>, %arg6 = %384: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_362", OutputName = "Add_362"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + %388 = xten_nn.subgraph (%arg5 = %373: tensor<1x64x12x20xbf16>, %arg6 = %387: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_363", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_363", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_363", + OutputName = "Concat_363", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc242) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc242) + } -> tensor<1x128x12x20xbf16> loc(#loc242) + %389 = xten_nn.subgraph (%arg5 = %388: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Resize_365", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Resize_365", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 128 : ui32, + config.dim_2 = 12 : ui32, + config.dim_3 = 20 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 24 : ui32, + config.output_W = 40 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_365", + OutputName = "Resize_365", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x24x40xbf16> loc(#loc243) + xten_nn.output %461 : tensor<1x128x24x40xbf16> loc(#loc243) + } -> tensor<1x128x24x40xbf16> loc(#loc243) + %390 = xten_nn.subgraph (%arg5 = %389: tensor<1x128x24x40xbf16>) attributes { + LayerName = "Slice_371", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + OutputName = "Slice_371", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 128 : ui32, + config.dim_h = 24 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 23 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_371", + OutputName = "Slice_371", + size = array, + start = array} : (tensor<1x128x24x40xbf16>) -> tensor<1x128x23x40xbf16> loc(#loc244) + xten_nn.output %461 : tensor<1x128x23x40xbf16> loc(#loc244) + } -> tensor<1x128x23x40xbf16> loc(#loc244) + %391 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>) attributes { + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc12) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc12) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_346", + PartOfOutputName = "AveragePool_346", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>) -> tensor<1x90x160x3xbf16> loc(#loc12) + %466 = tosa.transpose %465, %462 : (tensor<1x90x160x3xbf16>, tensor<4xi32>) -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %466 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %461 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + %392 = xten_nn.subgraph (%arg5 = %391: tensor<1x3x90x160xbf16>) attributes { + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x90x160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc245) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x3xbf16> loc(#loc245) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_347", + PartOfOutputName = "AveragePool_347", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x90x160x3xbf16>) -> tensor<1x45x80x3xbf16> loc(#loc245) + %466 = tosa.transpose %465, %462 : (tensor<1x45x80x3xbf16>, tensor<4xi32>) -> tensor<1x3x45x80xbf16> loc(#loc245) + xten_nn.output %466 : tensor<1x3x45x80xbf16> loc(#loc245) + } -> tensor<1x3x45x80xbf16> loc(#loc245) + xten_nn.output %461 : tensor<1x3x45x80xbf16> loc(#loc245) + } -> tensor<1x3x45x80xbf16> loc(#loc245) + %393 = xten_nn.subgraph (%arg5 = %392: tensor<1x3x45x80xbf16>) attributes { + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x45x80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 1], [0, 0]], + HWPaddingNotCounted = [[0, 1], [0, 0]], + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc246) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x3xbf16> loc(#loc246) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_348", + PartOfOutputName = "AveragePool_348", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x45x80x3xbf16>) -> tensor<1x23x40x3xbf16> loc(#loc246) + %466 = tosa.transpose %465, %462 : (tensor<1x23x40x3xbf16>, tensor<4xi32>) -> tensor<1x3x23x40xbf16> loc(#loc246) + xten_nn.output %466 : tensor<1x3x23x40xbf16> loc(#loc246) + } -> tensor<1x3x23x40xbf16> loc(#loc246) + xten_nn.output %461 : tensor<1x3x23x40xbf16> loc(#loc246) + } -> tensor<1x3x23x40xbf16> loc(#loc246) + %394 = xten_nn.subgraph (%arg5 = %390: tensor<1x128x23x40xbf16>, %arg6 = %217: tensor<1x40x23x40xbf16>, %arg7 = %393: tensor<1x3x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_372", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_372", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_372", + OutputName = "Concat_372", + axis = 1 : i32} : (tensor<1x128x23x40xbf16>, tensor<1x40x23x40xbf16>, tensor<1x3x23x40xbf16>) -> tensor<1x171x23x40xbf16> loc(#loc247) + xten_nn.output %461 : tensor<1x171x23x40xbf16> loc(#loc247) + } -> tensor<1x171x23x40xbf16> loc(#loc247) + %395 = xten_nn.subgraph (%arg5 = %394: tensor<1x171x23x40xbf16>, %arg6 = %26: tensor<80x171x3x3xbf16>, %arg7 = %25: tensor<80xbf16>) attributes { + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x171x23x40xbf16>, %arg9 = %arg6: tensor<80x171x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x171x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x171xbf16> loc(#loc348) + %465 = tosa.transpose %arg8, %463 : (tensor<1x171x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x171xbf16> loc(#loc348) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_373", + PartOfOutputName = "Conv_373", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x171xbf16>, tensor<80x3x3x171xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc248) + %467 = tosa.clamp %466 { + LayerName = "Relu_374", + OutputName = "Relu_374", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc249) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc348) + xten_nn.output %468 : tensor<1x80x23x40xbf16> loc(#loc249) + } -> tensor<1x80x23x40xbf16> loc(#loc348) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc348) + } -> tensor<1x80x23x40xbf16> loc(#loc348) + %396 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %397 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %398 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_376", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_376", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_376", + OutputName = "Concat_376", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc251) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc251) + } -> tensor<1x80x23x40xbf16> loc(#loc251) + %399 = xten_nn.subgraph (%arg5 = %398: tensor<1x80x23x40xbf16>, %arg6 = %24: tensor<80x80x3x3xbf16>, %arg7 = %23: tensor<80xbf16>) attributes { + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<80x80x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x80x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x80xbf16> loc(#loc252) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_377", + PartOfOutputName = "Conv_377", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<80x3x3x80xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %467 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + %400 = xten_nn.subgraph (%arg5 = %399: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_378", OutputName = "Sigmoid_378"} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %462 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + %401 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %402 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %403 = xten_nn.subgraph (%arg5 = %401: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_380", + OutputName = "Mul_380", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + %404 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %403: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_381", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_381", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_381", + OutputName = "Concat_381", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc256) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc256) + } -> tensor<1x80x23x40xbf16> loc(#loc256) + %405 = xten_nn.subgraph (%arg5 = %404: tensor<1x80x23x40xbf16>, %arg6 = %22: tensor<40x80x3x3xbf16>, %arg7 = %21: tensor<40xbf16>) attributes { + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<40x80x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x80x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x80xbf16> loc(#loc257) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc257) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_382", + PartOfOutputName = "Conv_382", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<40x3x3x80xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc257) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc257) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc257) + } -> tensor<1x40x23x40xbf16> loc(#loc257) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc257) + } -> tensor<1x40x23x40xbf16> loc(#loc257) + %406 = xten_nn.subgraph (%arg5 = %405: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_383", OutputName = "Tanh_383"} : (tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + %407 = xten_nn.subgraph (%arg5 = %402: tensor<1x40x23x40xbf16>, %arg6 = %406: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_387", + OutputName = "Mul_387", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + %408 = xten_nn.subgraph (%arg5 = %20: tensor<1x40x23x40xbf16>, %arg6 = %402: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_385", OutputName = "Sub_385"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc3) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc3) + } -> tensor<1x40x23x40xbf16> loc(#loc3) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc3) + } -> tensor<1x40x23x40xbf16> loc(#loc3) + %409 = xten_nn.subgraph (%arg5 = %408: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_386", + OutputName = "Mul_386", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + %410 = xten_nn.subgraph (%arg5 = %409: tensor<1x40x23x40xbf16>, %arg6 = %407: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_388", OutputName = "Add_388"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + %411 = xten_nn.subgraph (%arg5 = %396: tensor<1x40x23x40xbf16>, %arg6 = %410: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_389", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_389", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_389", + OutputName = "Concat_389", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc262) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc262) + } -> tensor<1x80x23x40xbf16> loc(#loc262) + %412 = xten_nn.subgraph (%arg5 = %411: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Resize_391", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Resize_391", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 80 : ui32, + config.dim_2 = 23 : ui32, + config.dim_3 = 40 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 46 : ui32, + config.output_W = 80 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_391", + OutputName = "Resize_391", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x46x80xbf16> loc(#loc263) + xten_nn.output %461 : tensor<1x80x46x80xbf16> loc(#loc263) + } -> tensor<1x80x46x80xbf16> loc(#loc263) + %413 = xten_nn.subgraph (%arg5 = %412: tensor<1x80x46x80xbf16>) attributes { + LayerName = "Slice_397", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + OutputName = "Slice_397", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 80 : ui32, + config.dim_h = 46 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 45 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_397", + OutputName = "Slice_397", + size = array, + start = array} : (tensor<1x80x46x80xbf16>) -> tensor<1x80x45x80xbf16> loc(#loc264) + xten_nn.output %461 : tensor<1x80x45x80xbf16> loc(#loc264) + } -> tensor<1x80x45x80xbf16> loc(#loc264) + %414 = xten_nn.subgraph (%arg5 = %413: tensor<1x80x45x80xbf16>, %arg6 = %181: tensor<1x24x45x80xbf16>, %arg7 = %392: tensor<1x3x45x80xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_398", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_398", + OutputName = "Concat_398", + axis = 1 : i32} : (tensor<1x80x45x80xbf16>, tensor<1x24x45x80xbf16>, tensor<1x3x45x80xbf16>) -> tensor<1x107x45x80xbf16> loc(#loc265) + xten_nn.output %461 : tensor<1x107x45x80xbf16> loc(#loc265) + } -> tensor<1x107x45x80xbf16> loc(#loc265) + %415 = xten_nn.subgraph (%arg5 = %414: tensor<1x107x45x80xbf16>, %arg6 = %19: tensor<40x107x3x3xbf16>, %arg7 = %18: tensor<40xbf16>) attributes { + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x107x45x80xbf16>, %arg9 = %arg6: tensor<40x107x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x107x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x107xbf16> loc(#loc349) + %465 = tosa.transpose %arg8, %463 : (tensor<1x107x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x107xbf16> loc(#loc349) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_399", + PartOfOutputName = "Conv_399", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x107xbf16>, tensor<40x3x3x107xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc266) + %467 = tosa.clamp %466 { + LayerName = "Relu_400", + OutputName = "Relu_400", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc267) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc349) + xten_nn.output %468 : tensor<1x40x45x80xbf16> loc(#loc267) + } -> tensor<1x40x45x80xbf16> loc(#loc349) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc349) + } -> tensor<1x40x45x80xbf16> loc(#loc349) + %416 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %417 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %418 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_402", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_402", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_402", + OutputName = "Concat_402", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc269) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc269) + } -> tensor<1x40x45x80xbf16> loc(#loc269) + %419 = xten_nn.subgraph (%arg5 = %418: tensor<1x40x45x80xbf16>, %arg6 = %17: tensor<40x40x3x3xbf16>, %arg7 = %16: tensor<40xbf16>) attributes { + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<40x40x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x40x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x40xbf16> loc(#loc270) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_403", + PartOfOutputName = "Conv_403", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<40x3x3x40xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %467 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + %420 = xten_nn.subgraph (%arg5 = %419: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_404", OutputName = "Sigmoid_404"} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %462 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + %421 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %422 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %423 = xten_nn.subgraph (%arg5 = %421: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_406", + OutputName = "Mul_406", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + %424 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %423: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_407", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_407", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_407", + OutputName = "Concat_407", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc274) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc274) + } -> tensor<1x40x45x80xbf16> loc(#loc274) + %425 = xten_nn.subgraph (%arg5 = %424: tensor<1x40x45x80xbf16>, %arg6 = %15: tensor<20x40x3x3xbf16>, %arg7 = %14: tensor<20xbf16>) attributes { + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<20x40x3x3xbf16>, %arg10 = %arg7: tensor<20xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<20x40x3x3xbf16>, tensor<4xi32>) -> tensor<20x3x3x40xbf16> loc(#loc275) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc275) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_408", + PartOfOutputName = "Conv_408", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<20x3x3x40xbf16>, tensor<20xbf16>) -> tensor<1x45x80x20xbf16> loc(#loc275) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x20xbf16>, tensor<4xi32>) -> tensor<1x20x45x80xbf16> loc(#loc275) + xten_nn.output %467 : tensor<1x20x45x80xbf16> loc(#loc275) + } -> tensor<1x20x45x80xbf16> loc(#loc275) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc275) + } -> tensor<1x20x45x80xbf16> loc(#loc275) + %426 = xten_nn.subgraph (%arg5 = %425: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_409", OutputName = "Tanh_409"} : (tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + %427 = xten_nn.subgraph (%arg5 = %422: tensor<1x20x45x80xbf16>, %arg6 = %426: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_413", + OutputName = "Mul_413", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + %428 = xten_nn.subgraph (%arg5 = %13: tensor<1x20x45x80xbf16>, %arg6 = %422: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_411", OutputName = "Sub_411"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc2) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc2) + } -> tensor<1x20x45x80xbf16> loc(#loc2) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc2) + } -> tensor<1x20x45x80xbf16> loc(#loc2) + %429 = xten_nn.subgraph (%arg5 = %428: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_412", + OutputName = "Mul_412", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + %430 = xten_nn.subgraph (%arg5 = %429: tensor<1x20x45x80xbf16>, %arg6 = %427: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_414", OutputName = "Add_414"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + %431 = xten_nn.subgraph (%arg5 = %416: tensor<1x20x45x80xbf16>, %arg6 = %430: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_415", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_415", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_415", + OutputName = "Concat_415", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc280) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc280) + } -> tensor<1x40x45x80xbf16> loc(#loc280) + %432 = xten_nn.subgraph (%arg5 = %431: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Resize_417", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Resize_417", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 40 : ui32, + config.dim_2 = 45 : ui32, + config.dim_3 = 80 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 90 : ui32, + config.output_W = 160 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_417", + OutputName = "Resize_417", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x90x160xbf16> loc(#loc281) + xten_nn.output %461 : tensor<1x40x90x160xbf16> loc(#loc281) + } -> tensor<1x40x90x160xbf16> loc(#loc281) + %433 = xten_nn.subgraph (%arg5 = %432: tensor<1x40x90x160xbf16>, %arg6 = %175: tensor<1x16x90x160xbf16>, %arg7 = %391: tensor<1x3x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_418", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_418", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_418", + OutputName = "Concat_418", + axis = 1 : i32} : (tensor<1x40x90x160xbf16>, tensor<1x16x90x160xbf16>, tensor<1x3x90x160xbf16>) -> tensor<1x59x90x160xbf16> loc(#loc282) + xten_nn.output %461 : tensor<1x59x90x160xbf16> loc(#loc282) + } -> tensor<1x59x90x160xbf16> loc(#loc282) + %434 = xten_nn.subgraph (%arg5 = %433: tensor<1x59x90x160xbf16>, %arg6 = %12: tensor<32x59x3x3xbf16>, %arg7 = %11: tensor<32xbf16>) attributes { + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x59x90x160xbf16>, %arg9 = %arg6: tensor<32x59x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x59x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x59xbf16> loc(#loc350) + %465 = tosa.transpose %arg8, %463 : (tensor<1x59x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x59xbf16> loc(#loc350) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_419", + PartOfOutputName = "Conv_419", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x59xbf16>, tensor<32x3x3x59xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc283) + %467 = tosa.clamp %466 { + LayerName = "Relu_420", + OutputName = "Relu_420", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc284) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc350) + xten_nn.output %468 : tensor<1x32x90x160xbf16> loc(#loc284) + } -> tensor<1x32x90x160xbf16> loc(#loc350) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc350) + } -> tensor<1x32x90x160xbf16> loc(#loc350) + %435 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %436 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %437 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_422", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_422", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_422", + OutputName = "Concat_422", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc286) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc286) + } -> tensor<1x32x90x160xbf16> loc(#loc286) + %438 = xten_nn.subgraph (%arg5 = %437: tensor<1x32x90x160xbf16>, %arg6 = %10: tensor<32x32x3x3xbf16>, %arg7 = %9: tensor<32xbf16>) attributes { + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<32x32x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x32x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x32xbf16> loc(#loc287) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_423", + PartOfOutputName = "Conv_423", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<32x3x3x32xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %467 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + %439 = xten_nn.subgraph (%arg5 = %438: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_424", OutputName = "Sigmoid_424"} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %462 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + %440 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %441 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %442 = xten_nn.subgraph (%arg5 = %440: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_426", + OutputName = "Mul_426", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + %443 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %442: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_427", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_427", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_427", + OutputName = "Concat_427", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc291) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc291) + } -> tensor<1x32x90x160xbf16> loc(#loc291) + %444 = xten_nn.subgraph (%arg5 = %443: tensor<1x32x90x160xbf16>, %arg6 = %8: tensor<16x32x3x3xbf16>, %arg7 = %7: tensor<16xbf16>) attributes { + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<16x32x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x32x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x32xbf16> loc(#loc292) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc292) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_428", + PartOfOutputName = "Conv_428", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<16x3x3x32xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc292) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc292) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc292) + } -> tensor<1x16x90x160xbf16> loc(#loc292) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc292) + } -> tensor<1x16x90x160xbf16> loc(#loc292) + %445 = xten_nn.subgraph (%arg5 = %444: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_429", OutputName = "Tanh_429"} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + %446 = xten_nn.subgraph (%arg5 = %441: tensor<1x16x90x160xbf16>, %arg6 = %445: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_433", + OutputName = "Mul_433", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + %447 = xten_nn.subgraph (%arg5 = %6: tensor<1x16x90x160xbf16>, %arg6 = %441: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_431", OutputName = "Sub_431"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc1) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc1) + } -> tensor<1x16x90x160xbf16> loc(#loc1) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc1) + } -> tensor<1x16x90x160xbf16> loc(#loc1) + %448 = xten_nn.subgraph (%arg5 = %447: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_432", + OutputName = "Mul_432", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + %449 = xten_nn.subgraph (%arg5 = %448: tensor<1x16x90x160xbf16>, %arg6 = %446: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_434", OutputName = "Add_434"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + %450 = xten_nn.subgraph (%arg5 = %435: tensor<1x16x90x160xbf16>, %arg6 = %449: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_435", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_435", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_435", + OutputName = "Concat_435", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc297) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc297) + } -> tensor<1x32x90x160xbf16> loc(#loc297) + %451 = xten_nn.subgraph (%arg5 = %450: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Resize_437", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Resize_437", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 32 : ui32, + config.dim_2 = 90 : ui32, + config.dim_3 = 160 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 180 : ui32, + config.output_W = 320 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_437", + OutputName = "Resize_437", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x180x320xbf16> loc(#loc298) + xten_nn.output %461 : tensor<1x32x180x320xbf16> loc(#loc298) + } -> tensor<1x32x180x320xbf16> loc(#loc298) + %452 = xten_nn.subgraph (%arg5 = %451: tensor<1x32x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_438", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Concat_438", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_438", + OutputName = "Concat_438", + axis = 1 : i32} : (tensor<1x32x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x35x180x320xbf16> loc(#loc299) + xten_nn.output %461 : tensor<1x35x180x320xbf16> loc(#loc299) + } -> tensor<1x35x180x320xbf16> loc(#loc299) + %453 = xten_nn.subgraph (%arg5 = %452: tensor<1x35x180x320xbf16>, %arg6 = %5: tensor<16x35x3x3xbf16>, %arg7 = %4: tensor<16xbf16>) attributes { + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x35x180x320xbf16>, %arg9 = %arg6: tensor<16x35x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x35x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x35xbf16> loc(#loc351) + %465 = tosa.transpose %arg8, %463 : (tensor<1x35x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x35xbf16> loc(#loc351) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_439", + PartOfOutputName = "Conv_439", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x35xbf16>, tensor<16x3x3x35xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc300) + %467 = tosa.clamp %466 { + LayerName = "Relu_440", + OutputName = "Relu_440", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc301) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc351) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc301) + } -> tensor<1x16x180x320xbf16> loc(#loc351) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc351) + } -> tensor<1x16x180x320xbf16> loc(#loc351) + %454 = xten_nn.subgraph (%arg5 = %453: tensor<1x16x180x320xbf16>, %arg6 = %3: tensor<16x16x3x3xbf16>, %arg7 = %2: tensor<16xbf16>) attributes { + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<16x16x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x16x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x16xbf16> loc(#loc352) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc352) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_441", + PartOfOutputName = "Conv_441", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<16x3x3x16xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc302) + %467 = tosa.clamp %466 { + LayerName = "Relu_442", + OutputName = "Relu_442", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc303) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc352) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc303) + } -> tensor<1x16x180x320xbf16> loc(#loc352) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc352) + } -> tensor<1x16x180x320xbf16> loc(#loc352) + %455 = xten_nn.subgraph (%arg5 = %454: tensor<1x16x180x320xbf16>, %arg6 = %1: tensor<4x16x1x1xbf16>, %arg7 = %0: tensor<4xbf16>) attributes { + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<4x16x1x1xbf16>, %arg10 = %arg7: tensor<4xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc304) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<4x16x1x1xbf16>) -> tensor<4x1x1x16xbf16> loc(#loc304) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc304) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_443", + PartOfOutputName = "Conv_443", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<4x1x1x16xbf16>, tensor<4xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc304) + %467 = tosa.transpose %466, %462 : (tensor<1x180x320x4xbf16>, tensor<4xi32>) -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %467 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %461 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + %456 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc305) + } -> tensor<1x3x180x320xbf16> loc(#loc305) + %457 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 4 : ui32, + config.start = 3 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc305) + } -> tensor<1x1x180x320xbf16> loc(#loc305) + %458 = xten_nn.subgraph (%arg5 = %457: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_447", + OutputName = "Clip_447", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc306) + xten_nn.output %462 : tensor<1x1x180x320xbf16> loc(#loc306) + } -> tensor<1x1x180x320xbf16> loc(#loc306) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc306) + } -> tensor<1x1x180x320xbf16> loc(#loc306) + %459 = xten_nn.subgraph (%arg5 = %456: tensor<1x3x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_445", OutputName = "Add_445"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + %460 = xten_nn.subgraph (%arg5 = %459: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_446", + OutputName = "Clip_446", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc307) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc307) + } -> tensor<1x3x180x320xbf16> loc(#loc307) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc307) + } -> tensor<1x3x180x320xbf16> loc(#loc307) + return %458, %449, %430, %410, %387, %460 : tensor<1x1x180x320xbf16>, tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16> loc(#loc) + } loc(#loc) +} loc(#loc) +#loc1 = loc("Sub_431") +#loc2 = loc("Sub_411") +#loc3 = loc("Sub_385") +#loc4 = loc("Sub_359") +#loc5 = loc("Div_16") +#loc6 = loc("Sub_14") +#loc7 = loc("Initializer_398") +#loc8 = loc("Div_2") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("Conv_17") +#loc14 = loc("Add_19") +#loc15 = loc("Clip_22") +#loc16 = loc("Div_24") +#loc17 = loc("Mul_25") +#loc18 = loc("Conv_26") +#loc19 = loc("Relu_27") +#loc20 = loc("Conv_28") +#loc21 = loc("Add_29") +#loc22 = loc("Conv_30") +#loc23 = loc("Relu_31") +#loc24 = loc("Conv_32") +#loc25 = loc("Relu_33") +#loc26 = loc("Conv_34") +#loc27 = loc("Conv_35") +#loc28 = loc("Relu_36") +#loc29 = loc("Conv_37") +#loc30 = loc("Relu_38") +#loc31 = loc("Conv_39") +#loc32 = loc("Add_40") +#loc33 = loc("Conv_41") +#loc34 = loc("Relu_42") +#loc35 = loc("Conv_43") +#loc36 = loc("Relu_44") +#loc37 = loc("GlobalAveragePool_45") +#loc38 = loc("Conv_46") +#loc39 = loc("Relu_47") +#loc40 = loc("Conv_48") +#loc41 = loc("Add_50") +#loc42 = loc("Clip_53") +#loc43 = loc("Div_55") +#loc44 = loc("Mul_56") +#loc45 = loc("Conv_57") +#loc46 = loc("Conv_58") +#loc47 = loc("Relu_59") +#loc48 = loc("Conv_60") +#loc49 = loc("Relu_61") +#loc50 = loc("GlobalAveragePool_62") +#loc51 = loc("Conv_63") +#loc52 = loc("Relu_64") +#loc53 = loc("Conv_65") +#loc54 = loc("Add_67") +#loc55 = loc("Clip_70") +#loc56 = loc("Div_72") +#loc57 = loc("Mul_73") +#loc58 = loc("Conv_74") +#loc59 = loc("Add_75") +#loc60 = loc("Conv_76") +#loc61 = loc("Relu_77") +#loc62 = loc("Conv_78") +#loc63 = loc("Relu_79") +#loc64 = loc("GlobalAveragePool_80") +#loc65 = loc("Conv_81") +#loc66 = loc("Relu_82") +#loc67 = loc("Conv_83") +#loc68 = loc("Add_85") +#loc69 = loc("Clip_88") +#loc70 = loc("Div_90") +#loc71 = loc("Mul_91") +#loc72 = loc("Conv_92") +#loc73 = loc("Add_93") +#loc74 = loc("Conv_94") +#loc75 = loc("Add_96") +#loc76 = loc("Clip_99") +#loc77 = loc("Div_101") +#loc78 = loc("Mul_102") +#loc79 = loc("Conv_103") +#loc80 = loc("Add_105") +#loc81 = loc("Clip_108") +#loc82 = loc("Div_110") +#loc83 = loc("Mul_111") +#loc84 = loc("Conv_112") +#loc85 = loc("Conv_113") +#loc86 = loc("Add_115") +#loc87 = loc("Clip_118") +#loc88 = loc("Div_120") +#loc89 = loc("Mul_121") +#loc90 = loc("Conv_122") +#loc91 = loc("Add_124") +#loc92 = loc("Clip_127") +#loc93 = loc("Div_129") +#loc94 = loc("Mul_130") +#loc95 = loc("Conv_131") +#loc96 = loc("Add_132") +#loc97 = loc("Conv_133") +#loc98 = loc("Add_135") +#loc99 = loc("Clip_138") +#loc100 = loc("Div_140") +#loc101 = loc("Mul_141") +#loc102 = loc("Conv_142") +#loc103 = loc("Add_144") +#loc104 = loc("Clip_147") +#loc105 = loc("Div_149") +#loc106 = loc("Mul_150") +#loc107 = loc("Conv_151") +#loc108 = loc("Add_152") +#loc109 = loc("Conv_153") +#loc110 = loc("Add_155") +#loc111 = loc("Clip_158") +#loc112 = loc("Div_160") +#loc113 = loc("Mul_161") +#loc114 = loc("Conv_162") +#loc115 = loc("Add_164") +#loc116 = loc("Clip_167") +#loc117 = loc("Div_169") +#loc118 = loc("Mul_170") +#loc119 = loc("Conv_171") +#loc120 = loc("Add_172") +#loc121 = loc("Conv_173") +#loc122 = loc("Add_175") +#loc123 = loc("Clip_178") +#loc124 = loc("Div_180") +#loc125 = loc("Mul_181") +#loc126 = loc("Conv_182") +#loc127 = loc("Add_184") +#loc128 = loc("Clip_187") +#loc129 = loc("Div_189") +#loc130 = loc("Mul_190") +#loc131 = loc("GlobalAveragePool_191") +#loc132 = loc("Conv_192") +#loc133 = loc("Relu_193") +#loc134 = loc("Conv_194") +#loc135 = loc("Add_196") +#loc136 = loc("Clip_199") +#loc137 = loc("Div_201") +#loc138 = loc("Mul_202") +#loc139 = loc("Conv_203") +#loc140 = loc("Conv_204") +#loc141 = loc("Add_206") +#loc142 = loc("Clip_209") +#loc143 = loc("Div_211") +#loc144 = loc("Mul_212") +#loc145 = loc("Conv_213") +#loc146 = loc("Add_215") +#loc147 = loc("Clip_218") +#loc148 = loc("Div_220") +#loc149 = loc("Mul_221") +#loc150 = loc("GlobalAveragePool_222") +#loc151 = loc("Conv_223") +#loc152 = loc("Relu_224") +#loc153 = loc("Conv_225") +#loc154 = loc("Add_227") +#loc155 = loc("Clip_230") +#loc156 = loc("Div_232") +#loc157 = loc("Mul_233") +#loc158 = loc("Conv_234") +#loc159 = loc("Add_235") +#loc160 = loc("Conv_236") +#loc161 = loc("Add_238") +#loc162 = loc("Clip_241") +#loc163 = loc("Div_243") +#loc164 = loc("Mul_244") +#loc165 = loc("Conv_245") +#loc166 = loc("Add_247") +#loc167 = loc("Clip_250") +#loc168 = loc("Div_252") +#loc169 = loc("Mul_253") +#loc170 = loc("GlobalAveragePool_254") +#loc171 = loc("Conv_255") +#loc172 = loc("Relu_256") +#loc173 = loc("Conv_257") +#loc174 = loc("Add_259") +#loc175 = loc("Clip_262") +#loc176 = loc("Div_264") +#loc177 = loc("Mul_265") +#loc178 = loc("Conv_266") +#loc179 = loc("Conv_267") +#loc180 = loc("Add_269") +#loc181 = loc("Clip_272") +#loc182 = loc("Div_274") +#loc183 = loc("Mul_275") +#loc184 = loc("Conv_276") +#loc185 = loc("Add_278") +#loc186 = loc("Clip_281") +#loc187 = loc("Div_283") +#loc188 = loc("Mul_284") +#loc189 = loc("GlobalAveragePool_285") +#loc190 = loc("Conv_286") +#loc191 = loc("Relu_287") +#loc192 = loc("Conv_288") +#loc193 = loc("Add_290") +#loc194 = loc("Clip_293") +#loc195 = loc("Div_295") +#loc196 = loc("Mul_296") +#loc197 = loc("Conv_297") +#loc198 = loc("Add_298") +#loc199 = loc("Conv_299") +#loc200 = loc("Add_301") +#loc201 = loc("Clip_304") +#loc202 = loc("Div_306") +#loc203 = loc("Mul_307") +#loc204 = loc("Conv_308") +#loc205 = loc("Add_310") +#loc206 = loc("Clip_313") +#loc207 = loc("Div_315") +#loc208 = loc("Mul_316") +#loc209 = loc("GlobalAveragePool_317") +#loc210 = loc("Conv_318") +#loc211 = loc("Relu_319") +#loc212 = loc("Conv_320") +#loc213 = loc("Add_322") +#loc214 = loc("Clip_325") +#loc215 = loc("Div_327") +#loc216 = loc("Mul_328") +#loc217 = loc("Conv_329") +#loc218 = loc("Add_330") +#loc219 = loc("Conv_331") +#loc220 = loc("Add_333") +#loc221 = loc("Clip_336") +#loc222 = loc("Div_338") +#loc223 = loc("Mul_339") +#loc224 = loc("GlobalAveragePool_342") +#loc225 = loc("Conv_343") +#loc226 = loc("Sigmoid_344") +#loc227 = loc("Mul_345") +#loc228 = loc("Conv_340") +#loc229 = loc("Relu_341") +#loc230 = loc("Split_349") +#loc231 = loc("Concat_350") +#loc232 = loc("Conv_351") +#loc233 = loc("Sigmoid_352") +#loc234 = loc("Split_353") +#loc235 = loc("Mul_354") +#loc236 = loc("Concat_355") +#loc237 = loc("Conv_356") +#loc238 = loc("Tanh_357") +#loc239 = loc("Mul_361") +#loc240 = loc("Mul_360") +#loc241 = loc("Add_362") +#loc242 = loc("Concat_363") +#loc243 = loc("Resize_365") +#loc244 = loc("Slice_371") +#loc245 = loc("AveragePool_347") +#loc246 = loc("AveragePool_348") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_380") +#loc256 = loc("Concat_381") +#loc257 = loc("Conv_382") +#loc258 = loc("Tanh_383") +#loc259 = loc("Mul_387") +#loc260 = loc("Mul_386") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_406") +#loc274 = loc("Concat_407") +#loc275 = loc("Conv_408") +#loc276 = loc("Tanh_409") +#loc277 = loc("Mul_413") +#loc278 = loc("Mul_412") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_426") +#loc291 = loc("Concat_427") +#loc292 = loc("Conv_428") +#loc293 = loc("Tanh_429") +#loc294 = loc("Mul_433") +#loc295 = loc("Mul_432") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_447") +#loc307 = loc("Clip_446") +#loc308 = loc(fused[#loc6, #loc7]) +#loc309 = loc(fused[#loc11, #loc9, #loc12]) +#loc310 = loc(fused[#loc9, #loc12, #loc11]) +#loc311 = loc(fused[#loc18, #loc19]) +#loc312 = loc(fused[#loc20, #loc21]) +#loc313 = loc(fused[#loc22, #loc23]) +#loc314 = loc(fused[#loc24, #loc25]) +#loc315 = loc(fused[#loc27, #loc28]) +#loc316 = loc(fused[#loc29, #loc30]) +#loc317 = loc(fused[#loc31, #loc32]) +#loc318 = loc(fused[#loc33, #loc34]) +#loc319 = loc(fused[#loc35, #loc36]) +#loc320 = loc(fused[#loc38, #loc39]) +#loc321 = loc(fused[#loc46, #loc47]) +#loc322 = loc(fused[#loc48, #loc49]) +#loc323 = loc(fused[#loc51, #loc52]) +#loc324 = loc(fused[#loc58, #loc59]) +#loc325 = loc(fused[#loc60, #loc61]) +#loc326 = loc(fused[#loc62, #loc63]) +#loc327 = loc(fused[#loc65, #loc66]) +#loc328 = loc(fused[#loc72, #loc73]) +#loc329 = loc(fused[#loc95, #loc96]) +#loc330 = loc(fused[#loc107, #loc108]) +#loc331 = loc(fused[#loc119, #loc120]) +#loc332 = loc(fused[#loc130, #loc131]) +#loc333 = loc(fused[#loc132, #loc133]) +#loc334 = loc(fused[#loc149, #loc150]) +#loc335 = loc(fused[#loc151, #loc152]) +#loc336 = loc(fused[#loc158, #loc159]) +#loc337 = loc(fused[#loc169, #loc170]) +#loc338 = loc(fused[#loc171, #loc172]) +#loc339 = loc(fused[#loc188, #loc189]) +#loc340 = loc(fused[#loc190, #loc191]) +#loc341 = loc(fused[#loc197, #loc198]) +#loc342 = loc(fused[#loc208, #loc209]) +#loc343 = loc(fused[#loc210, #loc211]) +#loc344 = loc(fused[#loc217, #loc218]) +#loc345 = loc(fused[#loc223, #loc224]) +#loc346 = loc(fused[#loc228, #loc229, #loc227]) +#loc347 = loc(fused[#loc228, #loc229]) +#loc348 = loc(fused[#loc248, #loc249]) +#loc349 = loc(fused[#loc266, #loc267]) +#loc350 = loc(fused[#loc283, #loc284]) +#loc351 = loc(fused[#loc300, #loc301]) +#loc352 = loc(fused[#loc302, #loc303]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.ll b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.ll new file mode 100644 index 0000000000000000000000000000000000000000..7f918d9e3422d223534f795925e8688a73c8034f --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.ll @@ -0,0 +1,428 @@ +; ModuleID = 'LLVMDialectModule' +source_filename = "LLVMDialectModule" +target datalayout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128" +target triple = "x86_64-unknown-linux-gnu" + +declare void @free(ptr) + +declare ptr @malloc(i64) + +declare { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } @forward_outlined_part_0(ptr, ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, ptr, ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, ptr, ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, ptr, ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, ptr, ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64) + +define { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } @forward(ptr %0, ptr %1, i64 %2, i64 %3, i64 %4, i64 %5, i64 %6, i64 %7, i64 %8, i64 %9, i64 %10, ptr %11, ptr %12, i64 %13, i64 %14, i64 %15, i64 %16, i64 %17, i64 %18, i64 %19, i64 %20, i64 %21, ptr %22, ptr %23, i64 %24, i64 %25, i64 %26, i64 %27, i64 %28, i64 %29, i64 %30, i64 %31, i64 %32, ptr %33, ptr %34, i64 %35, i64 %36, i64 %37, i64 %38, i64 %39, i64 %40, i64 %41, i64 %42, i64 %43, ptr %44, ptr %45, i64 %46, i64 %47, i64 %48, i64 %49, i64 %50, i64 %51, i64 %52, i64 %53, i64 %54) { + %56 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } undef, ptr %44, 0 + %57 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %56, ptr %45, 1 + %58 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %57, i64 %46, 2 + %59 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %58, i64 %47, 3, 0 + %60 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %59, i64 %51, 4, 0 + %61 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %60, i64 %48, 3, 1 + %62 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %61, i64 %52, 4, 1 + %63 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %62, i64 %49, 3, 2 + %64 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %63, i64 %53, 4, 2 + %65 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %64, i64 %50, 3, 3 + %66 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %65, i64 %54, 4, 3 + %67 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } undef, ptr %33, 0 + %68 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %67, ptr %34, 1 + %69 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %68, i64 %35, 2 + %70 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %69, i64 %36, 3, 0 + %71 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %70, i64 %40, 4, 0 + %72 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %71, i64 %37, 3, 1 + %73 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %72, i64 %41, 4, 1 + %74 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %73, i64 %38, 3, 2 + %75 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %74, i64 %42, 4, 2 + %76 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %75, i64 %39, 3, 3 + %77 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %76, i64 %43, 4, 3 + %78 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } undef, ptr %22, 0 + %79 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %78, ptr %23, 1 + %80 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %79, i64 %24, 2 + %81 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %80, i64 %25, 3, 0 + %82 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %81, i64 %29, 4, 0 + %83 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %82, i64 %26, 3, 1 + %84 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %83, i64 %30, 4, 1 + %85 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %84, i64 %27, 3, 2 + %86 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %85, i64 %31, 4, 2 + %87 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %86, i64 %28, 3, 3 + %88 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %87, i64 %32, 4, 3 + %89 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } undef, ptr %11, 0 + %90 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %89, ptr %12, 1 + %91 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %90, i64 %13, 2 + %92 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %91, i64 %14, 3, 0 + %93 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %92, i64 %18, 4, 0 + %94 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %93, i64 %15, 3, 1 + %95 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %94, i64 %19, 4, 1 + %96 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %95, i64 %16, 3, 2 + %97 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %96, i64 %20, 4, 2 + %98 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %97, i64 %17, 3, 3 + %99 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %98, i64 %21, 4, 3 + %100 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } undef, ptr %0, 0 + %101 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %100, ptr %1, 1 + %102 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %101, i64 %2, 2 + %103 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %102, i64 %3, 3, 0 + %104 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %103, i64 %7, 4, 0 + %105 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %104, i64 %4, 3, 1 + %106 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %105, i64 %8, 4, 1 + %107 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %106, i64 %5, 3, 2 + %108 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %107, i64 %9, 4, 2 + %109 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %108, i64 %6, 3, 3 + %110 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %109, i64 %10, 4, 3 + %111 = call ptr @malloc(i64 add (i64 ptrtoint (ptr getelementptr (float, ptr null, i64 230400) to i64), i64 64)) + %112 = ptrtoint ptr %111 to i64 + %113 = add i64 %112, 63 + %114 = urem i64 %113, 64 + %115 = sub i64 %113, %114 + %116 = inttoptr i64 %115 to ptr + %117 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } undef, ptr %111, 0 + %118 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %117, ptr %116, 1 + %119 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %118, i64 0, 2 + %120 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %119, i64 1, 3, 0 + %121 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %120, i64 180, 3, 1 + %122 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %121, i64 320, 3, 2 + %123 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %122, i64 4, 3, 3 + %124 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %123, i64 230400, 4, 0 + %125 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %124, i64 1280, 4, 1 + %126 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %125, i64 4, 4, 2 + %127 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %126, i64 1, 4, 3 + %128 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %110, 3, 0 + %129 = mul i64 1, %128 + %130 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %110, 3, 1 + %131 = mul i64 %129, %130 + %132 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %110, 3, 2 + %133 = mul i64 %131, %132 + %134 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %110, 3, 3 + %135 = mul i64 %133, %134 + %136 = mul i64 %135, ptrtoint (ptr getelementptr (float, ptr null, i32 1) to i64) + %137 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %110, 1 + %138 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %110, 2 + %139 = getelementptr float, ptr %137, i64 %138 + %140 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 1 + %141 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 2 + %142 = getelementptr float, ptr %140, i64 %141 + call void @llvm.memcpy.p0.p0.i64(ptr %142, ptr %139, i64 %136, i1 false) + %143 = call ptr @malloc(i64 add (i64 ptrtoint (ptr getelementptr (float, ptr null, i64 230400) to i64), i64 64)) + %144 = ptrtoint ptr %143 to i64 + %145 = add i64 %144, 63 + %146 = urem i64 %145, 64 + %147 = sub i64 %145, %146 + %148 = inttoptr i64 %147 to ptr + %149 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } undef, ptr %143, 0 + %150 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %149, ptr %148, 1 + %151 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %150, i64 0, 2 + %152 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %151, i64 1, 3, 0 + %153 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %152, i64 16, 3, 1 + %154 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %153, i64 90, 3, 2 + %155 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %154, i64 160, 3, 3 + %156 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %155, i64 230400, 4, 0 + %157 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %156, i64 14400, 4, 1 + %158 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %157, i64 160, 4, 2 + %159 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %158, i64 1, 4, 3 + %160 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %99, 3, 0 + %161 = mul i64 1, %160 + %162 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %99, 3, 1 + %163 = mul i64 %161, %162 + %164 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %99, 3, 2 + %165 = mul i64 %163, %164 + %166 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %99, 3, 3 + %167 = mul i64 %165, %166 + %168 = mul i64 %167, ptrtoint (ptr getelementptr (float, ptr null, i32 1) to i64) + %169 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %99, 1 + %170 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %99, 2 + %171 = getelementptr float, ptr %169, i64 %170 + %172 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 1 + %173 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 2 + %174 = getelementptr float, ptr %172, i64 %173 + call void @llvm.memcpy.p0.p0.i64(ptr %174, ptr %171, i64 %168, i1 false) + %175 = call ptr @malloc(i64 add (i64 ptrtoint (ptr getelementptr (float, ptr null, i64 72000) to i64), i64 64)) + %176 = ptrtoint ptr %175 to i64 + %177 = add i64 %176, 63 + %178 = urem i64 %177, 64 + %179 = sub i64 %177, %178 + %180 = inttoptr i64 %179 to ptr + %181 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } undef, ptr %175, 0 + %182 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %181, ptr %180, 1 + %183 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %182, i64 0, 2 + %184 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %183, i64 1, 3, 0 + %185 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %184, i64 20, 3, 1 + %186 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %185, i64 45, 3, 2 + %187 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %186, i64 80, 3, 3 + %188 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %187, i64 72000, 4, 0 + %189 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %188, i64 3600, 4, 1 + %190 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %189, i64 80, 4, 2 + %191 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %190, i64 1, 4, 3 + %192 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %88, 3, 0 + %193 = mul i64 1, %192 + %194 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %88, 3, 1 + %195 = mul i64 %193, %194 + %196 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %88, 3, 2 + %197 = mul i64 %195, %196 + %198 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %88, 3, 3 + %199 = mul i64 %197, %198 + %200 = mul i64 %199, ptrtoint (ptr getelementptr (float, ptr null, i32 1) to i64) + %201 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %88, 1 + %202 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %88, 2 + %203 = getelementptr float, ptr %201, i64 %202 + %204 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 1 + %205 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 2 + %206 = getelementptr float, ptr %204, i64 %205 + call void @llvm.memcpy.p0.p0.i64(ptr %206, ptr %203, i64 %200, i1 false) + %207 = call ptr @malloc(i64 add (i64 ptrtoint (ptr getelementptr (float, ptr null, i64 36800) to i64), i64 64)) + %208 = ptrtoint ptr %207 to i64 + %209 = add i64 %208, 63 + %210 = urem i64 %209, 64 + %211 = sub i64 %209, %210 + %212 = inttoptr i64 %211 to ptr + %213 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } undef, ptr %207, 0 + %214 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %213, ptr %212, 1 + %215 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %214, i64 0, 2 + %216 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %215, i64 1, 3, 0 + %217 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %216, i64 40, 3, 1 + %218 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %217, i64 23, 3, 2 + %219 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %218, i64 40, 3, 3 + %220 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %219, i64 36800, 4, 0 + %221 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %220, i64 920, 4, 1 + %222 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %221, i64 40, 4, 2 + %223 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %222, i64 1, 4, 3 + %224 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %77, 3, 0 + %225 = mul i64 1, %224 + %226 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %77, 3, 1 + %227 = mul i64 %225, %226 + %228 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %77, 3, 2 + %229 = mul i64 %227, %228 + %230 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %77, 3, 3 + %231 = mul i64 %229, %230 + %232 = mul i64 %231, ptrtoint (ptr getelementptr (float, ptr null, i32 1) to i64) + %233 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %77, 1 + %234 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %77, 2 + %235 = getelementptr float, ptr %233, i64 %234 + %236 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 1 + %237 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 2 + %238 = getelementptr float, ptr %236, i64 %237 + call void @llvm.memcpy.p0.p0.i64(ptr %238, ptr %235, i64 %232, i1 false) + %239 = call ptr @malloc(i64 add (i64 ptrtoint (ptr getelementptr (float, ptr null, i64 15360) to i64), i64 64)) + %240 = ptrtoint ptr %239 to i64 + %241 = add i64 %240, 63 + %242 = urem i64 %241, 64 + %243 = sub i64 %241, %242 + %244 = inttoptr i64 %243 to ptr + %245 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } undef, ptr %239, 0 + %246 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %245, ptr %244, 1 + %247 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %246, i64 0, 2 + %248 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %247, i64 1, 3, 0 + %249 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %248, i64 64, 3, 1 + %250 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %249, i64 12, 3, 2 + %251 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %250, i64 20, 3, 3 + %252 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %251, i64 15360, 4, 0 + %253 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %252, i64 240, 4, 1 + %254 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %253, i64 20, 4, 2 + %255 = insertvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %254, i64 1, 4, 3 + %256 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %66, 3, 0 + %257 = mul i64 1, %256 + %258 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %66, 3, 1 + %259 = mul i64 %257, %258 + %260 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %66, 3, 2 + %261 = mul i64 %259, %260 + %262 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %66, 3, 3 + %263 = mul i64 %261, %262 + %264 = mul i64 %263, ptrtoint (ptr getelementptr (float, ptr null, i32 1) to i64) + %265 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %66, 1 + %266 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %66, 2 + %267 = getelementptr float, ptr %265, i64 %266 + %268 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 1 + %269 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 2 + %270 = getelementptr float, ptr %268, i64 %269 + call void @llvm.memcpy.p0.p0.i64(ptr %270, ptr %267, i64 %264, i1 false) + %271 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 0 + %272 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 1 + %273 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 2 + %274 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 3, 0 + %275 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 3, 1 + %276 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 3, 2 + %277 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 3, 3 + %278 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 4, 0 + %279 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 4, 1 + %280 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 4, 2 + %281 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 4, 3 + %282 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 0 + %283 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 1 + %284 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 2 + %285 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 3, 0 + %286 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 3, 1 + %287 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 3, 2 + %288 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 3, 3 + %289 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 4, 0 + %290 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 4, 1 + %291 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 4, 2 + %292 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 4, 3 + %293 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 0 + %294 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 1 + %295 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 2 + %296 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 3, 0 + %297 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 3, 1 + %298 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 3, 2 + %299 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 3, 3 + %300 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 4, 0 + %301 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 4, 1 + %302 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 4, 2 + %303 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 4, 3 + %304 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 0 + %305 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 1 + %306 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 2 + %307 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 3, 0 + %308 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 3, 1 + %309 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 3, 2 + %310 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 3, 3 + %311 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 4, 0 + %312 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 4, 1 + %313 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 4, 2 + %314 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 4, 3 + %315 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 0 + %316 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 1 + %317 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 2 + %318 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 3, 0 + %319 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 3, 1 + %320 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 3, 2 + %321 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 3, 3 + %322 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 4, 0 + %323 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 4, 1 + %324 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 4, 2 + %325 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 4, 3 + %326 = call { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } @forward_outlined_part_0(ptr %271, ptr %272, i64 %273, i64 %274, i64 %275, i64 %276, i64 %277, i64 %278, i64 %279, i64 %280, i64 %281, ptr %282, ptr %283, i64 %284, i64 %285, i64 %286, i64 %287, i64 %288, i64 %289, i64 %290, i64 %291, i64 %292, ptr %293, ptr %294, i64 %295, i64 %296, i64 %297, i64 %298, i64 %299, i64 %300, i64 %301, i64 %302, i64 %303, ptr %304, ptr %305, i64 %306, i64 %307, i64 %308, i64 %309, i64 %310, i64 %311, i64 %312, i64 %313, i64 %314, ptr %315, ptr %316, i64 %317, i64 %318, i64 %319, i64 %320, i64 %321, i64 %322, i64 %323, i64 %324, i64 %325) + %327 = extractvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %326, 0 + %328 = extractvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %326, 1 + %329 = extractvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %326, 2 + %330 = extractvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %326, 3 + %331 = extractvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %326, 4 + %332 = extractvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %326, 5 + %333 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 1 + %334 = ptrtoint ptr %333 to i64 + %335 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %332, 1 + %336 = ptrtoint ptr %335 to i64 + %337 = icmp ne i64 %334, %336 + %338 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %327, 1 + %339 = ptrtoint ptr %338 to i64 + %340 = icmp ne i64 %334, %339 + %341 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %328, 1 + %342 = ptrtoint ptr %341 to i64 + %343 = icmp ne i64 %334, %342 + %344 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %329, 1 + %345 = ptrtoint ptr %344 to i64 + %346 = icmp ne i64 %334, %345 + %347 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %330, 1 + %348 = ptrtoint ptr %347 to i64 + %349 = icmp ne i64 %334, %348 + %350 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %331, 1 + %351 = ptrtoint ptr %350 to i64 + %352 = icmp ne i64 %334, %351 + %353 = and i1 %337, %340 + %354 = and i1 %353, %343 + %355 = and i1 %354, %346 + %356 = and i1 %355, %349 + %357 = and i1 %356, %352 + br i1 %357, label %358, label %360 + +358: ; preds = %55 + %359 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %127, 0 + call void @free(ptr %359) + br label %360 + +360: ; preds = %358, %55 + %361 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 1 + %362 = ptrtoint ptr %361 to i64 + %363 = icmp ne i64 %362, %336 + %364 = icmp ne i64 %362, %339 + %365 = icmp ne i64 %362, %342 + %366 = icmp ne i64 %362, %345 + %367 = icmp ne i64 %362, %348 + %368 = icmp ne i64 %362, %351 + %369 = and i1 %363, %364 + %370 = and i1 %369, %365 + %371 = and i1 %370, %366 + %372 = and i1 %371, %367 + %373 = and i1 %372, %368 + br i1 %373, label %374, label %376 + +374: ; preds = %360 + %375 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %159, 0 + call void @free(ptr %375) + br label %376 + +376: ; preds = %374, %360 + %377 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 1 + %378 = ptrtoint ptr %377 to i64 + %379 = icmp ne i64 %378, %336 + %380 = icmp ne i64 %378, %339 + %381 = icmp ne i64 %378, %342 + %382 = icmp ne i64 %378, %345 + %383 = icmp ne i64 %378, %348 + %384 = icmp ne i64 %378, %351 + %385 = and i1 %379, %380 + %386 = and i1 %385, %381 + %387 = and i1 %386, %382 + %388 = and i1 %387, %383 + %389 = and i1 %388, %384 + br i1 %389, label %390, label %392 + +390: ; preds = %376 + %391 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %191, 0 + call void @free(ptr %391) + br label %392 + +392: ; preds = %390, %376 + %393 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 1 + %394 = ptrtoint ptr %393 to i64 + %395 = icmp ne i64 %394, %336 + %396 = icmp ne i64 %394, %339 + %397 = icmp ne i64 %394, %342 + %398 = icmp ne i64 %394, %345 + %399 = icmp ne i64 %394, %348 + %400 = icmp ne i64 %394, %351 + %401 = and i1 %395, %396 + %402 = and i1 %401, %397 + %403 = and i1 %402, %398 + %404 = and i1 %403, %399 + %405 = and i1 %404, %400 + br i1 %405, label %406, label %408 + +406: ; preds = %392 + %407 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %223, 0 + call void @free(ptr %407) + br label %408 + +408: ; preds = %406, %392 + %409 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 1 + %410 = ptrtoint ptr %409 to i64 + %411 = icmp ne i64 %410, %336 + %412 = icmp ne i64 %410, %339 + %413 = icmp ne i64 %410, %342 + %414 = icmp ne i64 %410, %345 + %415 = icmp ne i64 %410, %348 + %416 = icmp ne i64 %410, %351 + %417 = and i1 %411, %412 + %418 = and i1 %417, %413 + %419 = and i1 %418, %414 + %420 = and i1 %419, %415 + %421 = and i1 %420, %416 + br i1 %421, label %422, label %424 + +422: ; preds = %408 + %423 = extractvalue { ptr, ptr, i64, [4 x i64], [4 x i64] } %255, 0 + call void @free(ptr %423) + br label %424 + +424: ; preds = %422, %408 + %425 = insertvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } undef, { ptr, ptr, i64, [4 x i64], [4 x i64] } %332, 0 + %426 = insertvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %425, { ptr, ptr, i64, [4 x i64], [4 x i64] } %327, 1 + %427 = insertvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %426, { ptr, ptr, i64, [4 x i64], [4 x i64] } %328, 2 + %428 = insertvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %427, { ptr, ptr, i64, [4 x i64], [4 x i64] } %329, 3 + %429 = insertvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %428, { ptr, ptr, i64, [4 x i64], [4 x i64] } %330, 4 + %430 = insertvalue { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %429, { ptr, ptr, i64, [4 x i64], [4 x i64] } %331, 5 + ret { { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] }, { ptr, ptr, i64, [4 x i64], [4 x i64] } } %430 +} + +; Function Attrs: nocallback nofree nounwind willreturn memory(argmem: readwrite) +declare void @llvm.memcpy.p0.p0.i64(ptr noalias nocapture writeonly, ptr noalias nocapture readonly, i64, i1 immarg) #0 + +attributes #0 = { nocallback nofree nounwind willreturn memory(argmem: readwrite) } + +!llvm.module.flags = !{!0} + +!0 = !{i32 2, !"Debug Info Version", i32 3} diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.llvm.mlir b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.llvm.mlir new file mode 100644 index 0000000000000000000000000000000000000000..daf3e8335e5962a3870ad344cec723a357425cb1 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.llvm.mlir @@ -0,0 +1,830 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + llvm.func @free(!llvm.ptr) loc(#loc) + llvm.func @malloc(i64) -> !llvm.ptr loc(#loc) + llvm.func @forward_outlined_part_0(!llvm.ptr, !llvm.ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, !llvm.ptr, !llvm.ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, !llvm.ptr, !llvm.ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, !llvm.ptr, !llvm.ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, !llvm.ptr, !llvm.ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64) -> !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> attributes { + aie_partition = 0 : i32, + kernel, + sym_visibility = "private"} loc(#loc308) + llvm.func @forward(%arg0: !llvm.ptr loc(unknown), %arg1: !llvm.ptr loc(unknown), %arg2: i64 loc(unknown), %arg3: i64 loc(unknown), %arg4: i64 loc(unknown), %arg5: i64 loc(unknown), %arg6: i64 loc(unknown), %arg7: i64 loc(unknown), %arg8: i64 loc(unknown), %arg9: i64 loc(unknown), %arg10: i64 loc(unknown), %arg11: !llvm.ptr loc(unknown), %arg12: !llvm.ptr loc(unknown), %arg13: i64 loc(unknown), %arg14: i64 loc(unknown), %arg15: i64 loc(unknown), %arg16: i64 loc(unknown), %arg17: i64 loc(unknown), %arg18: i64 loc(unknown), %arg19: i64 loc(unknown), %arg20: i64 loc(unknown), %arg21: i64 loc(unknown), %arg22: !llvm.ptr loc(unknown), %arg23: !llvm.ptr loc(unknown), %arg24: i64 loc(unknown), %arg25: i64 loc(unknown), %arg26: i64 loc(unknown), %arg27: i64 loc(unknown), %arg28: i64 loc(unknown), %arg29: i64 loc(unknown), %arg30: i64 loc(unknown), %arg31: i64 loc(unknown), %arg32: i64 loc(unknown), %arg33: !llvm.ptr loc(unknown), %arg34: !llvm.ptr loc(unknown), %arg35: i64 loc(unknown), %arg36: i64 loc(unknown), %arg37: i64 loc(unknown), %arg38: i64 loc(unknown), %arg39: i64 loc(unknown), %arg40: i64 loc(unknown), %arg41: i64 loc(unknown), %arg42: i64 loc(unknown), %arg43: i64 loc(unknown), %arg44: !llvm.ptr loc(unknown), %arg45: !llvm.ptr loc(unknown), %arg46: i64 loc(unknown), %arg47: i64 loc(unknown), %arg48: i64 loc(unknown), %arg49: i64 loc(unknown), %arg50: i64 loc(unknown), %arg51: i64 loc(unknown), %arg52: i64 loc(unknown), %arg53: i64 loc(unknown), %arg54: i64 loc(unknown)) -> !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> { + %0 = llvm.mlir.undef : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc) + %1 = llvm.mlir.undef : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %2 = llvm.insertvalue %arg44, %1[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %3 = llvm.insertvalue %arg45, %2[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %4 = llvm.insertvalue %arg46, %3[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %5 = llvm.insertvalue %arg47, %4[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %6 = llvm.insertvalue %arg51, %5[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %7 = llvm.insertvalue %arg48, %6[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %8 = llvm.insertvalue %arg52, %7[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %9 = llvm.insertvalue %arg49, %8[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %10 = llvm.insertvalue %arg53, %9[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %11 = llvm.insertvalue %arg50, %10[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %12 = llvm.insertvalue %arg54, %11[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %13 = llvm.insertvalue %arg33, %1[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %14 = llvm.insertvalue %arg34, %13[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %15 = llvm.insertvalue %arg35, %14[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %16 = llvm.insertvalue %arg36, %15[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %17 = llvm.insertvalue %arg40, %16[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %18 = llvm.insertvalue %arg37, %17[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %19 = llvm.insertvalue %arg41, %18[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %20 = llvm.insertvalue %arg38, %19[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %21 = llvm.insertvalue %arg42, %20[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %22 = llvm.insertvalue %arg39, %21[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %23 = llvm.insertvalue %arg43, %22[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %24 = llvm.insertvalue %arg22, %1[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %25 = llvm.insertvalue %arg23, %24[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %26 = llvm.insertvalue %arg24, %25[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %27 = llvm.insertvalue %arg25, %26[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %28 = llvm.insertvalue %arg29, %27[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %29 = llvm.insertvalue %arg26, %28[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %30 = llvm.insertvalue %arg30, %29[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %31 = llvm.insertvalue %arg27, %30[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %32 = llvm.insertvalue %arg31, %31[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %33 = llvm.insertvalue %arg28, %32[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %34 = llvm.insertvalue %arg32, %33[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %35 = llvm.insertvalue %arg11, %1[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %36 = llvm.insertvalue %arg12, %35[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %37 = llvm.insertvalue %arg13, %36[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %38 = llvm.insertvalue %arg14, %37[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %39 = llvm.insertvalue %arg18, %38[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %40 = llvm.insertvalue %arg15, %39[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %41 = llvm.insertvalue %arg19, %40[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %42 = llvm.insertvalue %arg16, %41[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %43 = llvm.insertvalue %arg20, %42[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %44 = llvm.insertvalue %arg17, %43[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %45 = llvm.insertvalue %arg21, %44[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %46 = llvm.insertvalue %arg0, %1[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %47 = llvm.insertvalue %arg1, %46[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %48 = llvm.insertvalue %arg2, %47[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %49 = llvm.insertvalue %arg3, %48[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %50 = llvm.insertvalue %arg7, %49[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %51 = llvm.insertvalue %arg4, %50[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %52 = llvm.insertvalue %arg8, %51[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %53 = llvm.insertvalue %arg5, %52[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %54 = llvm.insertvalue %arg9, %53[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %55 = llvm.insertvalue %arg6, %54[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %56 = llvm.insertvalue %arg10, %55[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %57 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %58 = llvm.mlir.constant(180 : index) : i64 loc(#loc308) + %59 = llvm.mlir.constant(320 : index) : i64 loc(#loc308) + %60 = llvm.mlir.constant(4 : index) : i64 loc(#loc308) + %61 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %62 = llvm.mlir.constant(1280 : index) : i64 loc(#loc308) + %63 = llvm.mlir.constant(230400 : index) : i64 loc(#loc308) + %64 = llvm.mlir.constant(230400 : index) : i64 loc(#loc308) + %65 = llvm.mlir.zero : !llvm.ptr loc(#loc308) + %66 = llvm.getelementptr %65[%64] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + %67 = llvm.ptrtoint %66 : !llvm.ptr to i64 loc(#loc308) + %68 = llvm.mlir.constant(64 : index) : i64 loc(#loc308) + %69 = llvm.add %67, %68 : i64 loc(#loc308) + %70 = llvm.call @malloc(%69) : (i64) -> !llvm.ptr loc(#loc308) + %71 = llvm.ptrtoint %70 : !llvm.ptr to i64 loc(#loc308) + %72 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %73 = llvm.sub %68, %72 : i64 loc(#loc308) + %74 = llvm.add %71, %73 : i64 loc(#loc308) + %75 = llvm.urem %74, %68 : i64 loc(#loc308) + %76 = llvm.sub %74, %75 : i64 loc(#loc308) + %77 = llvm.inttoptr %76 : i64 to !llvm.ptr loc(#loc308) + %78 = llvm.mlir.undef : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %79 = llvm.insertvalue %70, %78[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %80 = llvm.insertvalue %77, %79[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %81 = llvm.mlir.constant(0 : index) : i64 loc(#loc308) + %82 = llvm.insertvalue %81, %80[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %83 = llvm.insertvalue %57, %82[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %84 = llvm.insertvalue %58, %83[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %85 = llvm.insertvalue %59, %84[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %86 = llvm.insertvalue %60, %85[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %87 = llvm.insertvalue %63, %86[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %88 = llvm.insertvalue %62, %87[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %89 = llvm.insertvalue %60, %88[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %90 = llvm.insertvalue %61, %89[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %91 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %92 = llvm.extractvalue %56[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %93 = llvm.mul %91, %92 : i64 loc(#loc308) + %94 = llvm.extractvalue %56[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %95 = llvm.mul %93, %94 : i64 loc(#loc308) + %96 = llvm.extractvalue %56[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %97 = llvm.mul %95, %96 : i64 loc(#loc308) + %98 = llvm.extractvalue %56[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %99 = llvm.mul %97, %98 : i64 loc(#loc308) + %100 = llvm.mlir.zero : !llvm.ptr loc(#loc308) + %101 = llvm.getelementptr %100[1] : (!llvm.ptr) -> !llvm.ptr, f32 loc(#loc308) + %102 = llvm.ptrtoint %101 : !llvm.ptr to i64 loc(#loc308) + %103 = llvm.mul %99, %102 : i64 loc(#loc308) + %104 = llvm.extractvalue %56[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %105 = llvm.extractvalue %56[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %106 = llvm.getelementptr %104[%105] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + %107 = llvm.extractvalue %90[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %108 = llvm.extractvalue %90[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %109 = llvm.getelementptr %107[%108] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + "llvm.intr.memcpy"(%109, %106, %103) <{isVolatile = false}> : (!llvm.ptr, !llvm.ptr, i64) -> () loc(#loc308) + %110 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %111 = llvm.mlir.constant(16 : index) : i64 loc(#loc308) + %112 = llvm.mlir.constant(90 : index) : i64 loc(#loc308) + %113 = llvm.mlir.constant(160 : index) : i64 loc(#loc308) + %114 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %115 = llvm.mlir.constant(14400 : index) : i64 loc(#loc308) + %116 = llvm.mlir.constant(230400 : index) : i64 loc(#loc308) + %117 = llvm.mlir.constant(230400 : index) : i64 loc(#loc308) + %118 = llvm.mlir.zero : !llvm.ptr loc(#loc308) + %119 = llvm.getelementptr %118[%117] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + %120 = llvm.ptrtoint %119 : !llvm.ptr to i64 loc(#loc308) + %121 = llvm.mlir.constant(64 : index) : i64 loc(#loc308) + %122 = llvm.add %120, %121 : i64 loc(#loc308) + %123 = llvm.call @malloc(%122) : (i64) -> !llvm.ptr loc(#loc308) + %124 = llvm.ptrtoint %123 : !llvm.ptr to i64 loc(#loc308) + %125 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %126 = llvm.sub %121, %125 : i64 loc(#loc308) + %127 = llvm.add %124, %126 : i64 loc(#loc308) + %128 = llvm.urem %127, %121 : i64 loc(#loc308) + %129 = llvm.sub %127, %128 : i64 loc(#loc308) + %130 = llvm.inttoptr %129 : i64 to !llvm.ptr loc(#loc308) + %131 = llvm.mlir.undef : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %132 = llvm.insertvalue %123, %131[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %133 = llvm.insertvalue %130, %132[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %134 = llvm.mlir.constant(0 : index) : i64 loc(#loc308) + %135 = llvm.insertvalue %134, %133[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %136 = llvm.insertvalue %110, %135[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %137 = llvm.insertvalue %111, %136[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %138 = llvm.insertvalue %112, %137[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %139 = llvm.insertvalue %113, %138[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %140 = llvm.insertvalue %116, %139[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %141 = llvm.insertvalue %115, %140[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %142 = llvm.insertvalue %113, %141[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %143 = llvm.insertvalue %114, %142[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %144 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %145 = llvm.extractvalue %45[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %146 = llvm.mul %144, %145 : i64 loc(#loc308) + %147 = llvm.extractvalue %45[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %148 = llvm.mul %146, %147 : i64 loc(#loc308) + %149 = llvm.extractvalue %45[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %150 = llvm.mul %148, %149 : i64 loc(#loc308) + %151 = llvm.extractvalue %45[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %152 = llvm.mul %150, %151 : i64 loc(#loc308) + %153 = llvm.mlir.zero : !llvm.ptr loc(#loc308) + %154 = llvm.getelementptr %153[1] : (!llvm.ptr) -> !llvm.ptr, f32 loc(#loc308) + %155 = llvm.ptrtoint %154 : !llvm.ptr to i64 loc(#loc308) + %156 = llvm.mul %152, %155 : i64 loc(#loc308) + %157 = llvm.extractvalue %45[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %158 = llvm.extractvalue %45[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %159 = llvm.getelementptr %157[%158] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + %160 = llvm.extractvalue %143[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %161 = llvm.extractvalue %143[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %162 = llvm.getelementptr %160[%161] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + "llvm.intr.memcpy"(%162, %159, %156) <{isVolatile = false}> : (!llvm.ptr, !llvm.ptr, i64) -> () loc(#loc308) + %163 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %164 = llvm.mlir.constant(20 : index) : i64 loc(#loc308) + %165 = llvm.mlir.constant(45 : index) : i64 loc(#loc308) + %166 = llvm.mlir.constant(80 : index) : i64 loc(#loc308) + %167 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %168 = llvm.mlir.constant(3600 : index) : i64 loc(#loc308) + %169 = llvm.mlir.constant(72000 : index) : i64 loc(#loc308) + %170 = llvm.mlir.constant(72000 : index) : i64 loc(#loc308) + %171 = llvm.mlir.zero : !llvm.ptr loc(#loc308) + %172 = llvm.getelementptr %171[%170] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + %173 = llvm.ptrtoint %172 : !llvm.ptr to i64 loc(#loc308) + %174 = llvm.mlir.constant(64 : index) : i64 loc(#loc308) + %175 = llvm.add %173, %174 : i64 loc(#loc308) + %176 = llvm.call @malloc(%175) : (i64) -> !llvm.ptr loc(#loc308) + %177 = llvm.ptrtoint %176 : !llvm.ptr to i64 loc(#loc308) + %178 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %179 = llvm.sub %174, %178 : i64 loc(#loc308) + %180 = llvm.add %177, %179 : i64 loc(#loc308) + %181 = llvm.urem %180, %174 : i64 loc(#loc308) + %182 = llvm.sub %180, %181 : i64 loc(#loc308) + %183 = llvm.inttoptr %182 : i64 to !llvm.ptr loc(#loc308) + %184 = llvm.mlir.undef : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %185 = llvm.insertvalue %176, %184[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %186 = llvm.insertvalue %183, %185[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %187 = llvm.mlir.constant(0 : index) : i64 loc(#loc308) + %188 = llvm.insertvalue %187, %186[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %189 = llvm.insertvalue %163, %188[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %190 = llvm.insertvalue %164, %189[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %191 = llvm.insertvalue %165, %190[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %192 = llvm.insertvalue %166, %191[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %193 = llvm.insertvalue %169, %192[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %194 = llvm.insertvalue %168, %193[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %195 = llvm.insertvalue %166, %194[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %196 = llvm.insertvalue %167, %195[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %197 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %198 = llvm.extractvalue %34[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %199 = llvm.mul %197, %198 : i64 loc(#loc308) + %200 = llvm.extractvalue %34[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %201 = llvm.mul %199, %200 : i64 loc(#loc308) + %202 = llvm.extractvalue %34[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %203 = llvm.mul %201, %202 : i64 loc(#loc308) + %204 = llvm.extractvalue %34[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %205 = llvm.mul %203, %204 : i64 loc(#loc308) + %206 = llvm.mlir.zero : !llvm.ptr loc(#loc308) + %207 = llvm.getelementptr %206[1] : (!llvm.ptr) -> !llvm.ptr, f32 loc(#loc308) + %208 = llvm.ptrtoint %207 : !llvm.ptr to i64 loc(#loc308) + %209 = llvm.mul %205, %208 : i64 loc(#loc308) + %210 = llvm.extractvalue %34[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %211 = llvm.extractvalue %34[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %212 = llvm.getelementptr %210[%211] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + %213 = llvm.extractvalue %196[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %214 = llvm.extractvalue %196[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %215 = llvm.getelementptr %213[%214] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + "llvm.intr.memcpy"(%215, %212, %209) <{isVolatile = false}> : (!llvm.ptr, !llvm.ptr, i64) -> () loc(#loc308) + %216 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %217 = llvm.mlir.constant(40 : index) : i64 loc(#loc308) + %218 = llvm.mlir.constant(23 : index) : i64 loc(#loc308) + %219 = llvm.mlir.constant(40 : index) : i64 loc(#loc308) + %220 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %221 = llvm.mlir.constant(920 : index) : i64 loc(#loc308) + %222 = llvm.mlir.constant(36800 : index) : i64 loc(#loc308) + %223 = llvm.mlir.constant(36800 : index) : i64 loc(#loc308) + %224 = llvm.mlir.zero : !llvm.ptr loc(#loc308) + %225 = llvm.getelementptr %224[%223] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + %226 = llvm.ptrtoint %225 : !llvm.ptr to i64 loc(#loc308) + %227 = llvm.mlir.constant(64 : index) : i64 loc(#loc308) + %228 = llvm.add %226, %227 : i64 loc(#loc308) + %229 = llvm.call @malloc(%228) : (i64) -> !llvm.ptr loc(#loc308) + %230 = llvm.ptrtoint %229 : !llvm.ptr to i64 loc(#loc308) + %231 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %232 = llvm.sub %227, %231 : i64 loc(#loc308) + %233 = llvm.add %230, %232 : i64 loc(#loc308) + %234 = llvm.urem %233, %227 : i64 loc(#loc308) + %235 = llvm.sub %233, %234 : i64 loc(#loc308) + %236 = llvm.inttoptr %235 : i64 to !llvm.ptr loc(#loc308) + %237 = llvm.mlir.undef : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %238 = llvm.insertvalue %229, %237[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %239 = llvm.insertvalue %236, %238[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %240 = llvm.mlir.constant(0 : index) : i64 loc(#loc308) + %241 = llvm.insertvalue %240, %239[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %242 = llvm.insertvalue %216, %241[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %243 = llvm.insertvalue %217, %242[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %244 = llvm.insertvalue %218, %243[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %245 = llvm.insertvalue %219, %244[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %246 = llvm.insertvalue %222, %245[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %247 = llvm.insertvalue %221, %246[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %248 = llvm.insertvalue %219, %247[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %249 = llvm.insertvalue %220, %248[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %250 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %251 = llvm.extractvalue %23[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %252 = llvm.mul %250, %251 : i64 loc(#loc308) + %253 = llvm.extractvalue %23[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %254 = llvm.mul %252, %253 : i64 loc(#loc308) + %255 = llvm.extractvalue %23[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %256 = llvm.mul %254, %255 : i64 loc(#loc308) + %257 = llvm.extractvalue %23[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %258 = llvm.mul %256, %257 : i64 loc(#loc308) + %259 = llvm.mlir.zero : !llvm.ptr loc(#loc308) + %260 = llvm.getelementptr %259[1] : (!llvm.ptr) -> !llvm.ptr, f32 loc(#loc308) + %261 = llvm.ptrtoint %260 : !llvm.ptr to i64 loc(#loc308) + %262 = llvm.mul %258, %261 : i64 loc(#loc308) + %263 = llvm.extractvalue %23[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %264 = llvm.extractvalue %23[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %265 = llvm.getelementptr %263[%264] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + %266 = llvm.extractvalue %249[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %267 = llvm.extractvalue %249[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %268 = llvm.getelementptr %266[%267] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + "llvm.intr.memcpy"(%268, %265, %262) <{isVolatile = false}> : (!llvm.ptr, !llvm.ptr, i64) -> () loc(#loc308) + %269 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %270 = llvm.mlir.constant(64 : index) : i64 loc(#loc308) + %271 = llvm.mlir.constant(12 : index) : i64 loc(#loc308) + %272 = llvm.mlir.constant(20 : index) : i64 loc(#loc308) + %273 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %274 = llvm.mlir.constant(240 : index) : i64 loc(#loc308) + %275 = llvm.mlir.constant(15360 : index) : i64 loc(#loc308) + %276 = llvm.mlir.constant(15360 : index) : i64 loc(#loc308) + %277 = llvm.mlir.zero : !llvm.ptr loc(#loc308) + %278 = llvm.getelementptr %277[%276] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + %279 = llvm.ptrtoint %278 : !llvm.ptr to i64 loc(#loc308) + %280 = llvm.mlir.constant(64 : index) : i64 loc(#loc308) + %281 = llvm.add %279, %280 : i64 loc(#loc308) + %282 = llvm.call @malloc(%281) : (i64) -> !llvm.ptr loc(#loc308) + %283 = llvm.ptrtoint %282 : !llvm.ptr to i64 loc(#loc308) + %284 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %285 = llvm.sub %280, %284 : i64 loc(#loc308) + %286 = llvm.add %283, %285 : i64 loc(#loc308) + %287 = llvm.urem %286, %280 : i64 loc(#loc308) + %288 = llvm.sub %286, %287 : i64 loc(#loc308) + %289 = llvm.inttoptr %288 : i64 to !llvm.ptr loc(#loc308) + %290 = llvm.mlir.undef : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %291 = llvm.insertvalue %282, %290[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %292 = llvm.insertvalue %289, %291[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %293 = llvm.mlir.constant(0 : index) : i64 loc(#loc308) + %294 = llvm.insertvalue %293, %292[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %295 = llvm.insertvalue %269, %294[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %296 = llvm.insertvalue %270, %295[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %297 = llvm.insertvalue %271, %296[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %298 = llvm.insertvalue %272, %297[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %299 = llvm.insertvalue %275, %298[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %300 = llvm.insertvalue %274, %299[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %301 = llvm.insertvalue %272, %300[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %302 = llvm.insertvalue %273, %301[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %303 = llvm.mlir.constant(1 : index) : i64 loc(#loc308) + %304 = llvm.extractvalue %12[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %305 = llvm.mul %303, %304 : i64 loc(#loc308) + %306 = llvm.extractvalue %12[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %307 = llvm.mul %305, %306 : i64 loc(#loc308) + %308 = llvm.extractvalue %12[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %309 = llvm.mul %307, %308 : i64 loc(#loc308) + %310 = llvm.extractvalue %12[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %311 = llvm.mul %309, %310 : i64 loc(#loc308) + %312 = llvm.mlir.zero : !llvm.ptr loc(#loc308) + %313 = llvm.getelementptr %312[1] : (!llvm.ptr) -> !llvm.ptr, f32 loc(#loc308) + %314 = llvm.ptrtoint %313 : !llvm.ptr to i64 loc(#loc308) + %315 = llvm.mul %311, %314 : i64 loc(#loc308) + %316 = llvm.extractvalue %12[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %317 = llvm.extractvalue %12[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %318 = llvm.getelementptr %316[%317] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + %319 = llvm.extractvalue %302[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %320 = llvm.extractvalue %302[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %321 = llvm.getelementptr %319[%320] : (!llvm.ptr, i64) -> !llvm.ptr, f32 loc(#loc308) + "llvm.intr.memcpy"(%321, %318, %315) <{isVolatile = false}> : (!llvm.ptr, !llvm.ptr, i64) -> () loc(#loc308) + %322 = llvm.extractvalue %90[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %323 = llvm.extractvalue %90[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %324 = llvm.extractvalue %90[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %325 = llvm.extractvalue %90[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %326 = llvm.extractvalue %90[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %327 = llvm.extractvalue %90[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %328 = llvm.extractvalue %90[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %329 = llvm.extractvalue %90[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %330 = llvm.extractvalue %90[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %331 = llvm.extractvalue %90[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %332 = llvm.extractvalue %90[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %333 = llvm.extractvalue %143[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %334 = llvm.extractvalue %143[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %335 = llvm.extractvalue %143[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %336 = llvm.extractvalue %143[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %337 = llvm.extractvalue %143[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %338 = llvm.extractvalue %143[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %339 = llvm.extractvalue %143[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %340 = llvm.extractvalue %143[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %341 = llvm.extractvalue %143[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %342 = llvm.extractvalue %143[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %343 = llvm.extractvalue %143[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %344 = llvm.extractvalue %196[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %345 = llvm.extractvalue %196[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %346 = llvm.extractvalue %196[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %347 = llvm.extractvalue %196[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %348 = llvm.extractvalue %196[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %349 = llvm.extractvalue %196[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %350 = llvm.extractvalue %196[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %351 = llvm.extractvalue %196[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %352 = llvm.extractvalue %196[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %353 = llvm.extractvalue %196[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %354 = llvm.extractvalue %196[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %355 = llvm.extractvalue %249[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %356 = llvm.extractvalue %249[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %357 = llvm.extractvalue %249[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %358 = llvm.extractvalue %249[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %359 = llvm.extractvalue %249[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %360 = llvm.extractvalue %249[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %361 = llvm.extractvalue %249[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %362 = llvm.extractvalue %249[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %363 = llvm.extractvalue %249[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %364 = llvm.extractvalue %249[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %365 = llvm.extractvalue %249[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %366 = llvm.extractvalue %302[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %367 = llvm.extractvalue %302[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %368 = llvm.extractvalue %302[2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %369 = llvm.extractvalue %302[3, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %370 = llvm.extractvalue %302[3, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %371 = llvm.extractvalue %302[3, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %372 = llvm.extractvalue %302[3, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %373 = llvm.extractvalue %302[4, 0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %374 = llvm.extractvalue %302[4, 1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %375 = llvm.extractvalue %302[4, 2] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %376 = llvm.extractvalue %302[4, 3] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc308) + %377 = llvm.call @forward_outlined_part_0(%322, %323, %324, %325, %326, %327, %328, %329, %330, %331, %332, %333, %334, %335, %336, %337, %338, %339, %340, %341, %342, %343, %344, %345, %346, %347, %348, %349, %350, %351, %352, %353, %354, %355, %356, %357, %358, %359, %360, %361, %362, %363, %364, %365, %366, %367, %368, %369, %370, %371, %372, %373, %374, %375, %376) : (!llvm.ptr, !llvm.ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, !llvm.ptr, !llvm.ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, !llvm.ptr, !llvm.ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, !llvm.ptr, !llvm.ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64, !llvm.ptr, !llvm.ptr, i64, i64, i64, i64, i64, i64, i64, i64, i64) -> !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc308) + %378 = llvm.extractvalue %377[0] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc308) + %379 = llvm.extractvalue %377[1] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc308) + %380 = llvm.extractvalue %377[2] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc308) + %381 = llvm.extractvalue %377[3] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc308) + %382 = llvm.extractvalue %377[4] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc308) + %383 = llvm.extractvalue %377[5] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc308) + %384 = llvm.extractvalue %90[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %385 = llvm.ptrtoint %384 : !llvm.ptr to i64 loc(#loc) + %386 = llvm.extractvalue %383[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %387 = llvm.ptrtoint %386 : !llvm.ptr to i64 loc(#loc) + %388 = llvm.icmp "ne" %385, %387 : i64 loc(#loc) + %389 = llvm.extractvalue %378[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %390 = llvm.ptrtoint %389 : !llvm.ptr to i64 loc(#loc) + %391 = llvm.icmp "ne" %385, %390 : i64 loc(#loc) + %392 = llvm.extractvalue %379[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %393 = llvm.ptrtoint %392 : !llvm.ptr to i64 loc(#loc) + %394 = llvm.icmp "ne" %385, %393 : i64 loc(#loc) + %395 = llvm.extractvalue %380[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %396 = llvm.ptrtoint %395 : !llvm.ptr to i64 loc(#loc) + %397 = llvm.icmp "ne" %385, %396 : i64 loc(#loc) + %398 = llvm.extractvalue %381[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %399 = llvm.ptrtoint %398 : !llvm.ptr to i64 loc(#loc) + %400 = llvm.icmp "ne" %385, %399 : i64 loc(#loc) + %401 = llvm.extractvalue %382[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %402 = llvm.ptrtoint %401 : !llvm.ptr to i64 loc(#loc) + %403 = llvm.icmp "ne" %385, %402 : i64 loc(#loc) + %404 = llvm.and %388, %391 : i1 loc(#loc) + %405 = llvm.and %404, %394 : i1 loc(#loc) + %406 = llvm.and %405, %397 : i1 loc(#loc) + %407 = llvm.and %406, %400 : i1 loc(#loc) + %408 = llvm.and %407, %403 : i1 loc(#loc) + llvm.cond_br %408, ^bb1, ^bb2 loc(#loc) + ^bb1: // pred: ^bb0 + %409 = llvm.extractvalue %90[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + llvm.call @free(%409) : (!llvm.ptr) -> () loc(#loc) + llvm.br ^bb2 loc(#loc) + ^bb2: // 2 preds: ^bb0, ^bb1 + %410 = llvm.extractvalue %143[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %411 = llvm.ptrtoint %410 : !llvm.ptr to i64 loc(#loc) + %412 = llvm.icmp "ne" %411, %387 : i64 loc(#loc) + %413 = llvm.icmp "ne" %411, %390 : i64 loc(#loc) + %414 = llvm.icmp "ne" %411, %393 : i64 loc(#loc) + %415 = llvm.icmp "ne" %411, %396 : i64 loc(#loc) + %416 = llvm.icmp "ne" %411, %399 : i64 loc(#loc) + %417 = llvm.icmp "ne" %411, %402 : i64 loc(#loc) + %418 = llvm.and %412, %413 : i1 loc(#loc) + %419 = llvm.and %418, %414 : i1 loc(#loc) + %420 = llvm.and %419, %415 : i1 loc(#loc) + %421 = llvm.and %420, %416 : i1 loc(#loc) + %422 = llvm.and %421, %417 : i1 loc(#loc) + llvm.cond_br %422, ^bb3, ^bb4 loc(#loc) + ^bb3: // pred: ^bb2 + %423 = llvm.extractvalue %143[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + llvm.call @free(%423) : (!llvm.ptr) -> () loc(#loc) + llvm.br ^bb4 loc(#loc) + ^bb4: // 2 preds: ^bb2, ^bb3 + %424 = llvm.extractvalue %196[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %425 = llvm.ptrtoint %424 : !llvm.ptr to i64 loc(#loc) + %426 = llvm.icmp "ne" %425, %387 : i64 loc(#loc) + %427 = llvm.icmp "ne" %425, %390 : i64 loc(#loc) + %428 = llvm.icmp "ne" %425, %393 : i64 loc(#loc) + %429 = llvm.icmp "ne" %425, %396 : i64 loc(#loc) + %430 = llvm.icmp "ne" %425, %399 : i64 loc(#loc) + %431 = llvm.icmp "ne" %425, %402 : i64 loc(#loc) + %432 = llvm.and %426, %427 : i1 loc(#loc) + %433 = llvm.and %432, %428 : i1 loc(#loc) + %434 = llvm.and %433, %429 : i1 loc(#loc) + %435 = llvm.and %434, %430 : i1 loc(#loc) + %436 = llvm.and %435, %431 : i1 loc(#loc) + llvm.cond_br %436, ^bb5, ^bb6 loc(#loc) + ^bb5: // pred: ^bb4 + %437 = llvm.extractvalue %196[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + llvm.call @free(%437) : (!llvm.ptr) -> () loc(#loc) + llvm.br ^bb6 loc(#loc) + ^bb6: // 2 preds: ^bb4, ^bb5 + %438 = llvm.extractvalue %249[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %439 = llvm.ptrtoint %438 : !llvm.ptr to i64 loc(#loc) + %440 = llvm.icmp "ne" %439, %387 : i64 loc(#loc) + %441 = llvm.icmp "ne" %439, %390 : i64 loc(#loc) + %442 = llvm.icmp "ne" %439, %393 : i64 loc(#loc) + %443 = llvm.icmp "ne" %439, %396 : i64 loc(#loc) + %444 = llvm.icmp "ne" %439, %399 : i64 loc(#loc) + %445 = llvm.icmp "ne" %439, %402 : i64 loc(#loc) + %446 = llvm.and %440, %441 : i1 loc(#loc) + %447 = llvm.and %446, %442 : i1 loc(#loc) + %448 = llvm.and %447, %443 : i1 loc(#loc) + %449 = llvm.and %448, %444 : i1 loc(#loc) + %450 = llvm.and %449, %445 : i1 loc(#loc) + llvm.cond_br %450, ^bb7, ^bb8 loc(#loc) + ^bb7: // pred: ^bb6 + %451 = llvm.extractvalue %249[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + llvm.call @free(%451) : (!llvm.ptr) -> () loc(#loc) + llvm.br ^bb8 loc(#loc) + ^bb8: // 2 preds: ^bb6, ^bb7 + %452 = llvm.extractvalue %302[1] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + %453 = llvm.ptrtoint %452 : !llvm.ptr to i64 loc(#loc) + %454 = llvm.icmp "ne" %453, %387 : i64 loc(#loc) + %455 = llvm.icmp "ne" %453, %390 : i64 loc(#loc) + %456 = llvm.icmp "ne" %453, %393 : i64 loc(#loc) + %457 = llvm.icmp "ne" %453, %396 : i64 loc(#loc) + %458 = llvm.icmp "ne" %453, %399 : i64 loc(#loc) + %459 = llvm.icmp "ne" %453, %402 : i64 loc(#loc) + %460 = llvm.and %454, %455 : i1 loc(#loc) + %461 = llvm.and %460, %456 : i1 loc(#loc) + %462 = llvm.and %461, %457 : i1 loc(#loc) + %463 = llvm.and %462, %458 : i1 loc(#loc) + %464 = llvm.and %463, %459 : i1 loc(#loc) + llvm.cond_br %464, ^bb9, ^bb10 loc(#loc) + ^bb9: // pred: ^bb8 + %465 = llvm.extractvalue %302[0] : !llvm.struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)> loc(#loc) + llvm.call @free(%465) : (!llvm.ptr) -> () loc(#loc) + llvm.br ^bb10 loc(#loc) + ^bb10: // 2 preds: ^bb8, ^bb9 + %466 = llvm.insertvalue %383, %0[0] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc) + %467 = llvm.insertvalue %378, %466[1] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc) + %468 = llvm.insertvalue %379, %467[2] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc) + %469 = llvm.insertvalue %380, %468[3] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc) + %470 = llvm.insertvalue %381, %469[4] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc) + %471 = llvm.insertvalue %382, %470[5] : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc) + llvm.return %471 : !llvm.struct<(struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>, struct<(ptr, ptr, i64, array<4 x i64>, array<4 x i64>)>)> loc(#loc) + } loc(#loc) +} loc(#loc) +#loc1 = loc("Div_2") +#loc2 = loc("Sub_431") +#loc3 = loc("Sub_411") +#loc4 = loc("Sub_385") +#loc5 = loc("Sub_359") +#loc6 = loc("Div_16") +#loc7 = loc("Sub_14") +#loc8 = loc("Initializer_398") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("Conv_17") +#loc14 = loc("Add_19") +#loc15 = loc("Clip_22") +#loc16 = loc("Div_24") +#loc17 = loc("Mul_25") +#loc18 = loc("Conv_26") +#loc19 = loc("Relu_27") +#loc20 = loc("Conv_28") +#loc21 = loc("Add_29") +#loc22 = loc("Conv_30") +#loc23 = loc("Relu_31") +#loc24 = loc("Conv_32") +#loc25 = loc("Relu_33") +#loc26 = loc("Conv_34") +#loc27 = loc("Conv_35") +#loc28 = loc("Relu_36") +#loc29 = loc("Conv_37") +#loc30 = loc("Relu_38") +#loc31 = loc("Conv_39") +#loc32 = loc("Add_40") +#loc33 = loc("Conv_41") +#loc34 = loc("Relu_42") +#loc35 = loc("Conv_43") +#loc36 = loc("Relu_44") +#loc37 = loc("GlobalAveragePool_45") +#loc38 = loc("Conv_46") +#loc39 = loc("Relu_47") +#loc40 = loc("Conv_48") +#loc41 = loc("Add_50") +#loc42 = loc("Clip_53") +#loc43 = loc("Div_55") +#loc44 = loc("Mul_56") +#loc45 = loc("Conv_57") +#loc46 = loc("Conv_58") +#loc47 = loc("Relu_59") +#loc48 = loc("Conv_60") +#loc49 = loc("Relu_61") +#loc50 = loc("GlobalAveragePool_62") +#loc51 = loc("Conv_63") +#loc52 = loc("Relu_64") +#loc53 = loc("Conv_65") +#loc54 = loc("Add_67") +#loc55 = loc("Clip_70") +#loc56 = loc("Div_72") +#loc57 = loc("Mul_73") +#loc58 = loc("Conv_74") +#loc59 = loc("Add_75") +#loc60 = loc("Conv_76") +#loc61 = loc("Relu_77") +#loc62 = loc("Conv_78") +#loc63 = loc("Relu_79") +#loc64 = loc("GlobalAveragePool_80") +#loc65 = loc("Conv_81") +#loc66 = loc("Relu_82") +#loc67 = loc("Conv_83") +#loc68 = loc("Add_85") +#loc69 = loc("Clip_88") +#loc70 = loc("Div_90") +#loc71 = loc("Mul_91") +#loc72 = loc("Conv_92") +#loc73 = loc("Add_93") +#loc74 = loc("Conv_94") +#loc75 = loc("Add_96") +#loc76 = loc("Clip_99") +#loc77 = loc("Div_101") +#loc78 = loc("Mul_102") +#loc79 = loc("Conv_103") +#loc80 = loc("Add_105") +#loc81 = loc("Clip_108") +#loc82 = loc("Div_110") +#loc83 = loc("Mul_111") +#loc84 = loc("Conv_112") +#loc85 = loc("Conv_113") +#loc86 = loc("Add_115") +#loc87 = loc("Clip_118") +#loc88 = loc("Div_120") +#loc89 = loc("Mul_121") +#loc90 = loc("Conv_122") +#loc91 = loc("Add_124") +#loc92 = loc("Clip_127") +#loc93 = loc("Div_129") +#loc94 = loc("Mul_130") +#loc95 = loc("Conv_131") +#loc96 = loc("Add_132") +#loc97 = loc("Conv_133") +#loc98 = loc("Add_135") +#loc99 = loc("Clip_138") +#loc100 = loc("Div_140") +#loc101 = loc("Mul_141") +#loc102 = loc("Conv_142") +#loc103 = loc("Add_144") +#loc104 = loc("Clip_147") +#loc105 = loc("Div_149") +#loc106 = loc("Mul_150") +#loc107 = loc("Conv_151") +#loc108 = loc("Add_152") +#loc109 = loc("Conv_153") +#loc110 = loc("Add_155") +#loc111 = loc("Clip_158") +#loc112 = loc("Div_160") +#loc113 = loc("Mul_161") +#loc114 = loc("Conv_162") +#loc115 = loc("Add_164") +#loc116 = loc("Clip_167") +#loc117 = loc("Div_169") +#loc118 = loc("Mul_170") +#loc119 = loc("Conv_171") +#loc120 = loc("Add_172") +#loc121 = loc("Conv_173") +#loc122 = loc("Add_175") +#loc123 = loc("Clip_178") +#loc124 = loc("Div_180") +#loc125 = loc("Mul_181") +#loc126 = loc("Conv_182") +#loc127 = loc("Add_184") +#loc128 = loc("Clip_187") +#loc129 = loc("Div_189") +#loc130 = loc("Mul_190") +#loc131 = loc("GlobalAveragePool_191") +#loc132 = loc("Conv_192") +#loc133 = loc("Relu_193") +#loc134 = loc("Conv_194") +#loc135 = loc("Add_196") +#loc136 = loc("Clip_199") +#loc137 = loc("Div_201") +#loc138 = loc("Mul_202") +#loc139 = loc("Conv_203") +#loc140 = loc("Conv_204") +#loc141 = loc("Add_206") +#loc142 = loc("Clip_209") +#loc143 = loc("Div_211") +#loc144 = loc("Mul_212") +#loc145 = loc("Conv_213") +#loc146 = loc("Add_215") +#loc147 = loc("Clip_218") +#loc148 = loc("Div_220") +#loc149 = loc("Mul_221") +#loc150 = loc("GlobalAveragePool_222") +#loc151 = loc("Conv_223") +#loc152 = loc("Relu_224") +#loc153 = loc("Conv_225") +#loc154 = loc("Add_227") +#loc155 = loc("Clip_230") +#loc156 = loc("Div_232") +#loc157 = loc("Mul_233") +#loc158 = loc("Conv_234") +#loc159 = loc("Add_235") +#loc160 = loc("Conv_236") +#loc161 = loc("Add_238") +#loc162 = loc("Clip_241") +#loc163 = loc("Div_243") +#loc164 = loc("Mul_244") +#loc165 = loc("Conv_245") +#loc166 = loc("Add_247") +#loc167 = loc("Clip_250") +#loc168 = loc("Div_252") +#loc169 = loc("Mul_253") +#loc170 = loc("GlobalAveragePool_254") +#loc171 = loc("Conv_255") +#loc172 = loc("Relu_256") +#loc173 = loc("Conv_257") +#loc174 = loc("Add_259") +#loc175 = loc("Clip_262") +#loc176 = loc("Div_264") +#loc177 = loc("Mul_265") +#loc178 = loc("Conv_266") +#loc179 = loc("Conv_267") +#loc180 = loc("Add_269") +#loc181 = loc("Clip_272") +#loc182 = loc("Div_274") +#loc183 = loc("Mul_275") +#loc184 = loc("Conv_276") +#loc185 = loc("Add_278") +#loc186 = loc("Clip_281") +#loc187 = loc("Div_283") +#loc188 = loc("Mul_284") +#loc189 = loc("GlobalAveragePool_285") +#loc190 = loc("Conv_286") +#loc191 = loc("Relu_287") +#loc192 = loc("Conv_288") +#loc193 = loc("Add_290") +#loc194 = loc("Clip_293") +#loc195 = loc("Div_295") +#loc196 = loc("Mul_296") +#loc197 = loc("Conv_297") +#loc198 = loc("Add_298") +#loc199 = loc("Conv_299") +#loc200 = loc("Add_301") +#loc201 = loc("Clip_304") +#loc202 = loc("Div_306") +#loc203 = loc("Mul_307") +#loc204 = loc("Conv_308") +#loc205 = loc("Add_310") +#loc206 = loc("Clip_313") +#loc207 = loc("Div_315") +#loc208 = loc("Mul_316") +#loc209 = loc("GlobalAveragePool_317") +#loc210 = loc("Conv_318") +#loc211 = loc("Relu_319") +#loc212 = loc("Conv_320") +#loc213 = loc("Add_322") +#loc214 = loc("Clip_325") +#loc215 = loc("Div_327") +#loc216 = loc("Mul_328") +#loc217 = loc("Conv_329") +#loc218 = loc("Add_330") +#loc219 = loc("Conv_331") +#loc220 = loc("Add_333") +#loc221 = loc("Clip_336") +#loc222 = loc("Div_338") +#loc223 = loc("Mul_339") +#loc224 = loc("GlobalAveragePool_342") +#loc225 = loc("Conv_343") +#loc226 = loc("Sigmoid_344") +#loc227 = loc("Mul_345") +#loc228 = loc("Conv_340") +#loc229 = loc("Relu_341") +#loc230 = loc("Split_349") +#loc231 = loc("Concat_350") +#loc232 = loc("Conv_351") +#loc233 = loc("Sigmoid_352") +#loc234 = loc("Split_353") +#loc235 = loc("Mul_354") +#loc236 = loc("Concat_355") +#loc237 = loc("Conv_356") +#loc238 = loc("Tanh_357") +#loc239 = loc("Mul_361") +#loc240 = loc("Mul_360") +#loc241 = loc("Add_362") +#loc242 = loc("Concat_363") +#loc243 = loc("Resize_365") +#loc244 = loc("Slice_371") +#loc245 = loc("AveragePool_347") +#loc246 = loc("AveragePool_348") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_380") +#loc256 = loc("Concat_381") +#loc257 = loc("Conv_382") +#loc258 = loc("Tanh_383") +#loc259 = loc("Mul_387") +#loc260 = loc("Mul_386") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_406") +#loc274 = loc("Concat_407") +#loc275 = loc("Conv_408") +#loc276 = loc("Tanh_409") +#loc277 = loc("Mul_413") +#loc278 = loc("Mul_412") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_426") +#loc291 = loc("Concat_427") +#loc292 = loc("Conv_428") +#loc293 = loc("Tanh_429") +#loc294 = loc("Mul_433") +#loc295 = loc("Mul_432") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_446") +#loc307 = loc("Clip_447") +#loc308 = loc(fused[#loc1, #loc2, #loc3, #loc4, #loc5, #loc6, #loc7, #loc8, #loc9, #loc10, #loc11, #loc12, #loc13, #loc14, #loc15, #loc16, #loc17, #loc18, #loc19, #loc20, #loc21, #loc22, #loc23, #loc24, #loc25, #loc26, #loc27, #loc28, #loc29, #loc30, #loc31, #loc32, #loc33, #loc34, #loc35, #loc36, #loc37, #loc38, #loc39, #loc40, #loc41, #loc42, #loc43, #loc44, #loc45, #loc46, #loc47, #loc48, #loc49, #loc50, #loc51, #loc52, #loc53, #loc54, #loc55, #loc56, #loc57, #loc58, #loc59, #loc60, #loc61, #loc62, #loc63, #loc64, #loc65, #loc66, #loc67, #loc68, #loc69, #loc70, #loc71, #loc72, #loc73, #loc74, #loc75, #loc76, #loc77, #loc78, #loc79, #loc80, #loc81, #loc82, #loc83, #loc84, #loc85, #loc86, #loc87, #loc88, #loc89, #loc90, #loc91, #loc92, #loc93, #loc94, #loc95, #loc96, #loc97, #loc98, #loc99, #loc100, #loc101, #loc102, #loc103, #loc104, #loc105, #loc106, #loc107, #loc108, #loc109, #loc110, #loc111, #loc112, #loc113, #loc114, #loc115, #loc116, #loc117, #loc118, #loc119, #loc120, #loc121, #loc122, #loc123, #loc124, #loc125, #loc126, #loc127, #loc128, #loc129, #loc130, #loc131, #loc132, #loc133, #loc134, #loc135, #loc136, #loc137, #loc138, #loc139, #loc140, #loc141, #loc142, #loc143, #loc144, #loc145, #loc146, #loc147, #loc148, #loc149, #loc150, #loc151, #loc152, #loc153, #loc154, #loc155, #loc156, #loc157, #loc158, #loc159, #loc160, #loc161, #loc162, #loc163, #loc164, #loc165, #loc166, #loc167, #loc168, #loc169, #loc170, #loc171, #loc172, #loc173, #loc174, #loc175, #loc176, #loc177, #loc178, #loc179, #loc180, #loc181, #loc182, #loc183, #loc184, #loc185, #loc186, #loc187, #loc188, #loc189, #loc190, #loc191, #loc192, #loc193, #loc194, #loc195, #loc196, #loc197, #loc198, #loc199, #loc200, #loc201, #loc202, #loc203, #loc204, #loc205, #loc206, #loc207, #loc208, #loc209, #loc210, #loc211, #loc212, #loc213, #loc214, #loc215, #loc216, #loc217, #loc218, #loc219, #loc220, #loc221, #loc222, #loc223, #loc224, #loc225, #loc226, #loc227, #loc228, #loc229, #loc230, #loc231, #loc232, #loc233, #loc234, #loc235, #loc236, #loc237, #loc238, #loc239, #loc240, #loc241, #loc242, #loc243, #loc244, #loc245, #loc246, #loc247, #loc248, #loc249, #loc250, #loc251, #loc252, #loc253, #loc254, #loc255, #loc256, #loc257, #loc258, #loc259, #loc260, #loc261, #loc262, #loc263, #loc264, #loc265, #loc266, #loc267, #loc268, #loc269, #loc270, #loc271, #loc272, #loc273, #loc274, #loc275, #loc276, #loc277, #loc278, #loc279, #loc280, #loc281, #loc282, #loc283, #loc284, #loc285, #loc286, #loc287, #loc288, #loc289, #loc290, #loc291, #loc292, #loc293, #loc294, #loc295, #loc296, #loc297, #loc298, #loc299, #loc300, #loc301, #loc302, #loc303, #loc304, #loc305, #loc306, #loc307]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.o b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.o new file mode 100644 index 0000000000000000000000000000000000000000..8bd002778ad62e52e43706edc8af683ad03c21c0 Binary files /dev/null and b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.o differ diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.onnx b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.onnx new file mode 100644 index 0000000000000000000000000000000000000000..c808ae5ea6b47d22c3a8df3f631e050df634b77c --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.onnx @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71f96c285f8d7bab32597f49b7abc034cf73dc3ce830377c67d747b197d277cc +size 49136 diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.onnx.mlir b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.onnx.mlir new file mode 100644 index 0000000000000000000000000000000000000000..fccd349b50617bf76bfd7a68e1f9675f2d410977 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.onnx.mlir @@ -0,0 +1,1602 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func @main_graph(%arg0: tensor<1x180x320x4xf32> {onnx.name = "385"} loc(unknown), %arg1: tensor<1x16x90x160xf32> {onnx.name = "394"} loc(unknown), %arg2: tensor<1x20x45x80xf32> {onnx.name = "395"} loc(unknown), %arg3: tensor<1x40x23x40xf32> {onnx.name = "396"} loc(unknown), %arg4: tensor<1x64x12x20xf32> {onnx.name = "397"} loc(unknown)) -> (tensor<1x1x180x320xf32> {onnx.name = "921"}, tensor<1x16x90x160xf32> {onnx.name = "894"}, tensor<1x20x45x80xf32> {onnx.name = "868"}, tensor<1x40x23x40xf32> {onnx.name = "832"}, tensor<1x64x12x20xf32> {onnx.name = "796"}, tensor<1x3x180x320xf32> {onnx.name = "916"}) { + %0 = onnx.Constant dense<[[[-4.850000e-01]], [[-4.560000e-01]], [[-4.060000e-01]]]> : tensor<3x1x1xf32> loc(#loc) + %1 = onnx.Constant dense<[3, 1]> : tensor<2xi64> loc(#loc) + %2 = onnx.Constant dense<16> : tensor<2xi64> loc(#loc) + %3 = onnx.Constant dense<20> : tensor<2xi64> loc(#loc) + %4 = onnx.Constant dense<40> : tensor<2xi64> loc(#loc) + %5 = onnx.Constant dense<64> : tensor<2xi64> loc(#loc) + %6 = "onnx.NoValue"() {onnx_node_name = "onnx.NoValue_33", value} : () -> none loc(#loc) + %7 = onnx.Constant dense_resource<__elided__> : tensor<184x80x1x1xf32> loc(#loc1) + %8 = onnx.Constant dense_resource<__elided__> : tensor<184xf32> loc(#loc2) + %9 = onnx.Constant dense_resource<__elided__> : tensor<184x1x3x3xf32> loc(#loc3) + %10 = onnx.Constant dense_resource<__elided__> : tensor<184xf32> loc(#loc4) + %11 = onnx.Constant dense_resource<__elided__> : tensor<80x184x1x1xf32> loc(#loc5) + %12 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc6) + %13 = onnx.Constant dense_resource<__elided__> : tensor<184x80x1x1xf32> loc(#loc7) + %14 = onnx.Constant dense_resource<__elided__> : tensor<184xf32> loc(#loc8) + %15 = onnx.Constant dense_resource<__elided__> : tensor<184x1x3x3xf32> loc(#loc9) + %16 = onnx.Constant dense_resource<__elided__> : tensor<184xf32> loc(#loc10) + %17 = onnx.Constant dense_resource<__elided__> : tensor<80x184x1x1xf32> loc(#loc11) + %18 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc12) + %19 = onnx.Constant dense_resource<__elided__> : tensor<480x80x1x1xf32> loc(#loc13) + %20 = onnx.Constant dense_resource<__elided__> : tensor<480xf32> loc(#loc14) + %21 = onnx.Constant dense_resource<__elided__> : tensor<480x1x3x3xf32> loc(#loc15) + %22 = onnx.Constant dense_resource<__elided__> : tensor<480xf32> loc(#loc16) + %23 = onnx.Constant dense_resource<__elided__> : tensor<112x480x1x1xf32> loc(#loc17) + %24 = onnx.Constant dense_resource<__elided__> : tensor<112xf32> loc(#loc18) + %25 = onnx.Constant dense_resource<__elided__> : tensor<672x112x1x1xf32> loc(#loc19) + %26 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc20) + %27 = onnx.Constant dense_resource<__elided__> : tensor<672x1x3x3xf32> loc(#loc21) + %28 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc22) + %29 = onnx.Constant dense_resource<__elided__> : tensor<112x672x1x1xf32> loc(#loc23) + %30 = onnx.Constant dense_resource<__elided__> : tensor<112xf32> loc(#loc24) + %31 = onnx.Constant dense_resource<__elided__> : tensor<672x112x1x1xf32> loc(#loc25) + %32 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc26) + %33 = onnx.Constant dense_resource<__elided__> : tensor<672x1x5x5xf32> loc(#loc27) + %34 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc28) + %35 = onnx.Constant dense_resource<__elided__> : tensor<160x672x1x1xf32> loc(#loc29) + %36 = onnx.Constant dense_resource<__elided__> : tensor<160xf32> loc(#loc30) + %37 = onnx.Constant dense_resource<__elided__> : tensor<960x160x1x1xf32> loc(#loc31) + %38 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc32) + %39 = onnx.Constant dense_resource<__elided__> : tensor<960x1x5x5xf32> loc(#loc33) + %40 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc34) + %41 = onnx.Constant dense_resource<__elided__> : tensor<160x960x1x1xf32> loc(#loc35) + %42 = onnx.Constant dense_resource<__elided__> : tensor<160xf32> loc(#loc36) + %43 = onnx.Constant dense_resource<__elided__> : tensor<960x160x1x1xf32> loc(#loc37) + %44 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc38) + %45 = onnx.Constant dense_resource<__elided__> : tensor<960x1x5x5xf32> loc(#loc39) + %46 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc40) + %47 = onnx.Constant dense_resource<__elided__> : tensor<160x960x1x1xf32> loc(#loc41) + %48 = onnx.Constant dense_resource<__elided__> : tensor<160xf32> loc(#loc42) + %49 = onnx.Constant dense_resource<__elided__> : tensor<960x160x1x1xf32> loc(#loc43) + %50 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc44) + %51 = onnx.Constant dense_resource<__elided__> : tensor<128x960x1x1xf32> loc(#loc45) + %52 = onnx.Constant dense_resource<__elided__> : tensor<128xf32> loc(#loc46) + %53 = onnx.Constant dense_resource<__elided__> : tensor<80x171x3x3xf32> loc(#loc47) + %54 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc48) + %55 = onnx.Constant dense_resource<__elided__> : tensor<40x107x3x3xf32> loc(#loc49) + %56 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc50) + %57 = onnx.Constant dense_resource<__elided__> : tensor<32x59x3x3xf32> loc(#loc51) + %58 = onnx.Constant dense_resource<__elided__> : tensor<32xf32> loc(#loc52) + %59 = onnx.Constant dense_resource<__elided__> : tensor<16x35x3x3xf32> loc(#loc53) + %60 = onnx.Constant dense<[0.281471759, -0.0896756947, 0.0517602414, -0.266139954, 0.132527292, 0.684469878, -0.0511226803, 0.859402895, 0.504835129, 0.569725394, 0.217058718, -0.0543790609, -0.30986914, 0.451566547, 0.166573063, 0.415171683]> : tensor<16xf32> loc(#loc54) + %61 = onnx.Constant dense_resource<__elided__> : tensor<16x16x3x3xf32> loc(#loc55) + %62 = onnx.Constant dense<[-0.257117867, -0.332320929, -0.342930794, 0.882337093, -0.811691761, 1.04650748, 1.993430e-01, 0.471133053, 0.0722430944, 0.554342687, 1.3374486, 0.48697716, 1.31853354, 0.714223623, 1.16618729, 0.738572299]> : tensor<16xf32> loc(#loc56) + %63 = onnx.Constant dense<2> : tensor<1xi64> loc(#loc57) + %64 = onnx.Constant dense<[1.000000e+00, 1.000000e+00, 2.000000e+00, 2.000000e+00]> : tensor<4xf32> loc(#loc58) + %65 = onnx.Constant dense<2.550000e+02> : tensor loc(#loc59) + %66 = onnx.Constant dense<3> : tensor<1xi64> loc(#loc60) + %67 = onnx.Constant dense<0> : tensor<1xi64> loc(#loc61) + %68 = onnx.Constant dense<[[[2.290000e-01]], [[2.240000e-01]], [[2.250000e-01]]]> : tensor<3x1x1xf32> loc(#loc62) + %69 = onnx.Constant dense<0.000000e+00> : tensor loc(#loc63) + %70 = onnx.Constant dense<6.000000e+00> : tensor loc(#loc64) + %71 = onnx.Constant dense<3.000000e+00> : tensor loc(#loc65) + %72 = onnx.Constant dense<23> : tensor<1xi64> loc(#loc66) + %73 = onnx.Constant dense<45> : tensor<1xi64> loc(#loc67) + %74 = onnx.Constant dense<1> : tensor<1xi64> loc(#loc68) + %75 = onnx.Constant dense<1.000000e+00> : tensor loc(#loc69) + %76 = onnx.Constant dense_resource<__elided__> : tensor<16x3x3x3xf32> loc(#loc70) + %77 = onnx.Constant dense<[2.98861408, -1.22985208, 2.43826318, -3.98499513, 4.62797928, 2.54142761, 2.45345306, 2.64061832, 2.13576674, 2.30800247, -0.198341176, -0.427822977, -1.09159482, 4.85548782, 2.70597649, 2.6902504]> : tensor<16xf32> loc(#loc71) + %78 = onnx.Constant dense_resource<__elided__> : tensor<16x1x3x3xf32> loc(#loc72) + %79 = onnx.Constant dense<[-4.38406658, -1.06764766E-8, -0.704851329, -1.05036237E-8, -4.89120433E-9, 1.53249037, -0.0617836975, 2.16366434, 0.0416259095, -4.12739087E-9, -3.50249429E-9, -7.75795516E-9, -4.04315559E-9, 0.292217016, -0.010752866, 1.63358212]> : tensor<16xf32> loc(#loc73) + %80 = onnx.Constant dense_resource<__elided__> : tensor<16x16x1x1xf32> loc(#loc74) + %81 = onnx.Constant dense<[-1.31068802, 0.586562276, 5.67538071, 0.551027656, 2.19523954, 3.83854461, 0.0600251146, -2.18778157, -1.5404067, 2.044780e+00, -4.23846388, 0.703142225, -8.39978456E-5, 3.50620365, -0.531753063, -5.91183185]> : tensor<16xf32> loc(#loc75) + %82 = onnx.Constant dense_resource<__elided__> : tensor<64x16x1x1xf32> loc(#loc76) + %83 = onnx.Constant dense_resource<__elided__> : tensor<64xf32> loc(#loc77) + %84 = onnx.Constant dense_resource<__elided__> : tensor<64x1x3x3xf32> loc(#loc78) + %85 = onnx.Constant dense_resource<__elided__> : tensor<64xf32> loc(#loc79) + %86 = onnx.Constant dense_resource<__elided__> : tensor<24x64x1x1xf32> loc(#loc80) + %87 = onnx.Constant dense_resource<__elided__> : tensor<24xf32> loc(#loc81) + %88 = onnx.Constant dense_resource<__elided__> : tensor<72x24x1x1xf32> loc(#loc82) + %89 = onnx.Constant dense_resource<__elided__> : tensor<72xf32> loc(#loc83) + %90 = onnx.Constant dense_resource<__elided__> : tensor<72x1x3x3xf32> loc(#loc84) + %91 = onnx.Constant dense_resource<__elided__> : tensor<72xf32> loc(#loc85) + %92 = onnx.Constant dense_resource<__elided__> : tensor<24x72x1x1xf32> loc(#loc86) + %93 = onnx.Constant dense_resource<__elided__> : tensor<24xf32> loc(#loc87) + %94 = onnx.Constant dense_resource<__elided__> : tensor<72x24x1x1xf32> loc(#loc88) + %95 = onnx.Constant dense_resource<__elided__> : tensor<72xf32> loc(#loc89) + %96 = onnx.Constant dense_resource<__elided__> : tensor<72x1x5x5xf32> loc(#loc90) + %97 = onnx.Constant dense_resource<__elided__> : tensor<72xf32> loc(#loc91) + %98 = onnx.Constant dense_resource<__elided__> : tensor<40x72x1x1xf32> loc(#loc92) + %99 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc93) + %100 = onnx.Constant dense_resource<__elided__> : tensor<120x40x1x1xf32> loc(#loc94) + %101 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc95) + %102 = onnx.Constant dense_resource<__elided__> : tensor<120x1x5x5xf32> loc(#loc96) + %103 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc97) + %104 = onnx.Constant dense_resource<__elided__> : tensor<40x120x1x1xf32> loc(#loc98) + %105 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc99) + %106 = onnx.Constant dense_resource<__elided__> : tensor<120x40x1x1xf32> loc(#loc100) + %107 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc101) + %108 = onnx.Constant dense_resource<__elided__> : tensor<120x1x5x5xf32> loc(#loc102) + %109 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc103) + %110 = onnx.Constant dense_resource<__elided__> : tensor<40x120x1x1xf32> loc(#loc104) + %111 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc105) + %112 = onnx.Constant dense_resource<__elided__> : tensor<240x40x1x1xf32> loc(#loc106) + %113 = onnx.Constant dense_resource<__elided__> : tensor<240xf32> loc(#loc107) + %114 = onnx.Constant dense_resource<__elided__> : tensor<240x1x3x3xf32> loc(#loc108) + %115 = onnx.Constant dense_resource<__elided__> : tensor<240xf32> loc(#loc109) + %116 = onnx.Constant dense_resource<__elided__> : tensor<80x240x1x1xf32> loc(#loc110) + %117 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc111) + %118 = onnx.Constant dense_resource<__elided__> : tensor<200x80x1x1xf32> loc(#loc112) + %119 = onnx.Constant dense_resource<__elided__> : tensor<200xf32> loc(#loc113) + %120 = onnx.Constant dense_resource<__elided__> : tensor<200x1x3x3xf32> loc(#loc114) + %121 = onnx.Constant dense_resource<__elided__> : tensor<200xf32> loc(#loc115) + %122 = onnx.Constant dense_resource<__elided__> : tensor<80x200x1x1xf32> loc(#loc116) + %123 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc117) + %124 = onnx.Constant dense_resource<__elided__> : tensor<128x960x1x1xf32> loc(#loc118) + %125 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc119) + %126 = onnx.Constant dense_resource<__elided__> : tensor<120x480x1x1xf32> loc(#loc120) + %127 = onnx.Constant dense_resource<__elided__> : tensor<480xf32> loc(#loc121) + %128 = onnx.Constant dense_resource<__elided__> : tensor<480x120x1x1xf32> loc(#loc122) + %129 = onnx.Constant dense_resource<__elided__> : tensor<168xf32> loc(#loc123) + %130 = onnx.Constant dense_resource<__elided__> : tensor<168x672x1x1xf32> loc(#loc124) + %131 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc125) + %132 = onnx.Constant dense_resource<__elided__> : tensor<672x168x1x1xf32> loc(#loc126) + %133 = onnx.Constant dense_resource<__elided__> : tensor<168xf32> loc(#loc127) + %134 = onnx.Constant dense_resource<__elided__> : tensor<168x672x1x1xf32> loc(#loc128) + %135 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc129) + %136 = onnx.Constant dense_resource<__elided__> : tensor<672x168x1x1xf32> loc(#loc130) + %137 = onnx.Constant dense_resource<__elided__> : tensor<240xf32> loc(#loc131) + %138 = onnx.Constant dense_resource<__elided__> : tensor<240x960x1x1xf32> loc(#loc132) + %139 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc133) + %140 = onnx.Constant dense_resource<__elided__> : tensor<960x240x1x1xf32> loc(#loc134) + %141 = onnx.Constant dense_resource<__elided__> : tensor<240xf32> loc(#loc135) + %142 = onnx.Constant dense_resource<__elided__> : tensor<240x960x1x1xf32> loc(#loc136) + %143 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc137) + %144 = onnx.Constant dense_resource<__elided__> : tensor<960x240x1x1xf32> loc(#loc138) + %145 = onnx.Constant dense_resource<__elided__> : tensor<24xf32> loc(#loc139) + %146 = onnx.Constant dense_resource<__elided__> : tensor<24x72x1x1xf32> loc(#loc140) + %147 = onnx.Constant dense_resource<__elided__> : tensor<72xf32> loc(#loc141) + %148 = onnx.Constant dense_resource<__elided__> : tensor<72x24x1x1xf32> loc(#loc142) + %149 = onnx.Constant dense_resource<__elided__> : tensor<32xf32> loc(#loc143) + %150 = onnx.Constant dense_resource<__elided__> : tensor<32x120x1x1xf32> loc(#loc144) + %151 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc145) + %152 = onnx.Constant dense_resource<__elided__> : tensor<120x32x1x1xf32> loc(#loc146) + %153 = onnx.Constant dense_resource<__elided__> : tensor<32xf32> loc(#loc147) + %154 = onnx.Constant dense_resource<__elided__> : tensor<32x120x1x1xf32> loc(#loc148) + %155 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc149) + %156 = onnx.Constant dense_resource<__elided__> : tensor<120x32x1x1xf32> loc(#loc150) + %157 = onnx.Constant dense<[0.00544366054, 0.154367775, 0.115729354, 0.171141103, -0.168815523, 0.0456937179, 0.188233331, 0.0151384082, 0.242783383, -0.139173314, -0.24988465, -9.479440e-02, -0.055940561, -0.0512795448, -0.0738077834, 0.0476587117]> : tensor<16xf32> loc(#loc151) + %158 = onnx.Constant dense_resource<__elided__> : tensor<16x32x3x3xf32> loc(#loc152) + %159 = onnx.Constant dense_resource<__elided__> : tensor<32xf32> loc(#loc153) + %160 = onnx.Constant dense_resource<__elided__> : tensor<32x32x3x3xf32> loc(#loc154) + %161 = onnx.Constant dense_resource<__elided__> : tensor<20xf32> loc(#loc155) + %162 = onnx.Constant dense_resource<__elided__> : tensor<20x40x3x3xf32> loc(#loc156) + %163 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc157) + %164 = onnx.Constant dense_resource<__elided__> : tensor<40x40x3x3xf32> loc(#loc158) + %165 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc159) + %166 = onnx.Constant dense_resource<__elided__> : tensor<40x80x3x3xf32> loc(#loc160) + %167 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc161) + %168 = onnx.Constant dense_resource<__elided__> : tensor<80x80x3x3xf32> loc(#loc162) + %169 = onnx.Constant dense_resource<__elided__> : tensor<64xf32> loc(#loc163) + %170 = onnx.Constant dense_resource<__elided__> : tensor<64x128x3x3xf32> loc(#loc164) + %171 = onnx.Constant dense_resource<__elided__> : tensor<128xf32> loc(#loc165) + %172 = onnx.Constant dense_resource<__elided__> : tensor<128x128x3x3xf32> loc(#loc166) + %173 = onnx.Constant dense<[0.00409470545, 0.00284675183, 0.00200544903, 0.124928087]> : tensor<4xf32> loc(#loc167) + %174 = onnx.Constant dense_resource<__elided__> : tensor<4x16x1x1xf32> loc(#loc168) + %175 = "onnx.Div"(%arg0, %65) {onnx_node_name = "Div_2"} : (tensor<1x180x320x4xf32>, tensor) -> tensor<1x180x320x4xf32> loc(#loc169) + %176 = "onnx.Slice"(%175, %67, %66, %66, %74) {onnx_node_name = "Slice_7"} : (tensor<1x180x320x4xf32>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>) -> tensor<1x180x320x3xf32> loc(#loc170) + %177 = "onnx.Transpose"(%176) {onnx_node_name = "Transpose_8", perm = [0, 3, 1, 2]} : (tensor<1x180x320x3xf32>) -> tensor<1x3x180x320xf32> loc(#loc171) + %178 = "onnx.Add"(%177, %0) {onnx_node_name = "Sub_14-Initializer_398_30"} : (tensor<1x3x180x320xf32>, tensor<3x1x1xf32>) -> tensor<1x3x180x320xf32> loc(#loc476) + %179 = "onnx.Div"(%178, %68) {onnx_node_name = "Div_16"} : (tensor<1x3x180x320xf32>, tensor<3x1x1xf32>) -> tensor<1x3x180x320xf32> loc(#loc174) + %180 = "onnx.Conv"(%179, %76, %77) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_17", + pads = [1, 1, 1, 1], + strides = [2, 2]} : (tensor<1x3x180x320xf32>, tensor<16x3x3x3xf32>, tensor<16xf32>) -> tensor<1x16x90x160xf32> loc(#loc175) + %181 = "onnx.Add"(%180, %71) {onnx_node_name = "Add_19"} : (tensor<1x16x90x160xf32>, tensor) -> tensor<1x16x90x160xf32> loc(#loc176) + %182 = "onnx.Clip"(%181, %69, %70) {onnx_node_name = "Clip_22_12"} : (tensor<1x16x90x160xf32>, tensor, tensor) -> tensor<1x16x90x160xf32> loc(#loc177) + %183 = "onnx.Div"(%182, %70) {onnx_node_name = "Div_24"} : (tensor<1x16x90x160xf32>, tensor) -> tensor<1x16x90x160xf32> loc(#loc178) + %184 = "onnx.Mul"(%180, %183) {onnx_node_name = "Mul_25"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc179) + %185 = "onnx.Conv"(%184, %78, %79) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 16 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_26", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x16x90x160xf32>, tensor<16x1x3x3xf32>, tensor<16xf32>) -> tensor<1x16x90x160xf32> loc(#loc180) + %186 = "onnx.Relu"(%185) {onnx_node_name = "Relu_27"} : (tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc181) + %187 = "onnx.Conv"(%186, %80, %81) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_28", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x16x90x160xf32>, tensor<16x16x1x1xf32>, tensor<16xf32>) -> tensor<1x16x90x160xf32> loc(#loc182) + %188 = "onnx.Add"(%187, %184) {onnx_node_name = "Add_29"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc183) + %189 = "onnx.Conv"(%188, %82, %83) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_30", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x16x90x160xf32>, tensor<64x16x1x1xf32>, tensor<64xf32>) -> tensor<1x64x90x160xf32> loc(#loc184) + %190 = "onnx.Relu"(%189) {onnx_node_name = "Relu_31"} : (tensor<1x64x90x160xf32>) -> tensor<1x64x90x160xf32> loc(#loc185) + %191 = "onnx.Conv"(%190, %84, %85) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 64 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_32", + pads = [1, 1, 1, 1], + strides = [2, 2]} : (tensor<1x64x90x160xf32>, tensor<64x1x3x3xf32>, tensor<64xf32>) -> tensor<1x64x45x80xf32> loc(#loc186) + %192 = "onnx.Relu"(%191) {onnx_node_name = "Relu_33"} : (tensor<1x64x45x80xf32>) -> tensor<1x64x45x80xf32> loc(#loc187) + %193 = "onnx.Conv"(%192, %86, %87) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_34", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x64x45x80xf32>, tensor<24x64x1x1xf32>, tensor<24xf32>) -> tensor<1x24x45x80xf32> loc(#loc188) + %194 = "onnx.Conv"(%193, %88, %89) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_35", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x24x45x80xf32>, tensor<72x24x1x1xf32>, tensor<72xf32>) -> tensor<1x72x45x80xf32> loc(#loc189) + %195 = "onnx.Relu"(%194) {onnx_node_name = "Relu_36"} : (tensor<1x72x45x80xf32>) -> tensor<1x72x45x80xf32> loc(#loc190) + %196 = "onnx.Conv"(%195, %90, %91) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 72 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_37", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x72x45x80xf32>, tensor<72x1x3x3xf32>, tensor<72xf32>) -> tensor<1x72x45x80xf32> loc(#loc191) + %197 = "onnx.Relu"(%196) {onnx_node_name = "Relu_38"} : (tensor<1x72x45x80xf32>) -> tensor<1x72x45x80xf32> loc(#loc192) + %198 = "onnx.Conv"(%197, %92, %93) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_39", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x72x45x80xf32>, tensor<24x72x1x1xf32>, tensor<24xf32>) -> tensor<1x24x45x80xf32> loc(#loc193) + %199 = "onnx.Add"(%198, %193) {onnx_node_name = "Add_40"} : (tensor<1x24x45x80xf32>, tensor<1x24x45x80xf32>) -> tensor<1x24x45x80xf32> loc(#loc194) + %200 = "onnx.Conv"(%199, %94, %95) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_41", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x24x45x80xf32>, tensor<72x24x1x1xf32>, tensor<72xf32>) -> tensor<1x72x45x80xf32> loc(#loc195) + %201 = "onnx.Relu"(%200) {onnx_node_name = "Relu_42"} : (tensor<1x72x45x80xf32>) -> tensor<1x72x45x80xf32> loc(#loc196) + %202 = "onnx.Conv"(%201, %96, %97) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 72 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_43", + pads = [2, 2, 2, 2], + strides = [2, 2]} : (tensor<1x72x45x80xf32>, tensor<72x1x5x5xf32>, tensor<72xf32>) -> tensor<1x72x23x40xf32> loc(#loc197) + %203 = "onnx.Relu"(%202) {onnx_node_name = "Relu_44"} : (tensor<1x72x23x40xf32>) -> tensor<1x72x23x40xf32> loc(#loc198) + %204 = "onnx.ReduceMeanV13"(%203) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_45_45"} : (tensor<1x72x23x40xf32>) -> tensor<1x72x1x1xf32> loc(#loc199) + %205 = "onnx.Conv"(%204, %146, %145) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_46", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x72x1x1xf32>, tensor<24x72x1x1xf32>, tensor<24xf32>) -> tensor<1x24x1x1xf32> loc(#loc200) + %206 = "onnx.Relu"(%205) {onnx_node_name = "Relu_47"} : (tensor<1x24x1x1xf32>) -> tensor<1x24x1x1xf32> loc(#loc201) + %207 = "onnx.Conv"(%206, %148, %147) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_48", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x24x1x1xf32>, tensor<72x24x1x1xf32>, tensor<72xf32>) -> tensor<1x72x1x1xf32> loc(#loc202) + %208 = "onnx.Add"(%207, %71) {onnx_node_name = "Add_50"} : (tensor<1x72x1x1xf32>, tensor) -> tensor<1x72x1x1xf32> loc(#loc203) + %209 = "onnx.Clip"(%208, %69, %70) {onnx_node_name = "Clip_53_32"} : (tensor<1x72x1x1xf32>, tensor, tensor) -> tensor<1x72x1x1xf32> loc(#loc204) + %210 = "onnx.Div"(%209, %70) {onnx_node_name = "Div_55"} : (tensor<1x72x1x1xf32>, tensor) -> tensor<1x72x1x1xf32> loc(#loc205) + %211 = "onnx.Mul"(%210, %203) {onnx_node_name = "Mul_56"} : (tensor<1x72x1x1xf32>, tensor<1x72x23x40xf32>) -> tensor<1x72x23x40xf32> loc(#loc206) + %212 = "onnx.Conv"(%211, %98, %99) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_57", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x72x23x40xf32>, tensor<40x72x1x1xf32>, tensor<40xf32>) -> tensor<1x40x23x40xf32> loc(#loc207) + %213 = "onnx.Conv"(%212, %100, %101) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_58", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x40x23x40xf32>, tensor<120x40x1x1xf32>, tensor<120xf32>) -> tensor<1x120x23x40xf32> loc(#loc208) + %214 = "onnx.Relu"(%213) {onnx_node_name = "Relu_59"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc209) + %215 = "onnx.Conv"(%214, %102, %103) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 120 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_60", + pads = [2, 2, 2, 2], + strides = [1, 1]} : (tensor<1x120x23x40xf32>, tensor<120x1x5x5xf32>, tensor<120xf32>) -> tensor<1x120x23x40xf32> loc(#loc210) + %216 = "onnx.Relu"(%215) {onnx_node_name = "Relu_61"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc211) + %217 = "onnx.ReduceMeanV13"(%216) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_62_51"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x1x1xf32> loc(#loc212) + %218 = "onnx.Conv"(%217, %150, %149) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_63", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x120x1x1xf32>, tensor<32x120x1x1xf32>, tensor<32xf32>) -> tensor<1x32x1x1xf32> loc(#loc213) + %219 = "onnx.Relu"(%218) {onnx_node_name = "Relu_64"} : (tensor<1x32x1x1xf32>) -> tensor<1x32x1x1xf32> loc(#loc214) + %220 = "onnx.Conv"(%219, %152, %151) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_65", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x32x1x1xf32>, tensor<120x32x1x1xf32>, tensor<120xf32>) -> tensor<1x120x1x1xf32> loc(#loc215) + %221 = "onnx.Add"(%220, %71) {onnx_node_name = "Add_67"} : (tensor<1x120x1x1xf32>, tensor) -> tensor<1x120x1x1xf32> loc(#loc216) + %222 = "onnx.Clip"(%221, %69, %70) {onnx_node_name = "Clip_70_4"} : (tensor<1x120x1x1xf32>, tensor, tensor) -> tensor<1x120x1x1xf32> loc(#loc217) + %223 = "onnx.Div"(%222, %70) {onnx_node_name = "Div_72"} : (tensor<1x120x1x1xf32>, tensor) -> tensor<1x120x1x1xf32> loc(#loc218) + %224 = "onnx.Mul"(%223, %216) {onnx_node_name = "Mul_73"} : (tensor<1x120x1x1xf32>, tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc219) + %225 = "onnx.Conv"(%224, %104, %105) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_74", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x120x23x40xf32>, tensor<40x120x1x1xf32>, tensor<40xf32>) -> tensor<1x40x23x40xf32> loc(#loc220) + %226 = "onnx.Add"(%225, %212) {onnx_node_name = "Add_75"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc221) + %227 = "onnx.Conv"(%226, %106, %107) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_76", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x40x23x40xf32>, tensor<120x40x1x1xf32>, tensor<120xf32>) -> tensor<1x120x23x40xf32> loc(#loc222) + %228 = "onnx.Relu"(%227) {onnx_node_name = "Relu_77"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc223) + %229 = "onnx.Conv"(%228, %108, %109) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 120 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_78", + pads = [2, 2, 2, 2], + strides = [1, 1]} : (tensor<1x120x23x40xf32>, tensor<120x1x5x5xf32>, tensor<120xf32>) -> tensor<1x120x23x40xf32> loc(#loc224) + %230 = "onnx.Relu"(%229) {onnx_node_name = "Relu_79"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc225) + %231 = "onnx.ReduceMeanV13"(%230) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_80_18"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x1x1xf32> loc(#loc226) + %232 = "onnx.Conv"(%231, %154, %153) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_81", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x120x1x1xf32>, tensor<32x120x1x1xf32>, tensor<32xf32>) -> tensor<1x32x1x1xf32> loc(#loc227) + %233 = "onnx.Relu"(%232) {onnx_node_name = "Relu_82"} : (tensor<1x32x1x1xf32>) -> tensor<1x32x1x1xf32> loc(#loc228) + %234 = "onnx.Conv"(%233, %156, %155) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_83", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x32x1x1xf32>, tensor<120x32x1x1xf32>, tensor<120xf32>) -> tensor<1x120x1x1xf32> loc(#loc229) + %235 = "onnx.Add"(%234, %71) {onnx_node_name = "Add_85"} : (tensor<1x120x1x1xf32>, tensor) -> tensor<1x120x1x1xf32> loc(#loc230) + %236 = "onnx.Clip"(%235, %69, %70) {onnx_node_name = "Clip_88_48"} : (tensor<1x120x1x1xf32>, tensor, tensor) -> tensor<1x120x1x1xf32> loc(#loc231) + %237 = "onnx.Div"(%236, %70) {onnx_node_name = "Div_90"} : (tensor<1x120x1x1xf32>, tensor) -> tensor<1x120x1x1xf32> loc(#loc232) + %238 = "onnx.Mul"(%237, %230) {onnx_node_name = "Mul_91"} : (tensor<1x120x1x1xf32>, tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc233) + %239 = "onnx.Conv"(%238, %110, %111) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_92", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x120x23x40xf32>, tensor<40x120x1x1xf32>, tensor<40xf32>) -> tensor<1x40x23x40xf32> loc(#loc234) + %240 = "onnx.Add"(%239, %226) {onnx_node_name = "Add_93"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc235) + %241 = "onnx.Conv"(%240, %112, %113) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_94", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x40x23x40xf32>, tensor<240x40x1x1xf32>, tensor<240xf32>) -> tensor<1x240x23x40xf32> loc(#loc236) + %242 = "onnx.Add"(%241, %71) {onnx_node_name = "Add_96"} : (tensor<1x240x23x40xf32>, tensor) -> tensor<1x240x23x40xf32> loc(#loc237) + %243 = "onnx.Clip"(%242, %69, %70) {onnx_node_name = "Clip_99_36"} : (tensor<1x240x23x40xf32>, tensor, tensor) -> tensor<1x240x23x40xf32> loc(#loc238) + %244 = "onnx.Div"(%243, %70) {onnx_node_name = "Div_101"} : (tensor<1x240x23x40xf32>, tensor) -> tensor<1x240x23x40xf32> loc(#loc239) + %245 = "onnx.Mul"(%241, %244) {onnx_node_name = "Mul_102"} : (tensor<1x240x23x40xf32>, tensor<1x240x23x40xf32>) -> tensor<1x240x23x40xf32> loc(#loc240) + %246 = "onnx.Conv"(%245, %114, %115) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 240 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_103", + pads = [1, 1, 1, 1], + strides = [2, 2]} : (tensor<1x240x23x40xf32>, tensor<240x1x3x3xf32>, tensor<240xf32>) -> tensor<1x240x12x20xf32> loc(#loc241) + %247 = "onnx.Add"(%246, %71) {onnx_node_name = "Add_105"} : (tensor<1x240x12x20xf32>, tensor) -> tensor<1x240x12x20xf32> loc(#loc242) + %248 = "onnx.Clip"(%247, %69, %70) {onnx_node_name = "Clip_108_29"} : (tensor<1x240x12x20xf32>, tensor, tensor) -> tensor<1x240x12x20xf32> loc(#loc243) + %249 = "onnx.Div"(%248, %70) {onnx_node_name = "Div_110"} : (tensor<1x240x12x20xf32>, tensor) -> tensor<1x240x12x20xf32> loc(#loc244) + %250 = "onnx.Mul"(%246, %249) {onnx_node_name = "Mul_111"} : (tensor<1x240x12x20xf32>, tensor<1x240x12x20xf32>) -> tensor<1x240x12x20xf32> loc(#loc245) + %251 = "onnx.Conv"(%250, %116, %117) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_112", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x240x12x20xf32>, tensor<80x240x1x1xf32>, tensor<80xf32>) -> tensor<1x80x12x20xf32> loc(#loc246) + %252 = "onnx.Conv"(%251, %118, %119) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_113", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x80x12x20xf32>, tensor<200x80x1x1xf32>, tensor<200xf32>) -> tensor<1x200x12x20xf32> loc(#loc247) + %253 = "onnx.Add"(%252, %71) {onnx_node_name = "Add_115"} : (tensor<1x200x12x20xf32>, tensor) -> tensor<1x200x12x20xf32> loc(#loc248) + %254 = "onnx.Clip"(%253, %69, %70) {onnx_node_name = "Clip_118_42"} : (tensor<1x200x12x20xf32>, tensor, tensor) -> tensor<1x200x12x20xf32> loc(#loc249) + %255 = "onnx.Div"(%254, %70) {onnx_node_name = "Div_120"} : (tensor<1x200x12x20xf32>, tensor) -> tensor<1x200x12x20xf32> loc(#loc250) + %256 = "onnx.Mul"(%252, %255) {onnx_node_name = "Mul_121"} : (tensor<1x200x12x20xf32>, tensor<1x200x12x20xf32>) -> tensor<1x200x12x20xf32> loc(#loc251) + %257 = "onnx.Conv"(%256, %120, %121) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 200 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_122", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x200x12x20xf32>, tensor<200x1x3x3xf32>, tensor<200xf32>) -> tensor<1x200x12x20xf32> loc(#loc252) + %258 = "onnx.Add"(%257, %71) {onnx_node_name = "Add_124"} : (tensor<1x200x12x20xf32>, tensor) -> tensor<1x200x12x20xf32> loc(#loc253) + %259 = "onnx.Clip"(%258, %69, %70) {onnx_node_name = "Clip_127_39"} : (tensor<1x200x12x20xf32>, tensor, tensor) -> tensor<1x200x12x20xf32> loc(#loc254) + %260 = "onnx.Div"(%259, %70) {onnx_node_name = "Div_129"} : (tensor<1x200x12x20xf32>, tensor) -> tensor<1x200x12x20xf32> loc(#loc255) + %261 = "onnx.Mul"(%257, %260) {onnx_node_name = "Mul_130"} : (tensor<1x200x12x20xf32>, tensor<1x200x12x20xf32>) -> tensor<1x200x12x20xf32> loc(#loc256) + %262 = "onnx.Conv"(%261, %122, %123) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_131", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x200x12x20xf32>, tensor<80x200x1x1xf32>, tensor<80xf32>) -> tensor<1x80x12x20xf32> loc(#loc257) + %263 = "onnx.Add"(%262, %251) {onnx_node_name = "Add_132"} : (tensor<1x80x12x20xf32>, tensor<1x80x12x20xf32>) -> tensor<1x80x12x20xf32> loc(#loc258) + %264 = "onnx.Conv"(%263, %7, %8) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_133", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x80x12x20xf32>, tensor<184x80x1x1xf32>, tensor<184xf32>) -> tensor<1x184x12x20xf32> loc(#loc259) + %265 = "onnx.Add"(%264, %71) {onnx_node_name = "Add_135"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc260) + %266 = "onnx.Clip"(%265, %69, %70) {onnx_node_name = "Clip_138_46"} : (tensor<1x184x12x20xf32>, tensor, tensor) -> tensor<1x184x12x20xf32> loc(#loc261) + %267 = "onnx.Div"(%266, %70) {onnx_node_name = "Div_140"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc262) + %268 = "onnx.Mul"(%264, %267) {onnx_node_name = "Mul_141"} : (tensor<1x184x12x20xf32>, tensor<1x184x12x20xf32>) -> tensor<1x184x12x20xf32> loc(#loc263) + %269 = "onnx.Conv"(%268, %9, %10) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 184 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_142", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x184x12x20xf32>, tensor<184x1x3x3xf32>, tensor<184xf32>) -> tensor<1x184x12x20xf32> loc(#loc264) + %270 = "onnx.Add"(%269, %71) {onnx_node_name = "Add_144"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc265) + %271 = "onnx.Clip"(%270, %69, %70) {onnx_node_name = "Clip_147_7"} : (tensor<1x184x12x20xf32>, tensor, tensor) -> tensor<1x184x12x20xf32> loc(#loc266) + %272 = "onnx.Div"(%271, %70) {onnx_node_name = "Div_149"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc267) + %273 = "onnx.Mul"(%269, %272) {onnx_node_name = "Mul_150"} : (tensor<1x184x12x20xf32>, tensor<1x184x12x20xf32>) -> tensor<1x184x12x20xf32> loc(#loc268) + %274 = "onnx.Conv"(%273, %11, %12) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_151", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x184x12x20xf32>, tensor<80x184x1x1xf32>, tensor<80xf32>) -> tensor<1x80x12x20xf32> loc(#loc269) + %275 = "onnx.Add"(%274, %263) {onnx_node_name = "Add_152"} : (tensor<1x80x12x20xf32>, tensor<1x80x12x20xf32>) -> tensor<1x80x12x20xf32> loc(#loc270) + %276 = "onnx.Conv"(%275, %13, %14) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_153", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x80x12x20xf32>, tensor<184x80x1x1xf32>, tensor<184xf32>) -> tensor<1x184x12x20xf32> loc(#loc271) + %277 = "onnx.Add"(%276, %71) {onnx_node_name = "Add_155"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc272) + %278 = "onnx.Clip"(%277, %69, %70) {onnx_node_name = "Clip_158_24"} : (tensor<1x184x12x20xf32>, tensor, tensor) -> tensor<1x184x12x20xf32> loc(#loc273) + %279 = "onnx.Div"(%278, %70) {onnx_node_name = "Div_160"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc274) + %280 = "onnx.Mul"(%276, %279) {onnx_node_name = "Mul_161"} : (tensor<1x184x12x20xf32>, tensor<1x184x12x20xf32>) -> tensor<1x184x12x20xf32> loc(#loc275) + %281 = "onnx.Conv"(%280, %15, %16) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 184 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_162", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x184x12x20xf32>, tensor<184x1x3x3xf32>, tensor<184xf32>) -> tensor<1x184x12x20xf32> loc(#loc276) + %282 = "onnx.Add"(%281, %71) {onnx_node_name = "Add_164"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc277) + %283 = "onnx.Clip"(%282, %69, %70) {onnx_node_name = "Clip_167_37"} : (tensor<1x184x12x20xf32>, tensor, tensor) -> tensor<1x184x12x20xf32> loc(#loc278) + %284 = "onnx.Div"(%283, %70) {onnx_node_name = "Div_169"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc279) + %285 = "onnx.Mul"(%281, %284) {onnx_node_name = "Mul_170"} : (tensor<1x184x12x20xf32>, tensor<1x184x12x20xf32>) -> tensor<1x184x12x20xf32> loc(#loc280) + %286 = "onnx.Conv"(%285, %17, %18) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_171", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x184x12x20xf32>, tensor<80x184x1x1xf32>, tensor<80xf32>) -> tensor<1x80x12x20xf32> loc(#loc281) + %287 = "onnx.Add"(%286, %275) {onnx_node_name = "Add_172"} : (tensor<1x80x12x20xf32>, tensor<1x80x12x20xf32>) -> tensor<1x80x12x20xf32> loc(#loc282) + %288 = "onnx.Conv"(%287, %19, %20) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_173", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x80x12x20xf32>, tensor<480x80x1x1xf32>, tensor<480xf32>) -> tensor<1x480x12x20xf32> loc(#loc283) + %289 = "onnx.Add"(%288, %71) {onnx_node_name = "Add_175"} : (tensor<1x480x12x20xf32>, tensor) -> tensor<1x480x12x20xf32> loc(#loc284) + %290 = "onnx.Clip"(%289, %69, %70) {onnx_node_name = "Clip_178_50"} : (tensor<1x480x12x20xf32>, tensor, tensor) -> tensor<1x480x12x20xf32> loc(#loc285) + %291 = "onnx.Div"(%290, %70) {onnx_node_name = "Div_180"} : (tensor<1x480x12x20xf32>, tensor) -> tensor<1x480x12x20xf32> loc(#loc286) + %292 = "onnx.Mul"(%288, %291) {onnx_node_name = "Mul_181"} : (tensor<1x480x12x20xf32>, tensor<1x480x12x20xf32>) -> tensor<1x480x12x20xf32> loc(#loc287) + %293 = "onnx.Conv"(%292, %21, %22) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 480 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_182", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x480x12x20xf32>, tensor<480x1x3x3xf32>, tensor<480xf32>) -> tensor<1x480x12x20xf32> loc(#loc288) + %294 = "onnx.Add"(%293, %71) {onnx_node_name = "Add_184"} : (tensor<1x480x12x20xf32>, tensor) -> tensor<1x480x12x20xf32> loc(#loc289) + %295 = "onnx.Clip"(%294, %69, %70) {onnx_node_name = "Clip_187_0"} : (tensor<1x480x12x20xf32>, tensor, tensor) -> tensor<1x480x12x20xf32> loc(#loc290) + %296 = "onnx.Div"(%295, %70) {onnx_node_name = "Div_189"} : (tensor<1x480x12x20xf32>, tensor) -> tensor<1x480x12x20xf32> loc(#loc291) + %297 = "onnx.Mul"(%293, %296) {onnx_node_name = "Mul_190"} : (tensor<1x480x12x20xf32>, tensor<1x480x12x20xf32>) -> tensor<1x480x12x20xf32> loc(#loc292) + %298 = "onnx.ReduceMeanV13"(%297) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_191_26"} : (tensor<1x480x12x20xf32>) -> tensor<1x480x1x1xf32> loc(#loc293) + %299 = "onnx.Conv"(%298, %126, %125) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_192", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x480x1x1xf32>, tensor<120x480x1x1xf32>, tensor<120xf32>) -> tensor<1x120x1x1xf32> loc(#loc294) + %300 = "onnx.Relu"(%299) {onnx_node_name = "Relu_193"} : (tensor<1x120x1x1xf32>) -> tensor<1x120x1x1xf32> loc(#loc295) + %301 = "onnx.Conv"(%300, %128, %127) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_194", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x120x1x1xf32>, tensor<480x120x1x1xf32>, tensor<480xf32>) -> tensor<1x480x1x1xf32> loc(#loc296) + %302 = "onnx.Add"(%301, %71) {onnx_node_name = "Add_196"} : (tensor<1x480x1x1xf32>, tensor) -> tensor<1x480x1x1xf32> loc(#loc297) + %303 = "onnx.Clip"(%302, %69, %70) {onnx_node_name = "Clip_199_43"} : (tensor<1x480x1x1xf32>, tensor, tensor) -> tensor<1x480x1x1xf32> loc(#loc298) + %304 = "onnx.Div"(%303, %70) {onnx_node_name = "Div_201"} : (tensor<1x480x1x1xf32>, tensor) -> tensor<1x480x1x1xf32> loc(#loc299) + %305 = "onnx.Mul"(%304, %297) {onnx_node_name = "Mul_202"} : (tensor<1x480x1x1xf32>, tensor<1x480x12x20xf32>) -> tensor<1x480x12x20xf32> loc(#loc300) + %306 = "onnx.Conv"(%305, %23, %24) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_203", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x480x12x20xf32>, tensor<112x480x1x1xf32>, tensor<112xf32>) -> tensor<1x112x12x20xf32> loc(#loc301) + %307 = "onnx.Conv"(%306, %25, %26) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_204", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x112x12x20xf32>, tensor<672x112x1x1xf32>, tensor<672xf32>) -> tensor<1x672x12x20xf32> loc(#loc302) + %308 = "onnx.Add"(%307, %71) {onnx_node_name = "Add_206"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc303) + %309 = "onnx.Clip"(%308, %69, %70) {onnx_node_name = "Clip_209_28"} : (tensor<1x672x12x20xf32>, tensor, tensor) -> tensor<1x672x12x20xf32> loc(#loc304) + %310 = "onnx.Div"(%309, %70) {onnx_node_name = "Div_211"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc305) + %311 = "onnx.Mul"(%307, %310) {onnx_node_name = "Mul_212"} : (tensor<1x672x12x20xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc306) + %312 = "onnx.Conv"(%311, %27, %28) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 672 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_213", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x672x12x20xf32>, tensor<672x1x3x3xf32>, tensor<672xf32>) -> tensor<1x672x12x20xf32> loc(#loc307) + %313 = "onnx.Add"(%312, %71) {onnx_node_name = "Add_215"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc308) + %314 = "onnx.Clip"(%313, %69, %70) {onnx_node_name = "Clip_218_25"} : (tensor<1x672x12x20xf32>, tensor, tensor) -> tensor<1x672x12x20xf32> loc(#loc309) + %315 = "onnx.Div"(%314, %70) {onnx_node_name = "Div_220"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc310) + %316 = "onnx.Mul"(%312, %315) {onnx_node_name = "Mul_221"} : (tensor<1x672x12x20xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc311) + %317 = "onnx.ReduceMeanV13"(%316) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_222_47"} : (tensor<1x672x12x20xf32>) -> tensor<1x672x1x1xf32> loc(#loc312) + %318 = "onnx.Conv"(%317, %130, %129) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_223", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x672x1x1xf32>, tensor<168x672x1x1xf32>, tensor<168xf32>) -> tensor<1x168x1x1xf32> loc(#loc313) + %319 = "onnx.Relu"(%318) {onnx_node_name = "Relu_224"} : (tensor<1x168x1x1xf32>) -> tensor<1x168x1x1xf32> loc(#loc314) + %320 = "onnx.Conv"(%319, %132, %131) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_225", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x168x1x1xf32>, tensor<672x168x1x1xf32>, tensor<672xf32>) -> tensor<1x672x1x1xf32> loc(#loc315) + %321 = "onnx.Add"(%320, %71) {onnx_node_name = "Add_227"} : (tensor<1x672x1x1xf32>, tensor) -> tensor<1x672x1x1xf32> loc(#loc316) + %322 = "onnx.Clip"(%321, %69, %70) {onnx_node_name = "Clip_230_27"} : (tensor<1x672x1x1xf32>, tensor, tensor) -> tensor<1x672x1x1xf32> loc(#loc317) + %323 = "onnx.Div"(%322, %70) {onnx_node_name = "Div_232"} : (tensor<1x672x1x1xf32>, tensor) -> tensor<1x672x1x1xf32> loc(#loc318) + %324 = "onnx.Mul"(%323, %316) {onnx_node_name = "Mul_233"} : (tensor<1x672x1x1xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc319) + %325 = "onnx.Conv"(%324, %29, %30) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_234", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x672x12x20xf32>, tensor<112x672x1x1xf32>, tensor<112xf32>) -> tensor<1x112x12x20xf32> loc(#loc320) + %326 = "onnx.Add"(%325, %306) {onnx_node_name = "Add_235"} : (tensor<1x112x12x20xf32>, tensor<1x112x12x20xf32>) -> tensor<1x112x12x20xf32> loc(#loc321) + %327 = "onnx.Conv"(%326, %31, %32) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_236", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x112x12x20xf32>, tensor<672x112x1x1xf32>, tensor<672xf32>) -> tensor<1x672x12x20xf32> loc(#loc322) + %328 = "onnx.Add"(%327, %71) {onnx_node_name = "Add_238"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc323) + %329 = "onnx.Clip"(%328, %69, %70) {onnx_node_name = "Clip_241_35"} : (tensor<1x672x12x20xf32>, tensor, tensor) -> tensor<1x672x12x20xf32> loc(#loc324) + %330 = "onnx.Div"(%329, %70) {onnx_node_name = "Div_243"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc325) + %331 = "onnx.Mul"(%327, %330) {onnx_node_name = "Mul_244"} : (tensor<1x672x12x20xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc326) + %332 = "onnx.Conv"(%331, %33, %34) { + auto_pad = "NOTSET", + dilations = [2, 2], + group = 672 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_245", + pads = [4, 4, 4, 4], + strides = [1, 1]} : (tensor<1x672x12x20xf32>, tensor<672x1x5x5xf32>, tensor<672xf32>) -> tensor<1x672x12x20xf32> loc(#loc327) + %333 = "onnx.Add"(%332, %71) {onnx_node_name = "Add_247"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc328) + %334 = "onnx.Clip"(%333, %69, %70) {onnx_node_name = "Clip_250_40"} : (tensor<1x672x12x20xf32>, tensor, tensor) -> tensor<1x672x12x20xf32> loc(#loc329) + %335 = "onnx.Div"(%334, %70) {onnx_node_name = "Div_252"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc330) + %336 = "onnx.Mul"(%332, %335) {onnx_node_name = "Mul_253"} : (tensor<1x672x12x20xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc331) + %337 = "onnx.ReduceMeanV13"(%336) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_254_19"} : (tensor<1x672x12x20xf32>) -> tensor<1x672x1x1xf32> loc(#loc332) + %338 = "onnx.Conv"(%337, %134, %133) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_255", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x672x1x1xf32>, tensor<168x672x1x1xf32>, tensor<168xf32>) -> tensor<1x168x1x1xf32> loc(#loc333) + %339 = "onnx.Relu"(%338) {onnx_node_name = "Relu_256"} : (tensor<1x168x1x1xf32>) -> tensor<1x168x1x1xf32> loc(#loc334) + %340 = "onnx.Conv"(%339, %136, %135) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_257", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x168x1x1xf32>, tensor<672x168x1x1xf32>, tensor<672xf32>) -> tensor<1x672x1x1xf32> loc(#loc335) + %341 = "onnx.Add"(%340, %71) {onnx_node_name = "Add_259"} : (tensor<1x672x1x1xf32>, tensor) -> tensor<1x672x1x1xf32> loc(#loc336) + %342 = "onnx.Clip"(%341, %69, %70) {onnx_node_name = "Clip_262_49"} : (tensor<1x672x1x1xf32>, tensor, tensor) -> tensor<1x672x1x1xf32> loc(#loc337) + %343 = "onnx.Div"(%342, %70) {onnx_node_name = "Div_264"} : (tensor<1x672x1x1xf32>, tensor) -> tensor<1x672x1x1xf32> loc(#loc338) + %344 = "onnx.Mul"(%343, %336) {onnx_node_name = "Mul_265"} : (tensor<1x672x1x1xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc339) + %345 = "onnx.Conv"(%344, %35, %36) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_266", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x672x12x20xf32>, tensor<160x672x1x1xf32>, tensor<160xf32>) -> tensor<1x160x12x20xf32> loc(#loc340) + %346 = "onnx.Conv"(%345, %37, %38) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_267", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x160x12x20xf32>, tensor<960x160x1x1xf32>, tensor<960xf32>) -> tensor<1x960x12x20xf32> loc(#loc341) + %347 = "onnx.Add"(%346, %71) {onnx_node_name = "Add_269"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc342) + %348 = "onnx.Clip"(%347, %69, %70) {onnx_node_name = "Clip_272_20"} : (tensor<1x960x12x20xf32>, tensor, tensor) -> tensor<1x960x12x20xf32> loc(#loc343) + %349 = "onnx.Div"(%348, %70) {onnx_node_name = "Div_274"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc344) + %350 = "onnx.Mul"(%346, %349) {onnx_node_name = "Mul_275"} : (tensor<1x960x12x20xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc345) + %351 = "onnx.Conv"(%350, %39, %40) { + auto_pad = "NOTSET", + dilations = [2, 2], + group = 960 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_276", + pads = [4, 4, 4, 4], + strides = [1, 1]} : (tensor<1x960x12x20xf32>, tensor<960x1x5x5xf32>, tensor<960xf32>) -> tensor<1x960x12x20xf32> loc(#loc346) + %352 = "onnx.Add"(%351, %71) {onnx_node_name = "Add_278"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc347) + %353 = "onnx.Clip"(%352, %69, %70) {onnx_node_name = "Clip_281_2"} : (tensor<1x960x12x20xf32>, tensor, tensor) -> tensor<1x960x12x20xf32> loc(#loc348) + %354 = "onnx.Div"(%353, %70) {onnx_node_name = "Div_283"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc349) + %355 = "onnx.Mul"(%351, %354) {onnx_node_name = "Mul_284"} : (tensor<1x960x12x20xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc350) + %356 = "onnx.ReduceMeanV13"(%355) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_285_9"} : (tensor<1x960x12x20xf32>) -> tensor<1x960x1x1xf32> loc(#loc351) + %357 = "onnx.Conv"(%356, %138, %137) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_286", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x1x1xf32>, tensor<240x960x1x1xf32>, tensor<240xf32>) -> tensor<1x240x1x1xf32> loc(#loc352) + %358 = "onnx.Relu"(%357) {onnx_node_name = "Relu_287"} : (tensor<1x240x1x1xf32>) -> tensor<1x240x1x1xf32> loc(#loc353) + %359 = "onnx.Conv"(%358, %140, %139) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_288", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x240x1x1xf32>, tensor<960x240x1x1xf32>, tensor<960xf32>) -> tensor<1x960x1x1xf32> loc(#loc354) + %360 = "onnx.Add"(%359, %71) {onnx_node_name = "Add_290"} : (tensor<1x960x1x1xf32>, tensor) -> tensor<1x960x1x1xf32> loc(#loc355) + %361 = "onnx.Clip"(%360, %69, %70) {onnx_node_name = "Clip_293_44"} : (tensor<1x960x1x1xf32>, tensor, tensor) -> tensor<1x960x1x1xf32> loc(#loc356) + %362 = "onnx.Div"(%361, %70) {onnx_node_name = "Div_295"} : (tensor<1x960x1x1xf32>, tensor) -> tensor<1x960x1x1xf32> loc(#loc357) + %363 = "onnx.Mul"(%362, %355) {onnx_node_name = "Mul_296"} : (tensor<1x960x1x1xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc358) + %364 = "onnx.Conv"(%363, %41, %42) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_297", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x12x20xf32>, tensor<160x960x1x1xf32>, tensor<160xf32>) -> tensor<1x160x12x20xf32> loc(#loc359) + %365 = "onnx.Add"(%364, %345) {onnx_node_name = "Add_298"} : (tensor<1x160x12x20xf32>, tensor<1x160x12x20xf32>) -> tensor<1x160x12x20xf32> loc(#loc360) + %366 = "onnx.Conv"(%365, %43, %44) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_299", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x160x12x20xf32>, tensor<960x160x1x1xf32>, tensor<960xf32>) -> tensor<1x960x12x20xf32> loc(#loc361) + %367 = "onnx.Add"(%366, %71) {onnx_node_name = "Add_301"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc362) + %368 = "onnx.Clip"(%367, %69, %70) {onnx_node_name = "Clip_304_21"} : (tensor<1x960x12x20xf32>, tensor, tensor) -> tensor<1x960x12x20xf32> loc(#loc363) + %369 = "onnx.Div"(%368, %70) {onnx_node_name = "Div_306"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc364) + %370 = "onnx.Mul"(%366, %369) {onnx_node_name = "Mul_307"} : (tensor<1x960x12x20xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc365) + %371 = "onnx.Conv"(%370, %45, %46) { + auto_pad = "NOTSET", + dilations = [2, 2], + group = 960 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_308", + pads = [4, 4, 4, 4], + strides = [1, 1]} : (tensor<1x960x12x20xf32>, tensor<960x1x5x5xf32>, tensor<960xf32>) -> tensor<1x960x12x20xf32> loc(#loc366) + %372 = "onnx.Add"(%371, %71) {onnx_node_name = "Add_310"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc367) + %373 = "onnx.Clip"(%372, %69, %70) {onnx_node_name = "Clip_313_22"} : (tensor<1x960x12x20xf32>, tensor, tensor) -> tensor<1x960x12x20xf32> loc(#loc368) + %374 = "onnx.Div"(%373, %70) {onnx_node_name = "Div_315"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc369) + %375 = "onnx.Mul"(%371, %374) {onnx_node_name = "Mul_316"} : (tensor<1x960x12x20xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc370) + %376 = "onnx.ReduceMeanV13"(%375) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_317_8"} : (tensor<1x960x12x20xf32>) -> tensor<1x960x1x1xf32> loc(#loc371) + %377 = "onnx.Conv"(%376, %142, %141) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_318", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x1x1xf32>, tensor<240x960x1x1xf32>, tensor<240xf32>) -> tensor<1x240x1x1xf32> loc(#loc372) + %378 = "onnx.Relu"(%377) {onnx_node_name = "Relu_319"} : (tensor<1x240x1x1xf32>) -> tensor<1x240x1x1xf32> loc(#loc373) + %379 = "onnx.Conv"(%378, %144, %143) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_320", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x240x1x1xf32>, tensor<960x240x1x1xf32>, tensor<960xf32>) -> tensor<1x960x1x1xf32> loc(#loc374) + %380 = "onnx.Add"(%379, %71) {onnx_node_name = "Add_322"} : (tensor<1x960x1x1xf32>, tensor) -> tensor<1x960x1x1xf32> loc(#loc375) + %381 = "onnx.Clip"(%380, %69, %70) {onnx_node_name = "Clip_325_3"} : (tensor<1x960x1x1xf32>, tensor, tensor) -> tensor<1x960x1x1xf32> loc(#loc376) + %382 = "onnx.Div"(%381, %70) {onnx_node_name = "Div_327"} : (tensor<1x960x1x1xf32>, tensor) -> tensor<1x960x1x1xf32> loc(#loc377) + %383 = "onnx.Mul"(%382, %375) {onnx_node_name = "Mul_328"} : (tensor<1x960x1x1xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc378) + %384 = "onnx.Conv"(%383, %47, %48) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_329", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x12x20xf32>, tensor<160x960x1x1xf32>, tensor<160xf32>) -> tensor<1x160x12x20xf32> loc(#loc379) + %385 = "onnx.Add"(%384, %365) {onnx_node_name = "Add_330"} : (tensor<1x160x12x20xf32>, tensor<1x160x12x20xf32>) -> tensor<1x160x12x20xf32> loc(#loc380) + %386 = "onnx.Conv"(%385, %49, %50) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_331", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x160x12x20xf32>, tensor<960x160x1x1xf32>, tensor<960xf32>) -> tensor<1x960x12x20xf32> loc(#loc381) + %387 = "onnx.Add"(%386, %71) {onnx_node_name = "Add_333"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc382) + %388 = "onnx.Clip"(%387, %69, %70) {onnx_node_name = "Clip_336_53"} : (tensor<1x960x12x20xf32>, tensor, tensor) -> tensor<1x960x12x20xf32> loc(#loc383) + %389 = "onnx.Div"(%388, %70) {onnx_node_name = "Div_338"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc384) + %390 = "onnx.Mul"(%386, %389) {onnx_node_name = "Mul_339"} : (tensor<1x960x12x20xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc385) + %391 = "onnx.ReduceMeanV13"(%390) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_342_14"} : (tensor<1x960x12x20xf32>) -> tensor<1x960x1x1xf32> loc(#loc386) + %392 = "onnx.Conv"(%391, %124, %6) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_343", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x1x1xf32>, tensor<128x960x1x1xf32>, none) -> tensor<1x128x1x1xf32> loc(#loc387) + %393 = "onnx.Sigmoid"(%392) {onnx_node_name = "Sigmoid_344"} : (tensor<1x128x1x1xf32>) -> tensor<1x128x1x1xf32> loc(#loc388) + %394 = "onnx.Conv"(%390, %51, %52) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_340", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x12x20xf32>, tensor<128x960x1x1xf32>, tensor<128xf32>) -> tensor<1x128x12x20xf32> loc(#loc389) + %395 = "onnx.Relu"(%394) {onnx_node_name = "Relu_341"} : (tensor<1x128x12x20xf32>) -> tensor<1x128x12x20xf32> loc(#loc390) + %396 = "onnx.Mul"(%395, %393) {onnx_node_name = "Mul_345"} : (tensor<1x128x12x20xf32>, tensor<1x128x1x1xf32>) -> tensor<1x128x12x20xf32> loc(#loc391) + %397:2 = "onnx.Split"(%396, %5) {axis = 1 : si64, onnx_node_name = "Split_349_17"} : (tensor<1x128x12x20xf32>, tensor<2xi64>) -> (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) loc(#loc392) + %398 = "onnx.Concat"(%397#1, %arg4) {axis = 1 : si64, onnx_node_name = "Concat_350"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x128x12x20xf32> loc(#loc393) + %399 = "onnx.Conv"(%398, %172, %171) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_351", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x128x12x20xf32>, tensor<128x128x3x3xf32>, tensor<128xf32>) -> tensor<1x128x12x20xf32> loc(#loc394) + %400 = "onnx.Sigmoid"(%399) {onnx_node_name = "Sigmoid_352"} : (tensor<1x128x12x20xf32>) -> tensor<1x128x12x20xf32> loc(#loc395) + %401:2 = "onnx.Split"(%400, %5) {axis = 1 : si64, onnx_node_name = "Split_353_38"} : (tensor<1x128x12x20xf32>, tensor<2xi64>) -> (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) loc(#loc396) + %402 = "onnx.Mul"(%401#0, %arg4) {onnx_node_name = "Mul_354"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc397) + %403 = "onnx.Concat"(%397#1, %402) {axis = 1 : si64, onnx_node_name = "Concat_355"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x128x12x20xf32> loc(#loc398) + %404 = "onnx.Conv"(%403, %170, %169) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_356", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x128x12x20xf32>, tensor<64x128x3x3xf32>, tensor<64xf32>) -> tensor<1x64x12x20xf32> loc(#loc399) + %405 = "onnx.Tanh"(%404) {onnx_node_name = "Tanh_357"} : (tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc400) + %406 = "onnx.Mul"(%401#1, %405) {onnx_node_name = "Mul_361"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc401) + %407 = "onnx.Sub"(%75, %401#1) {onnx_node_name = "Sub_359"} : (tensor, tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc402) + %408 = "onnx.Mul"(%407, %arg4) {onnx_node_name = "Mul_360"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc403) + %409 = "onnx.Add"(%408, %406) {onnx_node_name = "Add_362"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc404) + %410 = "onnx.Concat"(%397#0, %409) {axis = 1 : si64, onnx_node_name = "Concat_363"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x128x12x20xf32> loc(#loc405) + %411 = "onnx.Resize"(%410, %6, %64, %6) { + antialias = 0 : si64, + coordinate_transformation_mode = "pytorch_half_pixel", + cubic_coeff_a = -7.500000e-01 : f32, + exclude_outside = 0 : si64, + extrapolation_value = 0.000000e+00 : f32, + keep_aspect_ratio_policy = "stretch", + mode = "linear", + nearest_mode = "floor", + onnx_node_name = "Resize_365_5"} : (tensor<1x128x12x20xf32>, none, tensor<4xf32>, none) -> tensor<1x128x24x40xf32> loc(#loc406) + %412 = "onnx.Slice"(%411, %67, %72, %63, %74) {onnx_node_name = "Slice_371"} : (tensor<1x128x24x40xf32>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>) -> tensor<1x128x23x40xf32> loc(#loc407) + %413 = "onnx.AveragePool"(%177) { + auto_pad = "NOTSET", + ceil_mode = 1 : si64, + count_include_pad = 0 : si64, + kernel_shape = [2, 2], + onnx_node_name = "AveragePool_346", + pads = [0, 0, 0, 0], + strides = [2, 2]} : (tensor<1x3x180x320xf32>) -> tensor<1x3x90x160xf32> loc(#loc408) + %414 = "onnx.AveragePool"(%413) { + auto_pad = "NOTSET", + ceil_mode = 1 : si64, + count_include_pad = 0 : si64, + kernel_shape = [2, 2], + onnx_node_name = "AveragePool_347", + pads = [0, 0, 0, 0], + strides = [2, 2]} : (tensor<1x3x90x160xf32>) -> tensor<1x3x45x80xf32> loc(#loc409) + %415 = "onnx.AveragePool"(%414) { + auto_pad = "NOTSET", + ceil_mode = 1 : si64, + count_include_pad = 0 : si64, + kernel_shape = [2, 2], + onnx_node_name = "AveragePool_348", + pads = [0, 0, 0, 0], + strides = [2, 2]} : (tensor<1x3x45x80xf32>) -> tensor<1x3x23x40xf32> loc(#loc410) + %416 = "onnx.Concat"(%412, %240, %415) {axis = 1 : si64, onnx_node_name = "Concat_372"} : (tensor<1x128x23x40xf32>, tensor<1x40x23x40xf32>, tensor<1x3x23x40xf32>) -> tensor<1x171x23x40xf32> loc(#loc411) + %417 = "onnx.Conv"(%416, %53, %54) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_373", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x171x23x40xf32>, tensor<80x171x3x3xf32>, tensor<80xf32>) -> tensor<1x80x23x40xf32> loc(#loc412) + %418 = "onnx.Relu"(%417) {onnx_node_name = "Relu_374"} : (tensor<1x80x23x40xf32>) -> tensor<1x80x23x40xf32> loc(#loc413) + %419:2 = "onnx.Split"(%418, %4) {axis = 1 : si64, onnx_node_name = "Split_375_52"} : (tensor<1x80x23x40xf32>, tensor<2xi64>) -> (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) loc(#loc414) + %420 = "onnx.Concat"(%419#1, %arg3) {axis = 1 : si64, onnx_node_name = "Concat_376"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x80x23x40xf32> loc(#loc415) + %421 = "onnx.Conv"(%420, %168, %167) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_377", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x80x23x40xf32>, tensor<80x80x3x3xf32>, tensor<80xf32>) -> tensor<1x80x23x40xf32> loc(#loc416) + %422 = "onnx.Sigmoid"(%421) {onnx_node_name = "Sigmoid_378"} : (tensor<1x80x23x40xf32>) -> tensor<1x80x23x40xf32> loc(#loc417) + %423:2 = "onnx.Split"(%422, %4) {axis = 1 : si64, onnx_node_name = "Split_379_31"} : (tensor<1x80x23x40xf32>, tensor<2xi64>) -> (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) loc(#loc418) + %424 = "onnx.Mul"(%423#0, %arg3) {onnx_node_name = "Mul_380"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc419) + %425 = "onnx.Concat"(%419#1, %424) {axis = 1 : si64, onnx_node_name = "Concat_381"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x80x23x40xf32> loc(#loc420) + %426 = "onnx.Conv"(%425, %166, %165) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_382", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x80x23x40xf32>, tensor<40x80x3x3xf32>, tensor<40xf32>) -> tensor<1x40x23x40xf32> loc(#loc421) + %427 = "onnx.Tanh"(%426) {onnx_node_name = "Tanh_383"} : (tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc422) + %428 = "onnx.Mul"(%423#1, %427) {onnx_node_name = "Mul_387"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc423) + %429 = "onnx.Sub"(%75, %423#1) {onnx_node_name = "Sub_385"} : (tensor, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc424) + %430 = "onnx.Mul"(%429, %arg3) {onnx_node_name = "Mul_386"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc425) + %431 = "onnx.Add"(%430, %428) {onnx_node_name = "Add_388"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc426) + %432 = "onnx.Concat"(%419#0, %431) {axis = 1 : si64, onnx_node_name = "Concat_389"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x80x23x40xf32> loc(#loc427) + %433 = "onnx.Resize"(%432, %6, %64, %6) { + antialias = 0 : si64, + coordinate_transformation_mode = "pytorch_half_pixel", + cubic_coeff_a = -7.500000e-01 : f32, + exclude_outside = 0 : si64, + extrapolation_value = 0.000000e+00 : f32, + keep_aspect_ratio_policy = "stretch", + mode = "linear", + nearest_mode = "floor", + onnx_node_name = "Resize_391_10"} : (tensor<1x80x23x40xf32>, none, tensor<4xf32>, none) -> tensor<1x80x46x80xf32> loc(#loc428) + %434 = "onnx.Slice"(%433, %67, %73, %63, %74) {onnx_node_name = "Slice_397"} : (tensor<1x80x46x80xf32>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>) -> tensor<1x80x45x80xf32> loc(#loc429) + %435 = "onnx.Concat"(%434, %199, %414) {axis = 1 : si64, onnx_node_name = "Concat_398"} : (tensor<1x80x45x80xf32>, tensor<1x24x45x80xf32>, tensor<1x3x45x80xf32>) -> tensor<1x107x45x80xf32> loc(#loc430) + %436 = "onnx.Conv"(%435, %55, %56) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_399", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x107x45x80xf32>, tensor<40x107x3x3xf32>, tensor<40xf32>) -> tensor<1x40x45x80xf32> loc(#loc431) + %437 = "onnx.Relu"(%436) {onnx_node_name = "Relu_400"} : (tensor<1x40x45x80xf32>) -> tensor<1x40x45x80xf32> loc(#loc432) + %438:2 = "onnx.Split"(%437, %3) {axis = 1 : si64, onnx_node_name = "Split_401_41"} : (tensor<1x40x45x80xf32>, tensor<2xi64>) -> (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) loc(#loc433) + %439 = "onnx.Concat"(%438#1, %arg2) {axis = 1 : si64, onnx_node_name = "Concat_402"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x40x45x80xf32> loc(#loc434) + %440 = "onnx.Conv"(%439, %164, %163) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_403", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x40x45x80xf32>, tensor<40x40x3x3xf32>, tensor<40xf32>) -> tensor<1x40x45x80xf32> loc(#loc435) + %441 = "onnx.Sigmoid"(%440) {onnx_node_name = "Sigmoid_404"} : (tensor<1x40x45x80xf32>) -> tensor<1x40x45x80xf32> loc(#loc436) + %442:2 = "onnx.Split"(%441, %3) {axis = 1 : si64, onnx_node_name = "Split_405_6"} : (tensor<1x40x45x80xf32>, tensor<2xi64>) -> (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) loc(#loc437) + %443 = "onnx.Mul"(%442#0, %arg2) {onnx_node_name = "Mul_406"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc438) + %444 = "onnx.Concat"(%438#1, %443) {axis = 1 : si64, onnx_node_name = "Concat_407"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x40x45x80xf32> loc(#loc439) + %445 = "onnx.Conv"(%444, %162, %161) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_408", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x40x45x80xf32>, tensor<20x40x3x3xf32>, tensor<20xf32>) -> tensor<1x20x45x80xf32> loc(#loc440) + %446 = "onnx.Tanh"(%445) {onnx_node_name = "Tanh_409"} : (tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc441) + %447 = "onnx.Mul"(%442#1, %446) {onnx_node_name = "Mul_413"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc442) + %448 = "onnx.Sub"(%75, %442#1) {onnx_node_name = "Sub_411"} : (tensor, tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc443) + %449 = "onnx.Mul"(%448, %arg2) {onnx_node_name = "Mul_412"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc444) + %450 = "onnx.Add"(%449, %447) {onnx_node_name = "Add_414"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc445) + %451 = "onnx.Concat"(%438#0, %450) {axis = 1 : si64, onnx_node_name = "Concat_415"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x40x45x80xf32> loc(#loc446) + %452 = "onnx.Resize"(%451, %6, %64, %6) { + antialias = 0 : si64, + coordinate_transformation_mode = "pytorch_half_pixel", + cubic_coeff_a = -7.500000e-01 : f32, + exclude_outside = 0 : si64, + extrapolation_value = 0.000000e+00 : f32, + keep_aspect_ratio_policy = "stretch", + mode = "linear", + nearest_mode = "floor", + onnx_node_name = "Resize_417_15"} : (tensor<1x40x45x80xf32>, none, tensor<4xf32>, none) -> tensor<1x40x90x160xf32> loc(#loc447) + %453 = "onnx.Concat"(%452, %188, %413) {axis = 1 : si64, onnx_node_name = "Concat_418"} : (tensor<1x40x90x160xf32>, tensor<1x16x90x160xf32>, tensor<1x3x90x160xf32>) -> tensor<1x59x90x160xf32> loc(#loc448) + %454 = "onnx.Conv"(%453, %57, %58) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_419", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x59x90x160xf32>, tensor<32x59x3x3xf32>, tensor<32xf32>) -> tensor<1x32x90x160xf32> loc(#loc449) + %455 = "onnx.Relu"(%454) {onnx_node_name = "Relu_420"} : (tensor<1x32x90x160xf32>) -> tensor<1x32x90x160xf32> loc(#loc450) + %456:2 = "onnx.Split"(%455, %2) {axis = 1 : si64, onnx_node_name = "Split_421_16"} : (tensor<1x32x90x160xf32>, tensor<2xi64>) -> (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) loc(#loc451) + %457 = "onnx.Concat"(%456#1, %arg1) {axis = 1 : si64, onnx_node_name = "Concat_422"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x32x90x160xf32> loc(#loc452) + %458 = "onnx.Conv"(%457, %160, %159) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_423", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x32x90x160xf32>, tensor<32x32x3x3xf32>, tensor<32xf32>) -> tensor<1x32x90x160xf32> loc(#loc453) + %459 = "onnx.Sigmoid"(%458) {onnx_node_name = "Sigmoid_424"} : (tensor<1x32x90x160xf32>) -> tensor<1x32x90x160xf32> loc(#loc454) + %460:2 = "onnx.Split"(%459, %2) {axis = 1 : si64, onnx_node_name = "Split_425_23"} : (tensor<1x32x90x160xf32>, tensor<2xi64>) -> (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) loc(#loc455) + %461 = "onnx.Mul"(%460#0, %arg1) {onnx_node_name = "Mul_426"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc456) + %462 = "onnx.Concat"(%456#1, %461) {axis = 1 : si64, onnx_node_name = "Concat_427"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x32x90x160xf32> loc(#loc457) + %463 = "onnx.Conv"(%462, %158, %157) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_428", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x32x90x160xf32>, tensor<16x32x3x3xf32>, tensor<16xf32>) -> tensor<1x16x90x160xf32> loc(#loc458) + %464 = "onnx.Tanh"(%463) {onnx_node_name = "Tanh_429"} : (tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc459) + %465 = "onnx.Mul"(%460#1, %464) {onnx_node_name = "Mul_433"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc460) + %466 = "onnx.Sub"(%75, %460#1) {onnx_node_name = "Sub_431"} : (tensor, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc461) + %467 = "onnx.Mul"(%466, %arg1) {onnx_node_name = "Mul_432"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc462) + %468 = "onnx.Add"(%467, %465) {onnx_node_name = "Add_434"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc463) + %469 = "onnx.Concat"(%456#0, %468) {axis = 1 : si64, onnx_node_name = "Concat_435"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x32x90x160xf32> loc(#loc464) + %470 = "onnx.Resize"(%469, %6, %64, %6) { + antialias = 0 : si64, + coordinate_transformation_mode = "pytorch_half_pixel", + cubic_coeff_a = -7.500000e-01 : f32, + exclude_outside = 0 : si64, + extrapolation_value = 0.000000e+00 : f32, + keep_aspect_ratio_policy = "stretch", + mode = "linear", + nearest_mode = "floor", + onnx_node_name = "Resize_437_11"} : (tensor<1x32x90x160xf32>, none, tensor<4xf32>, none) -> tensor<1x32x180x320xf32> loc(#loc465) + %471 = "onnx.Concat"(%470, %177) {axis = 1 : si64, onnx_node_name = "Concat_438"} : (tensor<1x32x180x320xf32>, tensor<1x3x180x320xf32>) -> tensor<1x35x180x320xf32> loc(#loc466) + %472 = "onnx.Conv"(%471, %59, %60) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_439", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x35x180x320xf32>, tensor<16x35x3x3xf32>, tensor<16xf32>) -> tensor<1x16x180x320xf32> loc(#loc467) + %473 = "onnx.Relu"(%472) {onnx_node_name = "Relu_440"} : (tensor<1x16x180x320xf32>) -> tensor<1x16x180x320xf32> loc(#loc468) + %474 = "onnx.Conv"(%473, %61, %62) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_441", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x16x180x320xf32>, tensor<16x16x3x3xf32>, tensor<16xf32>) -> tensor<1x16x180x320xf32> loc(#loc469) + %475 = "onnx.Relu"(%474) {onnx_node_name = "Relu_442"} : (tensor<1x16x180x320xf32>) -> tensor<1x16x180x320xf32> loc(#loc470) + %476 = "onnx.Conv"(%475, %174, %173) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_443", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x16x180x320xf32>, tensor<4x16x1x1xf32>, tensor<4xf32>) -> tensor<1x4x180x320xf32> loc(#loc471) + %477:2 = "onnx.Split"(%476, %1) {axis = 1 : si64, onnx_node_name = "Split_444_34"} : (tensor<1x4x180x320xf32>, tensor<2xi64>) -> (tensor<1x3x180x320xf32>, tensor<1x1x180x320xf32>) loc(#loc472) + %478 = "onnx.Clip"(%477#1, %69, %75) {onnx_node_name = "Clip_447_13"} : (tensor<1x1x180x320xf32>, tensor, tensor) -> tensor<1x1x180x320xf32> loc(#loc473) + %479 = "onnx.Add"(%477#0, %177) {onnx_node_name = "Add_445"} : (tensor<1x3x180x320xf32>, tensor<1x3x180x320xf32>) -> tensor<1x3x180x320xf32> loc(#loc474) + %480 = "onnx.Clip"(%479, %69, %75) {onnx_node_name = "Clip_446_1"} : (tensor<1x3x180x320xf32>, tensor, tensor) -> tensor<1x3x180x320xf32> loc(#loc475) + return %478, %468, %450, %431, %409, %480 : tensor<1x1x180x320xf32>, tensor<1x16x90x160xf32>, tensor<1x20x45x80xf32>, tensor<1x40x23x40xf32>, tensor<1x64x12x20xf32>, tensor<1x3x180x320xf32> loc(#loc) + } loc(#loc) + "onnx.EntryPoint"() {func = @main_graph} : () -> () loc(#loc) +} loc(#loc) +#loc1 = loc("Initializer_1001") +#loc2 = loc("Initializer_1002") +#loc3 = loc("Initializer_1004") +#loc4 = loc("Initializer_1005") +#loc5 = loc("Initializer_1007") +#loc6 = loc("Initializer_1008") +#loc7 = loc("Initializer_1010") +#loc8 = loc("Initializer_1011") +#loc9 = loc("Initializer_1013") +#loc10 = loc("Initializer_1014") +#loc11 = loc("Initializer_1016") +#loc12 = loc("Initializer_1017") +#loc13 = loc("Initializer_1019") +#loc14 = loc("Initializer_1020") +#loc15 = loc("Initializer_1022") +#loc16 = loc("Initializer_1023") +#loc17 = loc("Initializer_1025") +#loc18 = loc("Initializer_1026") +#loc19 = loc("Initializer_1028") +#loc20 = loc("Initializer_1029") +#loc21 = loc("Initializer_1031") +#loc22 = loc("Initializer_1032") +#loc23 = loc("Initializer_1034") +#loc24 = loc("Initializer_1035") +#loc25 = loc("Initializer_1037") +#loc26 = loc("Initializer_1038") +#loc27 = loc("Initializer_1040") +#loc28 = loc("Initializer_1041") +#loc29 = loc("Initializer_1043") +#loc30 = loc("Initializer_1044") +#loc31 = loc("Initializer_1046") +#loc32 = loc("Initializer_1047") +#loc33 = loc("Initializer_1049") +#loc34 = loc("Initializer_1050") +#loc35 = loc("Initializer_1052") +#loc36 = loc("Initializer_1053") +#loc37 = loc("Initializer_1055") +#loc38 = loc("Initializer_1056") +#loc39 = loc("Initializer_1058") +#loc40 = loc("Initializer_1059") +#loc41 = loc("Initializer_1061") +#loc42 = loc("Initializer_1062") +#loc43 = loc("Initializer_1064") +#loc44 = loc("Initializer_1065") +#loc45 = loc("Initializer_1067") +#loc46 = loc("Initializer_1068") +#loc47 = loc("Initializer_1070") +#loc48 = loc("Initializer_1071") +#loc49 = loc("Initializer_1073") +#loc50 = loc("Initializer_1074") +#loc51 = loc("Initializer_1076") +#loc52 = loc("Initializer_1077") +#loc53 = loc("Initializer_1079") +#loc54 = loc("Initializer_1080") +#loc55 = loc("Initializer_1082") +#loc56 = loc("Initializer_1083") +#loc57 = loc("Initializer_1086") +#loc58 = loc("Initializer_1090") +#loc59 = loc("Initializer_386") +#loc60 = loc("Initializer_388") +#loc61 = loc("Initializer_389") +#loc62 = loc("Initializer_400") +#loc63 = loc("Initializer_752") +#loc64 = loc("Initializer_755") +#loc65 = loc("Initializer_763") +#loc66 = loc("Initializer_809") +#loc67 = loc("Initializer_845") +#loc68 = loc("Initializer_847") +#loc69 = loc("Initializer_890") +#loc70 = loc("Initializer_929") +#loc71 = loc("Initializer_930") +#loc72 = loc("Initializer_932") +#loc73 = loc("Initializer_933") +#loc74 = loc("Initializer_935") +#loc75 = loc("Initializer_936") +#loc76 = loc("Initializer_938") +#loc77 = loc("Initializer_939") +#loc78 = loc("Initializer_941") +#loc79 = loc("Initializer_942") +#loc80 = loc("Initializer_944") +#loc81 = loc("Initializer_945") +#loc82 = loc("Initializer_947") +#loc83 = loc("Initializer_948") +#loc84 = loc("Initializer_950") +#loc85 = loc("Initializer_951") +#loc86 = loc("Initializer_953") +#loc87 = loc("Initializer_954") +#loc88 = loc("Initializer_956") +#loc89 = loc("Initializer_957") +#loc90 = loc("Initializer_959") +#loc91 = loc("Initializer_960") +#loc92 = loc("Initializer_962") +#loc93 = loc("Initializer_963") +#loc94 = loc("Initializer_965") +#loc95 = loc("Initializer_966") +#loc96 = loc("Initializer_968") +#loc97 = loc("Initializer_969") +#loc98 = loc("Initializer_971") +#loc99 = loc("Initializer_972") +#loc100 = loc("Initializer_974") +#loc101 = loc("Initializer_975") +#loc102 = loc("Initializer_977") +#loc103 = loc("Initializer_978") +#loc104 = loc("Initializer_980") +#loc105 = loc("Initializer_981") +#loc106 = loc("Initializer_983") +#loc107 = loc("Initializer_984") +#loc108 = loc("Initializer_986") +#loc109 = loc("Initializer_987") +#loc110 = loc("Initializer_989") +#loc111 = loc("Initializer_990") +#loc112 = loc("Initializer_992") +#loc113 = loc("Initializer_993") +#loc114 = loc("Initializer_995") +#loc115 = loc("Initializer_996") +#loc116 = loc("Initializer_998") +#loc117 = loc("Initializer_999") +#loc118 = loc("Initializer_aspp.aspp2.1.weight") +#loc119 = loc("Initializer_backbone.features.11.block.2.fc1.bias") +#loc120 = loc("Initializer_backbone.features.11.block.2.fc1.weight") +#loc121 = loc("Initializer_backbone.features.11.block.2.fc2.bias") +#loc122 = loc("Initializer_backbone.features.11.block.2.fc2.weight") +#loc123 = loc("Initializer_backbone.features.12.block.2.fc1.bias") +#loc124 = loc("Initializer_backbone.features.12.block.2.fc1.weight") +#loc125 = loc("Initializer_backbone.features.12.block.2.fc2.bias") +#loc126 = loc("Initializer_backbone.features.12.block.2.fc2.weight") +#loc127 = loc("Initializer_backbone.features.13.block.2.fc1.bias") +#loc128 = loc("Initializer_backbone.features.13.block.2.fc1.weight") +#loc129 = loc("Initializer_backbone.features.13.block.2.fc2.bias") +#loc130 = loc("Initializer_backbone.features.13.block.2.fc2.weight") +#loc131 = loc("Initializer_backbone.features.14.block.2.fc1.bias") +#loc132 = loc("Initializer_backbone.features.14.block.2.fc1.weight") +#loc133 = loc("Initializer_backbone.features.14.block.2.fc2.bias") +#loc134 = loc("Initializer_backbone.features.14.block.2.fc2.weight") +#loc135 = loc("Initializer_backbone.features.15.block.2.fc1.bias") +#loc136 = loc("Initializer_backbone.features.15.block.2.fc1.weight") +#loc137 = loc("Initializer_backbone.features.15.block.2.fc2.bias") +#loc138 = loc("Initializer_backbone.features.15.block.2.fc2.weight") +#loc139 = loc("Initializer_backbone.features.4.block.2.fc1.bias") +#loc140 = loc("Initializer_backbone.features.4.block.2.fc1.weight") +#loc141 = loc("Initializer_backbone.features.4.block.2.fc2.bias") +#loc142 = loc("Initializer_backbone.features.4.block.2.fc2.weight") +#loc143 = loc("Initializer_backbone.features.5.block.2.fc1.bias") +#loc144 = loc("Initializer_backbone.features.5.block.2.fc1.weight") +#loc145 = loc("Initializer_backbone.features.5.block.2.fc2.bias") +#loc146 = loc("Initializer_backbone.features.5.block.2.fc2.weight") +#loc147 = loc("Initializer_backbone.features.6.block.2.fc1.bias") +#loc148 = loc("Initializer_backbone.features.6.block.2.fc1.weight") +#loc149 = loc("Initializer_backbone.features.6.block.2.fc2.bias") +#loc150 = loc("Initializer_backbone.features.6.block.2.fc2.weight") +#loc151 = loc("Initializer_decoder.decode1.gru.hh.0.bias") +#loc152 = loc("Initializer_decoder.decode1.gru.hh.0.weight") +#loc153 = loc("Initializer_decoder.decode1.gru.ih.0.bias") +#loc154 = loc("Initializer_decoder.decode1.gru.ih.0.weight") +#loc155 = loc("Initializer_decoder.decode2.gru.hh.0.bias") +#loc156 = loc("Initializer_decoder.decode2.gru.hh.0.weight") +#loc157 = loc("Initializer_decoder.decode2.gru.ih.0.bias") +#loc158 = loc("Initializer_decoder.decode2.gru.ih.0.weight") +#loc159 = loc("Initializer_decoder.decode3.gru.hh.0.bias") +#loc160 = loc("Initializer_decoder.decode3.gru.hh.0.weight") +#loc161 = loc("Initializer_decoder.decode3.gru.ih.0.bias") +#loc162 = loc("Initializer_decoder.decode3.gru.ih.0.weight") +#loc163 = loc("Initializer_decoder.decode4.gru.hh.0.bias") +#loc164 = loc("Initializer_decoder.decode4.gru.hh.0.weight") +#loc165 = loc("Initializer_decoder.decode4.gru.ih.0.bias") +#loc166 = loc("Initializer_decoder.decode4.gru.ih.0.weight") +#loc167 = loc("Initializer_project_mat.conv.bias") +#loc168 = loc("Initializer_project_mat.conv.weight") +#loc169 = loc("Div_2") +#loc170 = loc("Slice_7") +#loc171 = loc("Transpose_8") +#loc172 = loc("Sub_14") +#loc173 = loc("Initializer_398") +#loc174 = loc("Div_16") +#loc175 = loc("Conv_17") +#loc176 = loc("Add_19") +#loc177 = loc("Clip_22") +#loc178 = loc("Div_24") +#loc179 = loc("Mul_25") +#loc180 = loc("Conv_26") +#loc181 = loc("Relu_27") +#loc182 = loc("Conv_28") +#loc183 = loc("Add_29") +#loc184 = loc("Conv_30") +#loc185 = loc("Relu_31") +#loc186 = loc("Conv_32") +#loc187 = loc("Relu_33") +#loc188 = loc("Conv_34") +#loc189 = loc("Conv_35") +#loc190 = loc("Relu_36") +#loc191 = loc("Conv_37") +#loc192 = loc("Relu_38") +#loc193 = loc("Conv_39") +#loc194 = loc("Add_40") +#loc195 = loc("Conv_41") +#loc196 = loc("Relu_42") +#loc197 = loc("Conv_43") +#loc198 = loc("Relu_44") +#loc199 = loc("GlobalAveragePool_45") +#loc200 = loc("Conv_46") +#loc201 = loc("Relu_47") +#loc202 = loc("Conv_48") +#loc203 = loc("Add_50") +#loc204 = loc("Clip_53") +#loc205 = loc("Div_55") +#loc206 = loc("Mul_56") +#loc207 = loc("Conv_57") +#loc208 = loc("Conv_58") +#loc209 = loc("Relu_59") +#loc210 = loc("Conv_60") +#loc211 = loc("Relu_61") +#loc212 = loc("GlobalAveragePool_62") +#loc213 = loc("Conv_63") +#loc214 = loc("Relu_64") +#loc215 = loc("Conv_65") +#loc216 = loc("Add_67") +#loc217 = loc("Clip_70") +#loc218 = loc("Div_72") +#loc219 = loc("Mul_73") +#loc220 = loc("Conv_74") +#loc221 = loc("Add_75") +#loc222 = loc("Conv_76") +#loc223 = loc("Relu_77") +#loc224 = loc("Conv_78") +#loc225 = loc("Relu_79") +#loc226 = loc("GlobalAveragePool_80") +#loc227 = loc("Conv_81") +#loc228 = loc("Relu_82") +#loc229 = loc("Conv_83") +#loc230 = loc("Add_85") +#loc231 = loc("Clip_88") +#loc232 = loc("Div_90") +#loc233 = loc("Mul_91") +#loc234 = loc("Conv_92") +#loc235 = loc("Add_93") +#loc236 = loc("Conv_94") +#loc237 = loc("Add_96") +#loc238 = loc("Clip_99") +#loc239 = loc("Div_101") +#loc240 = loc("Mul_102") +#loc241 = loc("Conv_103") +#loc242 = loc("Add_105") +#loc243 = loc("Clip_108") +#loc244 = loc("Div_110") +#loc245 = loc("Mul_111") +#loc246 = loc("Conv_112") +#loc247 = loc("Conv_113") +#loc248 = loc("Add_115") +#loc249 = loc("Clip_118") +#loc250 = loc("Div_120") +#loc251 = loc("Mul_121") +#loc252 = loc("Conv_122") +#loc253 = loc("Add_124") +#loc254 = loc("Clip_127") +#loc255 = loc("Div_129") +#loc256 = loc("Mul_130") +#loc257 = loc("Conv_131") +#loc258 = loc("Add_132") +#loc259 = loc("Conv_133") +#loc260 = loc("Add_135") +#loc261 = loc("Clip_138") +#loc262 = loc("Div_140") +#loc263 = loc("Mul_141") +#loc264 = loc("Conv_142") +#loc265 = loc("Add_144") +#loc266 = loc("Clip_147") +#loc267 = loc("Div_149") +#loc268 = loc("Mul_150") +#loc269 = loc("Conv_151") +#loc270 = loc("Add_152") +#loc271 = loc("Conv_153") +#loc272 = loc("Add_155") +#loc273 = loc("Clip_158") +#loc274 = loc("Div_160") +#loc275 = loc("Mul_161") +#loc276 = loc("Conv_162") +#loc277 = loc("Add_164") +#loc278 = loc("Clip_167") +#loc279 = loc("Div_169") +#loc280 = loc("Mul_170") +#loc281 = loc("Conv_171") +#loc282 = loc("Add_172") +#loc283 = loc("Conv_173") +#loc284 = loc("Add_175") +#loc285 = loc("Clip_178") +#loc286 = loc("Div_180") +#loc287 = loc("Mul_181") +#loc288 = loc("Conv_182") +#loc289 = loc("Add_184") +#loc290 = loc("Clip_187") +#loc291 = loc("Div_189") +#loc292 = loc("Mul_190") +#loc293 = loc("GlobalAveragePool_191") +#loc294 = loc("Conv_192") +#loc295 = loc("Relu_193") +#loc296 = loc("Conv_194") +#loc297 = loc("Add_196") +#loc298 = loc("Clip_199") +#loc299 = loc("Div_201") +#loc300 = loc("Mul_202") +#loc301 = loc("Conv_203") +#loc302 = loc("Conv_204") +#loc303 = loc("Add_206") +#loc304 = loc("Clip_209") +#loc305 = loc("Div_211") +#loc306 = loc("Mul_212") +#loc307 = loc("Conv_213") +#loc308 = loc("Add_215") +#loc309 = loc("Clip_218") +#loc310 = loc("Div_220") +#loc311 = loc("Mul_221") +#loc312 = loc("GlobalAveragePool_222") +#loc313 = loc("Conv_223") +#loc314 = loc("Relu_224") +#loc315 = loc("Conv_225") +#loc316 = loc("Add_227") +#loc317 = loc("Clip_230") +#loc318 = loc("Div_232") +#loc319 = loc("Mul_233") +#loc320 = loc("Conv_234") +#loc321 = loc("Add_235") +#loc322 = loc("Conv_236") +#loc323 = loc("Add_238") +#loc324 = loc("Clip_241") +#loc325 = loc("Div_243") +#loc326 = loc("Mul_244") +#loc327 = loc("Conv_245") +#loc328 = loc("Add_247") +#loc329 = loc("Clip_250") +#loc330 = loc("Div_252") +#loc331 = loc("Mul_253") +#loc332 = loc("GlobalAveragePool_254") +#loc333 = loc("Conv_255") +#loc334 = loc("Relu_256") +#loc335 = loc("Conv_257") +#loc336 = loc("Add_259") +#loc337 = loc("Clip_262") +#loc338 = loc("Div_264") +#loc339 = loc("Mul_265") +#loc340 = loc("Conv_266") +#loc341 = loc("Conv_267") +#loc342 = loc("Add_269") +#loc343 = loc("Clip_272") +#loc344 = loc("Div_274") +#loc345 = loc("Mul_275") +#loc346 = loc("Conv_276") +#loc347 = loc("Add_278") +#loc348 = loc("Clip_281") +#loc349 = loc("Div_283") +#loc350 = loc("Mul_284") +#loc351 = loc("GlobalAveragePool_285") +#loc352 = loc("Conv_286") +#loc353 = loc("Relu_287") +#loc354 = loc("Conv_288") +#loc355 = loc("Add_290") +#loc356 = loc("Clip_293") +#loc357 = loc("Div_295") +#loc358 = loc("Mul_296") +#loc359 = loc("Conv_297") +#loc360 = loc("Add_298") +#loc361 = loc("Conv_299") +#loc362 = loc("Add_301") +#loc363 = loc("Clip_304") +#loc364 = loc("Div_306") +#loc365 = loc("Mul_307") +#loc366 = loc("Conv_308") +#loc367 = loc("Add_310") +#loc368 = loc("Clip_313") +#loc369 = loc("Div_315") +#loc370 = loc("Mul_316") +#loc371 = loc("GlobalAveragePool_317") +#loc372 = loc("Conv_318") +#loc373 = loc("Relu_319") +#loc374 = loc("Conv_320") +#loc375 = loc("Add_322") +#loc376 = loc("Clip_325") +#loc377 = loc("Div_327") +#loc378 = loc("Mul_328") +#loc379 = loc("Conv_329") +#loc380 = loc("Add_330") +#loc381 = loc("Conv_331") +#loc382 = loc("Add_333") +#loc383 = loc("Clip_336") +#loc384 = loc("Div_338") +#loc385 = loc("Mul_339") +#loc386 = loc("GlobalAveragePool_342") +#loc387 = loc("Conv_343") +#loc388 = loc("Sigmoid_344") +#loc389 = loc("Conv_340") +#loc390 = loc("Relu_341") +#loc391 = loc("Mul_345") +#loc392 = loc("Split_349") +#loc393 = loc("Concat_350") +#loc394 = loc("Conv_351") +#loc395 = loc("Sigmoid_352") +#loc396 = loc("Split_353") +#loc397 = loc("Mul_354") +#loc398 = loc("Concat_355") +#loc399 = loc("Conv_356") +#loc400 = loc("Tanh_357") +#loc401 = loc("Mul_361") +#loc402 = loc("Sub_359") +#loc403 = loc("Mul_360") +#loc404 = loc("Add_362") +#loc405 = loc("Concat_363") +#loc406 = loc("Resize_365") +#loc407 = loc("Slice_371") +#loc408 = loc("AveragePool_346") +#loc409 = loc("AveragePool_347") +#loc410 = loc("AveragePool_348") +#loc411 = loc("Concat_372") +#loc412 = loc("Conv_373") +#loc413 = loc("Relu_374") +#loc414 = loc("Split_375") +#loc415 = loc("Concat_376") +#loc416 = loc("Conv_377") +#loc417 = loc("Sigmoid_378") +#loc418 = loc("Split_379") +#loc419 = loc("Mul_380") +#loc420 = loc("Concat_381") +#loc421 = loc("Conv_382") +#loc422 = loc("Tanh_383") +#loc423 = loc("Mul_387") +#loc424 = loc("Sub_385") +#loc425 = loc("Mul_386") +#loc426 = loc("Add_388") +#loc427 = loc("Concat_389") +#loc428 = loc("Resize_391") +#loc429 = loc("Slice_397") +#loc430 = loc("Concat_398") +#loc431 = loc("Conv_399") +#loc432 = loc("Relu_400") +#loc433 = loc("Split_401") +#loc434 = loc("Concat_402") +#loc435 = loc("Conv_403") +#loc436 = loc("Sigmoid_404") +#loc437 = loc("Split_405") +#loc438 = loc("Mul_406") +#loc439 = loc("Concat_407") +#loc440 = loc("Conv_408") +#loc441 = loc("Tanh_409") +#loc442 = loc("Mul_413") +#loc443 = loc("Sub_411") +#loc444 = loc("Mul_412") +#loc445 = loc("Add_414") +#loc446 = loc("Concat_415") +#loc447 = loc("Resize_417") +#loc448 = loc("Concat_418") +#loc449 = loc("Conv_419") +#loc450 = loc("Relu_420") +#loc451 = loc("Split_421") +#loc452 = loc("Concat_422") +#loc453 = loc("Conv_423") +#loc454 = loc("Sigmoid_424") +#loc455 = loc("Split_425") +#loc456 = loc("Mul_426") +#loc457 = loc("Concat_427") +#loc458 = loc("Conv_428") +#loc459 = loc("Tanh_429") +#loc460 = loc("Mul_433") +#loc461 = loc("Sub_431") +#loc462 = loc("Mul_432") +#loc463 = loc("Add_434") +#loc464 = loc("Concat_435") +#loc465 = loc("Resize_437") +#loc466 = loc("Concat_438") +#loc467 = loc("Conv_439") +#loc468 = loc("Relu_440") +#loc469 = loc("Conv_441") +#loc470 = loc("Relu_442") +#loc471 = loc("Conv_443") +#loc472 = loc("Split_444") +#loc473 = loc("Clip_447") +#loc474 = loc("Add_445") +#loc475 = loc("Clip_446") +#loc476 = loc(fused[#loc172, #loc173]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.par.x86.unwrapped.mlir b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.par.x86.unwrapped.mlir new file mode 100644 index 0000000000000000000000000000000000000000..f102a9ba159ea2b1314bca7b10a9d7ec6085d40d --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.par.x86.unwrapped.mlir @@ -0,0 +1,324 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func private @forward_outlined_part_0(tensor<1x180x320x4xbf16>, tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) attributes {aie_partition = 0 : i32, kernel} loc(#loc308) + func.func @forward(%arg0: tensor<1x180x320x4xbf16> {onnx.name = "385"} loc(unknown), %arg1: tensor<1x16x90x160xbf16> {onnx.name = "394"} loc(unknown), %arg2: tensor<1x20x45x80xbf16> {onnx.name = "395"} loc(unknown), %arg3: tensor<1x40x23x40xbf16> {onnx.name = "396"} loc(unknown), %arg4: tensor<1x64x12x20xbf16> {onnx.name = "397"} loc(unknown)) -> (tensor<1x1x180x320xbf16> {onnx.name = "921"}, tensor<1x16x90x160xbf16> {onnx.name = "894"}, tensor<1x20x45x80xbf16> {onnx.name = "868"}, tensor<1x40x23x40xbf16> {onnx.name = "832"}, tensor<1x64x12x20xbf16> {onnx.name = "796"}, tensor<1x3x180x320xbf16> {onnx.name = "916"}) { + %0:6 = call @forward_outlined_part_0(%arg0, %arg1, %arg2, %arg3, %arg4) : (tensor<1x180x320x4xbf16>, tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) loc(#loc308) + return %0#5, %0#0, %0#1, %0#2, %0#3, %0#4 : tensor<1x1x180x320xbf16>, tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16> loc(#loc) + } loc(#loc) +} loc(#loc) +#loc1 = loc("Div_2") +#loc2 = loc("Sub_431") +#loc3 = loc("Sub_411") +#loc4 = loc("Sub_385") +#loc5 = loc("Sub_359") +#loc6 = loc("Div_16") +#loc7 = loc("Sub_14") +#loc8 = loc("Initializer_398") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("Conv_17") +#loc14 = loc("Add_19") +#loc15 = loc("Clip_22") +#loc16 = loc("Div_24") +#loc17 = loc("Mul_25") +#loc18 = loc("Conv_26") +#loc19 = loc("Relu_27") +#loc20 = loc("Conv_28") +#loc21 = loc("Add_29") +#loc22 = loc("Conv_30") +#loc23 = loc("Relu_31") +#loc24 = loc("Conv_32") +#loc25 = loc("Relu_33") +#loc26 = loc("Conv_34") +#loc27 = loc("Conv_35") +#loc28 = loc("Relu_36") +#loc29 = loc("Conv_37") +#loc30 = loc("Relu_38") +#loc31 = loc("Conv_39") +#loc32 = loc("Add_40") +#loc33 = loc("Conv_41") +#loc34 = loc("Relu_42") +#loc35 = loc("Conv_43") +#loc36 = loc("Relu_44") +#loc37 = loc("GlobalAveragePool_45") +#loc38 = loc("Conv_46") +#loc39 = loc("Relu_47") +#loc40 = loc("Conv_48") +#loc41 = loc("Add_50") +#loc42 = loc("Clip_53") +#loc43 = loc("Div_55") +#loc44 = loc("Mul_56") +#loc45 = loc("Conv_57") +#loc46 = loc("Conv_58") +#loc47 = loc("Relu_59") +#loc48 = loc("Conv_60") +#loc49 = loc("Relu_61") +#loc50 = loc("GlobalAveragePool_62") +#loc51 = loc("Conv_63") +#loc52 = loc("Relu_64") +#loc53 = loc("Conv_65") +#loc54 = loc("Add_67") +#loc55 = loc("Clip_70") +#loc56 = loc("Div_72") +#loc57 = loc("Mul_73") +#loc58 = loc("Conv_74") +#loc59 = loc("Add_75") +#loc60 = loc("Conv_76") +#loc61 = loc("Relu_77") +#loc62 = loc("Conv_78") +#loc63 = loc("Relu_79") +#loc64 = loc("GlobalAveragePool_80") +#loc65 = loc("Conv_81") +#loc66 = loc("Relu_82") +#loc67 = loc("Conv_83") +#loc68 = loc("Add_85") +#loc69 = loc("Clip_88") +#loc70 = loc("Div_90") +#loc71 = loc("Mul_91") +#loc72 = loc("Conv_92") +#loc73 = loc("Add_93") +#loc74 = loc("Conv_94") +#loc75 = loc("Add_96") +#loc76 = loc("Clip_99") +#loc77 = loc("Div_101") +#loc78 = loc("Mul_102") +#loc79 = loc("Conv_103") +#loc80 = loc("Add_105") +#loc81 = loc("Clip_108") +#loc82 = loc("Div_110") +#loc83 = loc("Mul_111") +#loc84 = loc("Conv_112") +#loc85 = loc("Conv_113") +#loc86 = loc("Add_115") +#loc87 = loc("Clip_118") +#loc88 = loc("Div_120") +#loc89 = loc("Mul_121") +#loc90 = loc("Conv_122") +#loc91 = loc("Add_124") +#loc92 = loc("Clip_127") +#loc93 = loc("Div_129") +#loc94 = loc("Mul_130") +#loc95 = loc("Conv_131") +#loc96 = loc("Add_132") +#loc97 = loc("Conv_133") +#loc98 = loc("Add_135") +#loc99 = loc("Clip_138") +#loc100 = loc("Div_140") +#loc101 = loc("Mul_141") +#loc102 = loc("Conv_142") +#loc103 = loc("Add_144") +#loc104 = loc("Clip_147") +#loc105 = loc("Div_149") +#loc106 = loc("Mul_150") +#loc107 = loc("Conv_151") +#loc108 = loc("Add_152") +#loc109 = loc("Conv_153") +#loc110 = loc("Add_155") +#loc111 = loc("Clip_158") +#loc112 = loc("Div_160") +#loc113 = loc("Mul_161") +#loc114 = loc("Conv_162") +#loc115 = loc("Add_164") +#loc116 = loc("Clip_167") +#loc117 = loc("Div_169") +#loc118 = loc("Mul_170") +#loc119 = loc("Conv_171") +#loc120 = loc("Add_172") +#loc121 = loc("Conv_173") +#loc122 = loc("Add_175") +#loc123 = loc("Clip_178") +#loc124 = loc("Div_180") +#loc125 = loc("Mul_181") +#loc126 = loc("Conv_182") +#loc127 = loc("Add_184") +#loc128 = loc("Clip_187") +#loc129 = loc("Div_189") +#loc130 = loc("Mul_190") +#loc131 = loc("GlobalAveragePool_191") +#loc132 = loc("Conv_192") +#loc133 = loc("Relu_193") +#loc134 = loc("Conv_194") +#loc135 = loc("Add_196") +#loc136 = loc("Clip_199") +#loc137 = loc("Div_201") +#loc138 = loc("Mul_202") +#loc139 = loc("Conv_203") +#loc140 = loc("Conv_204") +#loc141 = loc("Add_206") +#loc142 = loc("Clip_209") +#loc143 = loc("Div_211") +#loc144 = loc("Mul_212") +#loc145 = loc("Conv_213") +#loc146 = loc("Add_215") +#loc147 = loc("Clip_218") +#loc148 = loc("Div_220") +#loc149 = loc("Mul_221") +#loc150 = loc("GlobalAveragePool_222") +#loc151 = loc("Conv_223") +#loc152 = loc("Relu_224") +#loc153 = loc("Conv_225") +#loc154 = loc("Add_227") +#loc155 = loc("Clip_230") +#loc156 = loc("Div_232") +#loc157 = loc("Mul_233") +#loc158 = loc("Conv_234") +#loc159 = loc("Add_235") +#loc160 = loc("Conv_236") +#loc161 = loc("Add_238") +#loc162 = loc("Clip_241") +#loc163 = loc("Div_243") +#loc164 = loc("Mul_244") +#loc165 = loc("Conv_245") +#loc166 = loc("Add_247") +#loc167 = loc("Clip_250") +#loc168 = loc("Div_252") +#loc169 = loc("Mul_253") +#loc170 = loc("GlobalAveragePool_254") +#loc171 = loc("Conv_255") +#loc172 = loc("Relu_256") +#loc173 = loc("Conv_257") +#loc174 = loc("Add_259") +#loc175 = loc("Clip_262") +#loc176 = loc("Div_264") +#loc177 = loc("Mul_265") +#loc178 = loc("Conv_266") +#loc179 = loc("Conv_267") +#loc180 = loc("Add_269") +#loc181 = loc("Clip_272") +#loc182 = loc("Div_274") +#loc183 = loc("Mul_275") +#loc184 = loc("Conv_276") +#loc185 = loc("Add_278") +#loc186 = loc("Clip_281") +#loc187 = loc("Div_283") +#loc188 = loc("Mul_284") +#loc189 = loc("GlobalAveragePool_285") +#loc190 = loc("Conv_286") +#loc191 = loc("Relu_287") +#loc192 = loc("Conv_288") +#loc193 = loc("Add_290") +#loc194 = loc("Clip_293") +#loc195 = loc("Div_295") +#loc196 = loc("Mul_296") +#loc197 = loc("Conv_297") +#loc198 = loc("Add_298") +#loc199 = loc("Conv_299") +#loc200 = loc("Add_301") +#loc201 = loc("Clip_304") +#loc202 = loc("Div_306") +#loc203 = loc("Mul_307") +#loc204 = loc("Conv_308") +#loc205 = loc("Add_310") +#loc206 = loc("Clip_313") +#loc207 = loc("Div_315") +#loc208 = loc("Mul_316") +#loc209 = loc("GlobalAveragePool_317") +#loc210 = loc("Conv_318") +#loc211 = loc("Relu_319") +#loc212 = loc("Conv_320") +#loc213 = loc("Add_322") +#loc214 = loc("Clip_325") +#loc215 = loc("Div_327") +#loc216 = loc("Mul_328") +#loc217 = loc("Conv_329") +#loc218 = loc("Add_330") +#loc219 = loc("Conv_331") +#loc220 = loc("Add_333") +#loc221 = loc("Clip_336") +#loc222 = loc("Div_338") +#loc223 = loc("Mul_339") +#loc224 = loc("GlobalAveragePool_342") +#loc225 = loc("Conv_343") +#loc226 = loc("Sigmoid_344") +#loc227 = loc("Mul_345") +#loc228 = loc("Conv_340") +#loc229 = loc("Relu_341") +#loc230 = loc("Split_349") +#loc231 = loc("Concat_350") +#loc232 = loc("Conv_351") +#loc233 = loc("Sigmoid_352") +#loc234 = loc("Split_353") +#loc235 = loc("Mul_354") +#loc236 = loc("Concat_355") +#loc237 = loc("Conv_356") +#loc238 = loc("Tanh_357") +#loc239 = loc("Mul_361") +#loc240 = loc("Mul_360") +#loc241 = loc("Add_362") +#loc242 = loc("Concat_363") +#loc243 = loc("Resize_365") +#loc244 = loc("Slice_371") +#loc245 = loc("AveragePool_347") +#loc246 = loc("AveragePool_348") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_380") +#loc256 = loc("Concat_381") +#loc257 = loc("Conv_382") +#loc258 = loc("Tanh_383") +#loc259 = loc("Mul_387") +#loc260 = loc("Mul_386") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_406") +#loc274 = loc("Concat_407") +#loc275 = loc("Conv_408") +#loc276 = loc("Tanh_409") +#loc277 = loc("Mul_413") +#loc278 = loc("Mul_412") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_426") +#loc291 = loc("Concat_427") +#loc292 = loc("Conv_428") +#loc293 = loc("Tanh_429") +#loc294 = loc("Mul_433") +#loc295 = loc("Mul_432") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_446") +#loc307 = loc("Clip_447") +#loc308 = loc(fused[#loc1, #loc2, #loc3, #loc4, #loc5, #loc6, #loc7, #loc8, #loc9, #loc10, #loc11, #loc12, #loc13, #loc14, #loc15, #loc16, #loc17, #loc18, #loc19, #loc20, #loc21, #loc22, #loc23, #loc24, #loc25, #loc26, #loc27, #loc28, #loc29, #loc30, #loc31, #loc32, #loc33, #loc34, #loc35, #loc36, #loc37, #loc38, #loc39, #loc40, #loc41, #loc42, #loc43, #loc44, #loc45, #loc46, #loc47, #loc48, #loc49, #loc50, #loc51, #loc52, #loc53, #loc54, #loc55, #loc56, #loc57, #loc58, #loc59, #loc60, #loc61, #loc62, #loc63, #loc64, #loc65, #loc66, #loc67, #loc68, #loc69, #loc70, #loc71, #loc72, #loc73, #loc74, #loc75, #loc76, #loc77, #loc78, #loc79, #loc80, #loc81, #loc82, #loc83, #loc84, #loc85, #loc86, #loc87, #loc88, #loc89, #loc90, #loc91, #loc92, #loc93, #loc94, #loc95, #loc96, #loc97, #loc98, #loc99, #loc100, #loc101, #loc102, #loc103, #loc104, #loc105, #loc106, #loc107, #loc108, #loc109, #loc110, #loc111, #loc112, #loc113, #loc114, #loc115, #loc116, #loc117, #loc118, #loc119, #loc120, #loc121, #loc122, #loc123, #loc124, #loc125, #loc126, #loc127, #loc128, #loc129, #loc130, #loc131, #loc132, #loc133, #loc134, #loc135, #loc136, #loc137, #loc138, #loc139, #loc140, #loc141, #loc142, #loc143, #loc144, #loc145, #loc146, #loc147, #loc148, #loc149, #loc150, #loc151, #loc152, #loc153, #loc154, #loc155, #loc156, #loc157, #loc158, #loc159, #loc160, #loc161, #loc162, #loc163, #loc164, #loc165, #loc166, #loc167, #loc168, #loc169, #loc170, #loc171, #loc172, #loc173, #loc174, #loc175, #loc176, #loc177, #loc178, #loc179, #loc180, #loc181, #loc182, #loc183, #loc184, #loc185, #loc186, #loc187, #loc188, #loc189, #loc190, #loc191, #loc192, #loc193, #loc194, #loc195, #loc196, #loc197, #loc198, #loc199, #loc200, #loc201, #loc202, #loc203, #loc204, #loc205, #loc206, #loc207, #loc208, #loc209, #loc210, #loc211, #loc212, #loc213, #loc214, #loc215, #loc216, #loc217, #loc218, #loc219, #loc220, #loc221, #loc222, #loc223, #loc224, #loc225, #loc226, #loc227, #loc228, #loc229, #loc230, #loc231, #loc232, #loc233, #loc234, #loc235, #loc236, #loc237, #loc238, #loc239, #loc240, #loc241, #loc242, #loc243, #loc244, #loc245, #loc246, #loc247, #loc248, #loc249, #loc250, #loc251, #loc252, #loc253, #loc254, #loc255, #loc256, #loc257, #loc258, #loc259, #loc260, #loc261, #loc262, #loc263, #loc264, #loc265, #loc266, #loc267, #loc268, #loc269, #loc270, #loc271, #loc272, #loc273, #loc274, #loc275, #loc276, #loc277, #loc278, #loc279, #loc280, #loc281, #loc282, #loc283, #loc284, #loc285, #loc286, #loc287, #loc288, #loc289, #loc290, #loc291, #loc292, #loc293, #loc294, #loc295, #loc296, #loc297, #loc298, #loc299, #loc300, #loc301, #loc302, #loc303, #loc304, #loc305, #loc306, #loc307]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.par.x86.wrapped.mlir b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.par.x86.wrapped.mlir new file mode 100644 index 0000000000000000000000000000000000000000..6c9e3f0f681ec13f5da27a57ac5237d2832fb610 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.par.x86.wrapped.mlir @@ -0,0 +1,24904 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func private @forward_outlined_part_0(%arg0: tensor<1x180x320x4xbf16> loc(unknown), %arg1: tensor<1x16x90x160xbf16> loc(unknown), %arg2: tensor<1x20x45x80xbf16> loc(unknown), %arg3: tensor<1x40x23x40xbf16> loc(unknown), %arg4: tensor<1x64x12x20xbf16> loc(unknown)) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) attributes {aie_partition = 0 : i32, kernel} { + %0 = xten_nn.subgraph (%arg5 = %arg0: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.906250e-03 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.906250e-03> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc1) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_2", + OutputName = "Div_2", + shift = 0 : i8} : (tensor<1x180x320x4xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %463 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %461 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + %1 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/biases"} -> tensor<4xbf16> loc(#loc) + %2 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/weights"} -> tensor<4x16x1x1xbf16> loc(#loc) + %3 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/biases"} -> tensor<16xbf16> loc(#loc) + %4 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/weights"} -> tensor<16x16x3x3xbf16> loc(#loc) + %5 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/biases"} -> tensor<16xbf16> loc(#loc) + %6 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/weights"} -> tensor<16x35x3x3xbf16> loc(#loc) + %7 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_431/Constant_0_0"} -> tensor<1x16x90x160xbf16> loc(#loc2) + %8 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/biases"} -> tensor<16xbf16> loc(#loc) + %9 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/weights"} -> tensor<16x32x3x3xbf16> loc(#loc) + %10 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/biases"} -> tensor<32xbf16> loc(#loc) + %11 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/weights"} -> tensor<32x32x3x3xbf16> loc(#loc) + %12 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/biases"} -> tensor<32xbf16> loc(#loc) + %13 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/weights"} -> tensor<32x59x3x3xbf16> loc(#loc) + %14 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_411/Constant_0_0"} -> tensor<1x20x45x80xbf16> loc(#loc3) + %15 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/biases"} -> tensor<20xbf16> loc(#loc) + %16 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/weights"} -> tensor<20x40x3x3xbf16> loc(#loc) + %17 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/biases"} -> tensor<40xbf16> loc(#loc) + %18 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/weights"} -> tensor<40x40x3x3xbf16> loc(#loc) + %19 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/biases"} -> tensor<40xbf16> loc(#loc) + %20 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/weights"} -> tensor<40x107x3x3xbf16> loc(#loc) + %21 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_385/Constant_0_0"} -> tensor<1x40x23x40xbf16> loc(#loc4) + %22 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/biases"} -> tensor<40xbf16> loc(#loc) + %23 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/weights"} -> tensor<40x80x3x3xbf16> loc(#loc) + %24 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/biases"} -> tensor<80xbf16> loc(#loc) + %25 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/weights"} -> tensor<80x80x3x3xbf16> loc(#loc) + %26 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/biases"} -> tensor<80xbf16> loc(#loc) + %27 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/weights"} -> tensor<80x171x3x3xbf16> loc(#loc) + %28 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_359/Constant_0_0"} -> tensor<1x64x12x20xbf16> loc(#loc5) + %29 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/biases"} -> tensor<64xbf16> loc(#loc) + %30 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/weights"} -> tensor<64x128x3x3xbf16> loc(#loc) + %31 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/biases"} -> tensor<128xbf16> loc(#loc) + %32 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/weights"} -> tensor<128x128x3x3xbf16> loc(#loc) + %33 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/biases"} -> tensor<128xbf16> loc(#loc) + %34 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %35 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/biases"} -> tensor<128xbf16> loc(#loc) + %36 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %37 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/biases"} -> tensor<960xbf16> loc(#loc) + %38 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %39 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/biases"} -> tensor<160xbf16> loc(#loc) + %40 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %41 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/biases"} -> tensor<960xbf16> loc(#loc) + %42 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %43 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/biases"} -> tensor<240xbf16> loc(#loc) + %44 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %45 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/biases"} -> tensor<960xbf16> loc(#loc) + %46 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %47 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/biases"} -> tensor<960xbf16> loc(#loc) + %48 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %49 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/biases"} -> tensor<160xbf16> loc(#loc) + %50 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %51 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/biases"} -> tensor<960xbf16> loc(#loc) + %52 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %53 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/biases"} -> tensor<240xbf16> loc(#loc) + %54 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %55 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/biases"} -> tensor<960xbf16> loc(#loc) + %56 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %57 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/biases"} -> tensor<960xbf16> loc(#loc) + %58 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %59 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/biases"} -> tensor<160xbf16> loc(#loc) + %60 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/weights"} -> tensor<160x672x1x1xbf16> loc(#loc) + %61 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/biases"} -> tensor<672xbf16> loc(#loc) + %62 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %63 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/biases"} -> tensor<168xbf16> loc(#loc) + %64 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %65 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/biases"} -> tensor<672xbf16> loc(#loc) + %66 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/weights"} -> tensor<672x1x9x9xbf16> loc(#loc) + %67 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/biases"} -> tensor<672xbf16> loc(#loc) + %68 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %69 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/biases"} -> tensor<112xbf16> loc(#loc) + %70 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/weights"} -> tensor<112x672x1x1xbf16> loc(#loc) + %71 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/biases"} -> tensor<672xbf16> loc(#loc) + %72 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %73 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/biases"} -> tensor<168xbf16> loc(#loc) + %74 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %75 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/biases"} -> tensor<672xbf16> loc(#loc) + %76 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/weights"} -> tensor<672x1x3x3xbf16> loc(#loc) + %77 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/biases"} -> tensor<672xbf16> loc(#loc) + %78 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %79 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/biases"} -> tensor<112xbf16> loc(#loc) + %80 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/weights"} -> tensor<112x480x1x1xbf16> loc(#loc) + %81 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/biases"} -> tensor<480xbf16> loc(#loc) + %82 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/weights"} -> tensor<480x120x1x1xbf16> loc(#loc) + %83 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/biases"} -> tensor<120xbf16> loc(#loc) + %84 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/weights"} -> tensor<120x480x1x1xbf16> loc(#loc) + %85 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/biases"} -> tensor<480xbf16> loc(#loc) + %86 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/weights"} -> tensor<480x1x3x3xbf16> loc(#loc) + %87 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/biases"} -> tensor<480xbf16> loc(#loc) + %88 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/weights"} -> tensor<480x80x1x1xbf16> loc(#loc) + %89 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/biases"} -> tensor<80xbf16> loc(#loc) + %90 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %91 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/biases"} -> tensor<184xbf16> loc(#loc) + %92 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %93 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/biases"} -> tensor<184xbf16> loc(#loc) + %94 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %95 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/biases"} -> tensor<80xbf16> loc(#loc) + %96 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %97 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/biases"} -> tensor<184xbf16> loc(#loc) + %98 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %99 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/biases"} -> tensor<184xbf16> loc(#loc) + %100 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %101 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/biases"} -> tensor<80xbf16> loc(#loc) + %102 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/weights"} -> tensor<80x200x1x1xbf16> loc(#loc) + %103 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/biases"} -> tensor<200xbf16> loc(#loc) + %104 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/weights"} -> tensor<200x1x3x3xbf16> loc(#loc) + %105 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/biases"} -> tensor<200xbf16> loc(#loc) + %106 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/weights"} -> tensor<200x80x1x1xbf16> loc(#loc) + %107 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/biases"} -> tensor<80xbf16> loc(#loc) + %108 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/weights"} -> tensor<80x240x1x1xbf16> loc(#loc) + %109 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/biases"} -> tensor<240xbf16> loc(#loc) + %110 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/weights"} -> tensor<240x1x3x3xbf16> loc(#loc) + %111 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/biases"} -> tensor<240xbf16> loc(#loc) + %112 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/weights"} -> tensor<240x40x1x1xbf16> loc(#loc) + %113 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/biases"} -> tensor<40xbf16> loc(#loc) + %114 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %115 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/biases"} -> tensor<120xbf16> loc(#loc) + %116 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %117 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/biases"} -> tensor<32xbf16> loc(#loc) + %118 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %119 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/biases"} -> tensor<120xbf16> loc(#loc) + %120 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %121 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/biases"} -> tensor<120xbf16> loc(#loc) + %122 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %123 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/biases"} -> tensor<40xbf16> loc(#loc) + %124 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %125 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/biases"} -> tensor<120xbf16> loc(#loc) + %126 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %127 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/biases"} -> tensor<32xbf16> loc(#loc) + %128 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %129 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/biases"} -> tensor<120xbf16> loc(#loc) + %130 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %131 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/biases"} -> tensor<120xbf16> loc(#loc) + %132 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %133 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/biases"} -> tensor<40xbf16> loc(#loc) + %134 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/weights"} -> tensor<40x72x1x1xbf16> loc(#loc) + %135 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/biases"} -> tensor<72xbf16> loc(#loc) + %136 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %137 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/biases"} -> tensor<24xbf16> loc(#loc) + %138 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %139 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/biases"} -> tensor<72xbf16> loc(#loc) + %140 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/weights"} -> tensor<72x1x5x5xbf16> loc(#loc) + %141 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/biases"} -> tensor<72xbf16> loc(#loc) + %142 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %143 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/biases"} -> tensor<24xbf16> loc(#loc) + %144 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %145 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/biases"} -> tensor<72xbf16> loc(#loc) + %146 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/weights"} -> tensor<72x1x3x3xbf16> loc(#loc) + %147 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/biases"} -> tensor<72xbf16> loc(#loc) + %148 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %149 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/biases"} -> tensor<24xbf16> loc(#loc) + %150 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/weights"} -> tensor<24x64x1x1xbf16> loc(#loc) + %151 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/biases"} -> tensor<64xbf16> loc(#loc) + %152 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/weights"} -> tensor<64x1x3x3xbf16> loc(#loc) + %153 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/biases"} -> tensor<64xbf16> loc(#loc) + %154 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/weights"} -> tensor<64x16x1x1xbf16> loc(#loc) + %155 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/biases"} -> tensor<16xbf16> loc(#loc) + %156 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/weights"} -> tensor<16x16x1x1xbf16> loc(#loc) + %157 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/biases"} -> tensor<16xbf16> loc(#loc) + %158 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/weights"} -> tensor<16x1x3x3xbf16> loc(#loc) + %159 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/biases"} -> tensor<16xbf16> loc(#loc) + %160 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/weights"} -> tensor<16x3x3x3xbf16> loc(#loc) + %161 = xten_nn.load_external_const {file = "constants.h5", key = "Div_16/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc6) + %162 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_14/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc309) + %163 = xten_nn.subgraph (%arg5 = %0: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Slice_7", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Slice_7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "W", + config.dim_c = 184 : ui32, + config.dim_h = 320 : ui32, + config.dim_w = 4 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_7", + OutputName = "Slice_7", + size = array, + start = array} : (tensor<1x180x320x4xbf16>) -> tensor<1x180x320x3xbf16> loc(#loc9) + xten_nn.output %461 : tensor<1x180x320x3xbf16> loc(#loc9) + } -> tensor<1x180x320x3xbf16> loc(#loc9) + %164 = xten_nn.subgraph (%arg5 = %163: tensor<1x180x320x3xbf16>) attributes { + LayerName = "Generated-#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Generated-#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "BufferPadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_0_padded = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_1_padded = 23 : ui32, + config.dim_2 = 3 : ui32, + config.dim_2_padded = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_padded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x180x320x3xbf16> loc(#loc10) + } -> tensor<1x180x320x3xbf16> loc(#loc10) + %165 = xten_nn.subgraph (%arg5 = %164: tensor<1x180x320x3xbf16>) attributes { + LayerName = "Generated-#2", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Generated-#3", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_2 = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 10 : ui32 + }} { + %461 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc11) + %462 = tosa.transpose %arg5, %461 : (tensor<1x180x320x3xbf16>, tensor<4xi32>) -> tensor<1x3x180x320xbf16> loc(#loc311) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc311) + } -> tensor<1x3x180x320xbf16> loc(#loc310) + %166 = xten_nn.subgraph (%arg5 = %165: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Generated-#4", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Generated-#5", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 184 : ui32, + config.dim_0_unpadded = 180 : ui32, + config.dim_1 = 1 : ui32, + config.dim_1_unpadded = 1 : ui32, + config.dim_2 = 320 : ui32, + config.dim_2_unpadded = 320 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x3x180x320xbf16> loc(#loc10) + } -> tensor<1x3x180x320xbf16> loc(#loc10) + %167 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>, %arg6 = %162: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Initializer_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Initializer_398", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Sub_14", OutputName = "Initializer_398"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc309) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc309) + } -> tensor<1x3x180x320xbf16> loc(#loc309) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc309) + } -> tensor<1x3x180x320xbf16> loc(#loc309) + %168 = xten_nn.subgraph (%arg5 = %167: tensor<1x3x180x320xbf16>, %arg6 = %161: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + OutputName = "Div_16", + PartOfLayerName = "Div_16", + shift = 0 : i8} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + %169 = xten_nn.subgraph (%arg5 = %168: tensor<1x3x180x320xbf16>, %arg6 = %160: tensor<16x3x3x3xbf16>, %arg7 = %159: tensor<16xbf16>) attributes { + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x3x180x320xbf16>, %arg9 = %arg6: tensor<16x3x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 2 : ui8, + config.stride_w = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x3x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x3xbf16> loc(#loc13) + %465 = tosa.transpose %arg8, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc13) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_17", + PartOfOutputName = "Conv_17", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>, tensor<16x3x3x3xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc13) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc13) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc13) + } -> tensor<1x16x90x160xbf16> loc(#loc13) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc13) + } -> tensor<1x16x90x160xbf16> loc(#loc13) + %170 = xten_nn.subgraph (%arg5 = %169: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_19", OutputName = "Add_19"} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc14) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc14) + } -> tensor<1x16x90x160xbf16> loc(#loc14) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc14) + } -> tensor<1x16x90x160xbf16> loc(#loc14) + %171 = xten_nn.subgraph (%arg5 = %170: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_22", + OutputName = "Clip_22", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + %172 = xten_nn.subgraph (%arg5 = %171: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_24", + OutputName = "Div_24", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + %173 = xten_nn.subgraph (%arg5 = %169: tensor<1x16x90x160xbf16>, %arg6 = %172: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_25", + OutputName = "Mul_25", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + %174 = xten_nn.subgraph (%arg5 = %173: tensor<1x16x90x160xbf16>, %arg6 = %158: tensor<16x1x3x3xbf16>, %arg7 = %157: tensor<16xbf16>) attributes { + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<16x1x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc312) + %465 = tosa.transpose %arg9, %464 : (tensor<16x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x16x1xbf16> loc(#loc312) + %466 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc312) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_26", + PartOfOutputName = "Conv_26", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<3x3x16x1xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc18) + %468 = tosa.clamp %467 { + LayerName = "Relu_27", + OutputName = "Relu_27", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc19) + %469 = tosa.transpose %468, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc312) + xten_nn.output %469 : tensor<1x16x90x160xbf16> loc(#loc19) + } -> tensor<1x16x90x160xbf16> loc(#loc312) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc312) + } -> tensor<1x16x90x160xbf16> loc(#loc312) + %175 = xten_nn.subgraph (%arg5 = %174: tensor<1x16x90x160xbf16>, %arg6 = %156: tensor<16x16x1x1xbf16>, %arg7 = %155: tensor<16xbf16>, %arg8 = %173: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Conv_28", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x16x90x160xbf16>, %arg10 = %arg6: tensor<16x16x1x1xbf16>, %arg11 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_28", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_28", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc20) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<16x16x1x1xbf16>) -> tensor<16x1x1x16xbf16> loc(#loc20) + %466 = tosa.transpose %arg9, %464 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_28", + PartOfOutputName = "Conv_28", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<16x1x1x16xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %468 = tosa.transpose %467, %463 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc20) + xten_nn.output %468 : tensor<1x16x90x160xbf16> loc(#loc20) + } -> tensor<1x16x90x160xbf16> loc(#loc20) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x16x90x160xbf16>, %arg10 = %arg8: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_29", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_29", OutputName = "Add_29"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc313) + %176 = xten_nn.subgraph (%arg5 = %175: tensor<1x16x90x160xbf16>, %arg6 = %154: tensor<64x16x1x1xbf16>, %arg7 = %153: tensor<64xbf16>) attributes { + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<64x16x1x1xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc314) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<64x16x1x1xbf16>) -> tensor<64x1x1x16xbf16> loc(#loc314) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc314) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_30", + PartOfOutputName = "Conv_30", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<64x1x1x16xbf16>, tensor<64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc22) + %467 = tosa.clamp %466 { + LayerName = "Relu_31", + OutputName = "Relu_31", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc23) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x64xbf16>, tensor<4xi32>) -> tensor<1x64x90x160xbf16> loc(#loc314) + xten_nn.output %468 : tensor<1x64x90x160xbf16> loc(#loc23) + } -> tensor<1x64x90x160xbf16> loc(#loc314) + xten_nn.output %461 : tensor<1x64x90x160xbf16> loc(#loc314) + } -> tensor<1x64x90x160xbf16> loc(#loc314) + %177 = xten_nn.subgraph (%arg5 = %176: tensor<1x64x90x160xbf16>, %arg6 = %152: tensor<64x1x3x3xbf16>, %arg7 = %151: tensor<64xbf16>) attributes { + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x90x160xbf16>, %arg9 = %arg6: tensor<64x1x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc315) + %465 = tosa.transpose %arg9, %464 : (tensor<64x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x64x1xbf16> loc(#loc315) + %466 = tosa.transpose %arg8, %463 : (tensor<1x64x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x64xbf16> loc(#loc315) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_32", + PartOfOutputName = "Conv_32", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x64xbf16>, tensor<3x3x64x1xbf16>, tensor<64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc24) + %468 = tosa.clamp %467 { + LayerName = "Relu_33", + OutputName = "Relu_33", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc25) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x64xbf16>, tensor<4xi32>) -> tensor<1x64x45x80xbf16> loc(#loc315) + xten_nn.output %469 : tensor<1x64x45x80xbf16> loc(#loc25) + } -> tensor<1x64x45x80xbf16> loc(#loc315) + xten_nn.output %461 : tensor<1x64x45x80xbf16> loc(#loc315) + } -> tensor<1x64x45x80xbf16> loc(#loc315) + %178 = xten_nn.subgraph (%arg5 = %177: tensor<1x64x45x80xbf16>, %arg6 = %150: tensor<24x64x1x1xbf16>, %arg7 = %149: tensor<24xbf16>) attributes { + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x45x80xbf16>, %arg9 = %arg6: tensor<24x64x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc26) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x64x1x1xbf16>) -> tensor<24x1x1x64xbf16> loc(#loc26) + %465 = tosa.transpose %arg8, %463 : (tensor<1x64x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x64xbf16> loc(#loc26) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_34", + PartOfOutputName = "Conv_34", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x64xbf16>, tensor<24x1x1x64xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc26) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc26) + xten_nn.output %467 : tensor<1x24x45x80xbf16> loc(#loc26) + } -> tensor<1x24x45x80xbf16> loc(#loc26) + xten_nn.output %461 : tensor<1x24x45x80xbf16> loc(#loc26) + } -> tensor<1x24x45x80xbf16> loc(#loc26) + %179 = xten_nn.subgraph (%arg5 = %178: tensor<1x24x45x80xbf16>, %arg6 = %148: tensor<72x24x1x1xbf16>, %arg7 = %147: tensor<72xbf16>) attributes { + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc316) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc316) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc316) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_35", + PartOfOutputName = "Conv_35", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc27) + %467 = tosa.clamp %466 { + LayerName = "Relu_36", + OutputName = "Relu_36", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc28) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc316) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc28) + } -> tensor<1x72x45x80xbf16> loc(#loc316) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc316) + } -> tensor<1x72x45x80xbf16> loc(#loc316) + %180 = xten_nn.subgraph (%arg5 = %179: tensor<1x72x45x80xbf16>, %arg6 = %146: tensor<72x1x3x3xbf16>, %arg7 = %145: tensor<72xbf16>) attributes { + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x3x3xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc317) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x72x1xbf16> loc(#loc317) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc317) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_37", + PartOfOutputName = "Conv_37", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<3x3x72x1xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc29) + %468 = tosa.clamp %467 { + LayerName = "Relu_38", + OutputName = "Relu_38", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc30) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc317) + xten_nn.output %469 : tensor<1x72x45x80xbf16> loc(#loc30) + } -> tensor<1x72x45x80xbf16> loc(#loc317) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc317) + } -> tensor<1x72x45x80xbf16> loc(#loc317) + %181 = xten_nn.subgraph (%arg5 = %180: tensor<1x72x45x80xbf16>, %arg6 = %144: tensor<24x72x1x1xbf16>, %arg7 = %143: tensor<24xbf16>, %arg8 = %178: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Conv_39", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x72x45x80xbf16>, %arg10 = %arg6: tensor<24x72x1x1xbf16>, %arg11 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_39", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_39", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc31) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc31) + %466 = tosa.transpose %arg9, %464 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc31) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_39", + PartOfOutputName = "Conv_39", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc31) + %468 = tosa.transpose %467, %463 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc31) + xten_nn.output %468 : tensor<1x24x45x80xbf16> loc(#loc31) + } -> tensor<1x24x45x80xbf16> loc(#loc31) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x24x45x80xbf16>, %arg10 = %arg8: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Add_40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_40", OutputName = "Add_40"} : (tensor<1x24x45x80xbf16>, tensor<1x24x45x80xbf16>) -> tensor<1x24x45x80xbf16> loc(#loc32) + xten_nn.output %463 : tensor<1x24x45x80xbf16> loc(#loc32) + } -> tensor<1x24x45x80xbf16> loc(#loc32) + xten_nn.output %462 : tensor<1x24x45x80xbf16> loc(#loc32) + } -> tensor<1x24x45x80xbf16> loc(#loc318) + %182 = xten_nn.subgraph (%arg5 = %181: tensor<1x24x45x80xbf16>, %arg6 = %142: tensor<72x24x1x1xbf16>, %arg7 = %141: tensor<72xbf16>) attributes { + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc319) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc319) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc319) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_41", + PartOfOutputName = "Conv_41", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc33) + %467 = tosa.clamp %466 { + LayerName = "Relu_42", + OutputName = "Relu_42", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc34) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc319) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc34) + } -> tensor<1x72x45x80xbf16> loc(#loc319) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc319) + } -> tensor<1x72x45x80xbf16> loc(#loc319) + %183 = xten_nn.subgraph (%arg5 = %182: tensor<1x72x45x80xbf16>, %arg6 = %140: tensor<72x1x5x5xbf16>, %arg7 = %139: tensor<72xbf16>) attributes { + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x5x5xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 1]], + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc320) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x72x1xbf16> loc(#loc320) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc320) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_43", + PartOfOutputName = "Conv_43", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<5x5x72x1xbf16>, tensor<72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc35) + %468 = tosa.clamp %467 { + LayerName = "Relu_44", + OutputName = "Relu_44", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc36) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x72xbf16>, tensor<4xi32>) -> tensor<1x72x23x40xbf16> loc(#loc320) + xten_nn.output %469 : tensor<1x72x23x40xbf16> loc(#loc36) + } -> tensor<1x72x23x40xbf16> loc(#loc320) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc320) + } -> tensor<1x72x23x40xbf16> loc(#loc320) + %184 = xten_nn.subgraph (%arg5 = %183: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Generated-#6", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 9 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x72x23x40xbf16>) -> tensor<1x72x1x920xbf16> loc(#loc37) + xten_nn.output %461 : tensor<1x72x1x920xbf16> loc(#loc37) + } -> tensor<1x72x1x920xbf16> loc(#loc37) + %185 = xten_nn.subgraph (%arg5 = %184: tensor<1x72x1x920xbf16>) attributes { + LayerName = "Generated-#8", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#9", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x920xbf16>) attributes { + LayerName = "Generated-#8", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#9", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 72 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x72x1x920xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc37) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc37) + } -> tensor<1x72x1x1xbf16> loc(#loc37) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc37) + } -> tensor<1x72x1x1xbf16> loc(#loc37) + %186 = xten_nn.subgraph (%arg5 = %185: tensor<1x72x1x1xbf16>, %arg6 = %138: tensor<24x72x1x1xbf16>, %arg7 = %137: tensor<24xbf16>) attributes { + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x1x1xbf16>, %arg9 = %arg6: tensor<24x72x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc321) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc321) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_46", + PartOfOutputName = "Conv_46", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc38) + %465 = tosa.clamp %464 { + LayerName = "Relu_47", + OutputName = "Relu_47", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc39) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x24xbf16>) -> tensor<1x24x1x1xbf16> loc(#loc321) + xten_nn.output %466 : tensor<1x24x1x1xbf16> loc(#loc39) + } -> tensor<1x24x1x1xbf16> loc(#loc321) + xten_nn.output %461 : tensor<1x24x1x1xbf16> loc(#loc321) + } -> tensor<1x24x1x1xbf16> loc(#loc321) + %187 = xten_nn.subgraph (%arg5 = %186: tensor<1x24x1x1xbf16>, %arg6 = %136: tensor<72x24x1x1xbf16>, %arg7 = %135: tensor<72xbf16>) attributes { + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x1x1xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc40) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x24x1x1xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc40) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_48", + PartOfOutputName = "Conv_48", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc40) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x72xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc40) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc40) + } -> tensor<1x72x1x1xbf16> loc(#loc40) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc40) + } -> tensor<1x72x1x1xbf16> loc(#loc40) + %188 = xten_nn.subgraph (%arg5 = %187: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_50", OutputName = "Add_50"} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc41) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc41) + } -> tensor<1x72x1x1xbf16> loc(#loc41) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc41) + } -> tensor<1x72x1x1xbf16> loc(#loc41) + %189 = xten_nn.subgraph (%arg5 = %188: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_53", + OutputName = "Clip_53", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + %190 = xten_nn.subgraph (%arg5 = %189: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_55", + OutputName = "Div_55", + shift = 0 : i8} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + %191 = xten_nn.subgraph (%arg5 = %190: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Generated-#10", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#11", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 72 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + %192 = xten_nn.subgraph (%arg5 = %191: tensor<1x72x23x40xbf16>, %arg6 = %183: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x72x23x40xbf16>, %arg8 = %arg6: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_56", + OutputName = "Mul_56", + shift = 0 : i8} : (tensor<1x72x23x40xbf16>, tensor<1x72x23x40xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %462 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc44) + } -> tensor<1x72x23x40xbf16> loc(#loc44) + %193 = xten_nn.subgraph (%arg5 = %192: tensor<1x72x23x40xbf16>, %arg6 = %134: tensor<40x72x1x1xbf16>, %arg7 = %133: tensor<40xbf16>) attributes { + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x23x40xbf16>, %arg9 = %arg6: tensor<40x72x1x1xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc45) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<40x72x1x1xbf16>) -> tensor<40x1x1x72xbf16> loc(#loc45) + %465 = tosa.transpose %arg8, %463 : (tensor<1x72x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x72xbf16> loc(#loc45) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_57", + PartOfOutputName = "Conv_57", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x72xbf16>, tensor<40x1x1x72xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc45) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc45) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc45) + } -> tensor<1x40x23x40xbf16> loc(#loc45) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc45) + } -> tensor<1x40x23x40xbf16> loc(#loc45) + %194 = xten_nn.subgraph (%arg5 = %193: tensor<1x40x23x40xbf16>, %arg6 = %132: tensor<120x40x1x1xbf16>, %arg7 = %131: tensor<120xbf16>) attributes { + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc322) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc322) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc322) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_58", + PartOfOutputName = "Conv_58", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc46) + %467 = tosa.clamp %466 { + LayerName = "Relu_59", + OutputName = "Relu_59", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc47) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc322) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc47) + } -> tensor<1x120x23x40xbf16> loc(#loc322) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc322) + } -> tensor<1x120x23x40xbf16> loc(#loc322) + %195 = xten_nn.subgraph (%arg5 = %194: tensor<1x120x23x40xbf16>, %arg6 = %130: tensor<120x1x5x5xbf16>, %arg7 = %129: tensor<120xbf16>) attributes { + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc323) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc323) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc323) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_60", + PartOfOutputName = "Conv_60", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc48) + %468 = tosa.clamp %467 { + LayerName = "Relu_61", + OutputName = "Relu_61", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc49) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc323) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc49) + } -> tensor<1x120x23x40xbf16> loc(#loc323) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc323) + } -> tensor<1x120x23x40xbf16> loc(#loc323) + %196 = xten_nn.subgraph (%arg5 = %195: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Generated-#12", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#13", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc50) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc50) + } -> tensor<1x120x1x920xbf16> loc(#loc50) + %197 = xten_nn.subgraph (%arg5 = %196: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#15", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#15", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc50) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc50) + } -> tensor<1x120x1x1xbf16> loc(#loc50) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc50) + } -> tensor<1x120x1x1xbf16> loc(#loc50) + %198 = xten_nn.subgraph (%arg5 = %197: tensor<1x120x1x1xbf16>, %arg6 = %128: tensor<32x120x1x1xbf16>, %arg7 = %127: tensor<32xbf16>) attributes { + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc324) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc324) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_63", + PartOfOutputName = "Conv_63", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc51) + %465 = tosa.clamp %464 { + LayerName = "Relu_64", + OutputName = "Relu_64", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc52) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc324) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc52) + } -> tensor<1x32x1x1xbf16> loc(#loc324) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc324) + } -> tensor<1x32x1x1xbf16> loc(#loc324) + %199 = xten_nn.subgraph (%arg5 = %198: tensor<1x32x1x1xbf16>, %arg6 = %126: tensor<120x32x1x1xbf16>, %arg7 = %125: tensor<120xbf16>) attributes { + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc53) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc53) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_65", + PartOfOutputName = "Conv_65", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc53) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc53) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc53) + } -> tensor<1x120x1x1xbf16> loc(#loc53) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc53) + } -> tensor<1x120x1x1xbf16> loc(#loc53) + %200 = xten_nn.subgraph (%arg5 = %199: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_67", OutputName = "Add_67"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc54) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc54) + } -> tensor<1x120x1x1xbf16> loc(#loc54) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc54) + } -> tensor<1x120x1x1xbf16> loc(#loc54) + %201 = xten_nn.subgraph (%arg5 = %200: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_70", + OutputName = "Clip_70", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + %202 = xten_nn.subgraph (%arg5 = %201: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_72", + OutputName = "Div_72", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + %203 = xten_nn.subgraph (%arg5 = %202: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Generated-#16", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#17", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + %204 = xten_nn.subgraph (%arg5 = %203: tensor<1x120x23x40xbf16>, %arg6 = %195: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_73", + OutputName = "Mul_73", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc57) + } -> tensor<1x120x23x40xbf16> loc(#loc57) + %205 = xten_nn.subgraph (%arg5 = %204: tensor<1x120x23x40xbf16>, %arg6 = %124: tensor<40x120x1x1xbf16>, %arg7 = %123: tensor<40xbf16>, %arg8 = %193: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_74", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_74", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_74", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc58) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc58) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc58) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_74", + PartOfOutputName = "Conv_74", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc58) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc58) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc58) + } -> tensor<1x40x23x40xbf16> loc(#loc58) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_75", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_75", OutputName = "Add_75"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc59) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc59) + } -> tensor<1x40x23x40xbf16> loc(#loc59) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc59) + } -> tensor<1x40x23x40xbf16> loc(#loc325) + %206 = xten_nn.subgraph (%arg5 = %205: tensor<1x40x23x40xbf16>, %arg6 = %122: tensor<120x40x1x1xbf16>, %arg7 = %121: tensor<120xbf16>) attributes { + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc326) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc326) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc326) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_76", + PartOfOutputName = "Conv_76", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc60) + %467 = tosa.clamp %466 { + LayerName = "Relu_77", + OutputName = "Relu_77", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc61) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc326) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc61) + } -> tensor<1x120x23x40xbf16> loc(#loc326) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc326) + } -> tensor<1x120x23x40xbf16> loc(#loc326) + %207 = xten_nn.subgraph (%arg5 = %206: tensor<1x120x23x40xbf16>, %arg6 = %120: tensor<120x1x5x5xbf16>, %arg7 = %119: tensor<120xbf16>) attributes { + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc327) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc327) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc327) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_78", + PartOfOutputName = "Conv_78", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc62) + %468 = tosa.clamp %467 { + LayerName = "Relu_79", + OutputName = "Relu_79", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc63) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc327) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc63) + } -> tensor<1x120x23x40xbf16> loc(#loc327) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc327) + } -> tensor<1x120x23x40xbf16> loc(#loc327) + %208 = xten_nn.subgraph (%arg5 = %207: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Generated-#18", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Generated-#19", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc64) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc64) + } -> tensor<1x120x1x920xbf16> loc(#loc64) + %209 = xten_nn.subgraph (%arg5 = %208: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#20", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#21", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "Generated-#20", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "Generated-#21", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc64) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc64) + } -> tensor<1x120x1x1xbf16> loc(#loc64) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc64) + } -> tensor<1x120x1x1xbf16> loc(#loc64) + %210 = xten_nn.subgraph (%arg5 = %209: tensor<1x120x1x1xbf16>, %arg6 = %118: tensor<32x120x1x1xbf16>, %arg7 = %117: tensor<32xbf16>) attributes { + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc328) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc328) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_81", + PartOfOutputName = "Conv_81", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc65) + %465 = tosa.clamp %464 { + LayerName = "Relu_82", + OutputName = "Relu_82", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc66) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc328) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc66) + } -> tensor<1x32x1x1xbf16> loc(#loc328) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc328) + } -> tensor<1x32x1x1xbf16> loc(#loc328) + %211 = xten_nn.subgraph (%arg5 = %210: tensor<1x32x1x1xbf16>, %arg6 = %116: tensor<120x32x1x1xbf16>, %arg7 = %115: tensor<120xbf16>) attributes { + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc67) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc67) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_83", + PartOfOutputName = "Conv_83", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc67) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc67) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc67) + } -> tensor<1x120x1x1xbf16> loc(#loc67) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc67) + } -> tensor<1x120x1x1xbf16> loc(#loc67) + %212 = xten_nn.subgraph (%arg5 = %211: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_85", OutputName = "Add_85"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc68) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc68) + } -> tensor<1x120x1x1xbf16> loc(#loc68) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc68) + } -> tensor<1x120x1x1xbf16> loc(#loc68) + %213 = xten_nn.subgraph (%arg5 = %212: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_88", + OutputName = "Clip_88", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + %214 = xten_nn.subgraph (%arg5 = %213: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_90", + OutputName = "Div_90", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + %215 = xten_nn.subgraph (%arg5 = %214: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Generated-#22", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#23", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + %216 = xten_nn.subgraph (%arg5 = %215: tensor<1x120x23x40xbf16>, %arg6 = %207: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_91", + OutputName = "Mul_91", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc71) + } -> tensor<1x120x23x40xbf16> loc(#loc71) + %217 = xten_nn.subgraph (%arg5 = %216: tensor<1x120x23x40xbf16>, %arg6 = %114: tensor<40x120x1x1xbf16>, %arg7 = %113: tensor<40xbf16>, %arg8 = %205: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_92", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_92", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_92", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc72) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc72) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc72) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_92", + PartOfOutputName = "Conv_92", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc72) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc72) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc72) + } -> tensor<1x40x23x40xbf16> loc(#loc72) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_93", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_93", OutputName = "Add_93"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc73) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc73) + } -> tensor<1x40x23x40xbf16> loc(#loc73) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc73) + } -> tensor<1x40x23x40xbf16> loc(#loc329) + %218 = xten_nn.subgraph (%arg5 = %217: tensor<1x40x23x40xbf16>, %arg6 = %112: tensor<240x40x1x1xbf16>, %arg7 = %111: tensor<240xbf16>) attributes { + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<240x40x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc74) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x40x1x1xbf16>) -> tensor<240x1x1x40xbf16> loc(#loc74) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc74) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_94", + PartOfOutputName = "Conv_94", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<240x1x1x40xbf16>, tensor<240xbf16>) -> tensor<1x23x40x240xbf16> loc(#loc74) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x240xbf16>, tensor<4xi32>) -> tensor<1x240x23x40xbf16> loc(#loc74) + xten_nn.output %467 : tensor<1x240x23x40xbf16> loc(#loc74) + } -> tensor<1x240x23x40xbf16> loc(#loc74) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc74) + } -> tensor<1x240x23x40xbf16> loc(#loc74) + %219 = xten_nn.subgraph (%arg5 = %218: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_96", OutputName = "Add_96"} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc75) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc75) + } -> tensor<1x240x23x40xbf16> loc(#loc75) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc75) + } -> tensor<1x240x23x40xbf16> loc(#loc75) + %220 = xten_nn.subgraph (%arg5 = %219: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_99", + OutputName = "Clip_99", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + %221 = xten_nn.subgraph (%arg5 = %220: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_101", + OutputName = "Div_101", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + %222 = xten_nn.subgraph (%arg5 = %218: tensor<1x240x23x40xbf16>, %arg6 = %221: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x23x40xbf16>, %arg8 = %arg6: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_102", + OutputName = "Mul_102", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + %223 = xten_nn.subgraph (%arg5 = %222: tensor<1x240x23x40xbf16>, %arg6 = %110: tensor<240x1x3x3xbf16>, %arg7 = %109: tensor<240xbf16>) attributes { + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x23x40xbf16>, %arg9 = %arg6: tensor<240x1x3x3xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 0]], + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc79) + %465 = tosa.transpose %arg9, %464 : (tensor<240x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x240x1xbf16> loc(#loc79) + %466 = tosa.transpose %arg8, %463 : (tensor<1x240x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x240xbf16> loc(#loc79) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_103", + PartOfOutputName = "Conv_103", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x240xbf16>, tensor<3x3x240x1xbf16>, tensor<240xbf16>) -> tensor<1x12x20x240xbf16> loc(#loc79) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x240xbf16>, tensor<4xi32>) -> tensor<1x240x12x20xbf16> loc(#loc79) + xten_nn.output %468 : tensor<1x240x12x20xbf16> loc(#loc79) + } -> tensor<1x240x12x20xbf16> loc(#loc79) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc79) + } -> tensor<1x240x12x20xbf16> loc(#loc79) + %224 = xten_nn.subgraph (%arg5 = %223: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_105", OutputName = "Add_105"} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc80) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc80) + } -> tensor<1x240x12x20xbf16> loc(#loc80) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc80) + } -> tensor<1x240x12x20xbf16> loc(#loc80) + %225 = xten_nn.subgraph (%arg5 = %224: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_108", + OutputName = "Clip_108", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + %226 = xten_nn.subgraph (%arg5 = %225: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_110", + OutputName = "Div_110", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + %227 = xten_nn.subgraph (%arg5 = %223: tensor<1x240x12x20xbf16>, %arg6 = %226: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x12x20xbf16>, %arg8 = %arg6: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_111", + OutputName = "Mul_111", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + %228 = xten_nn.subgraph (%arg5 = %227: tensor<1x240x12x20xbf16>, %arg6 = %108: tensor<80x240x1x1xbf16>, %arg7 = %107: tensor<80xbf16>) attributes { + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x12x20xbf16>, %arg9 = %arg6: tensor<80x240x1x1xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc84) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<80x240x1x1xbf16>) -> tensor<80x1x1x240xbf16> loc(#loc84) + %465 = tosa.transpose %arg8, %463 : (tensor<1x240x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x240xbf16> loc(#loc84) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_112", + PartOfOutputName = "Conv_112", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x240xbf16>, tensor<80x1x1x240xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc84) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc84) + xten_nn.output %467 : tensor<1x80x12x20xbf16> loc(#loc84) + } -> tensor<1x80x12x20xbf16> loc(#loc84) + xten_nn.output %461 : tensor<1x80x12x20xbf16> loc(#loc84) + } -> tensor<1x80x12x20xbf16> loc(#loc84) + %229 = xten_nn.subgraph (%arg5 = %228: tensor<1x80x12x20xbf16>, %arg6 = %106: tensor<200x80x1x1xbf16>, %arg7 = %105: tensor<200xbf16>) attributes { + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<200x80x1x1xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc85) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<200x80x1x1xbf16>) -> tensor<200x1x1x80xbf16> loc(#loc85) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc85) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_113", + PartOfOutputName = "Conv_113", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<200x1x1x80xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc85) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc85) + xten_nn.output %467 : tensor<1x200x12x20xbf16> loc(#loc85) + } -> tensor<1x200x12x20xbf16> loc(#loc85) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc85) + } -> tensor<1x200x12x20xbf16> loc(#loc85) + %230 = xten_nn.subgraph (%arg5 = %229: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_115", OutputName = "Add_115"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc86) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc86) + } -> tensor<1x200x12x20xbf16> loc(#loc86) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc86) + } -> tensor<1x200x12x20xbf16> loc(#loc86) + %231 = xten_nn.subgraph (%arg5 = %230: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_118", + OutputName = "Clip_118", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + %232 = xten_nn.subgraph (%arg5 = %231: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_120", + OutputName = "Div_120", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + %233 = xten_nn.subgraph (%arg5 = %229: tensor<1x200x12x20xbf16>, %arg6 = %232: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_121", + OutputName = "Mul_121", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + %234 = xten_nn.subgraph (%arg5 = %233: tensor<1x200x12x20xbf16>, %arg6 = %104: tensor<200x1x3x3xbf16>, %arg7 = %103: tensor<200xbf16>) attributes { + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x200x12x20xbf16>, %arg9 = %arg6: tensor<200x1x3x3xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc90) + %465 = tosa.transpose %arg9, %464 : (tensor<200x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x200x1xbf16> loc(#loc90) + %466 = tosa.transpose %arg8, %463 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc90) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_122", + PartOfOutputName = "Conv_122", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<3x3x200x1xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc90) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %468 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + %235 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_124", OutputName = "Add_124"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + %236 = xten_nn.subgraph (%arg5 = %235: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_127", + OutputName = "Clip_127", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + %237 = xten_nn.subgraph (%arg5 = %236: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_129", + OutputName = "Div_129", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + %238 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>, %arg6 = %237: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_130", + OutputName = "Mul_130", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + %239 = xten_nn.subgraph (%arg5 = %238: tensor<1x200x12x20xbf16>, %arg6 = %102: tensor<80x200x1x1xbf16>, %arg7 = %101: tensor<80xbf16>, %arg8 = %228: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_131", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x200x12x20xbf16>, %arg10 = %arg6: tensor<80x200x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_131", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_131", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc95) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x200x1x1xbf16>) -> tensor<80x1x1x200xbf16> loc(#loc95) + %466 = tosa.transpose %arg9, %464 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc95) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_131", + PartOfOutputName = "Conv_131", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<80x1x1x200xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc95) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc95) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc95) + } -> tensor<1x80x12x20xbf16> loc(#loc95) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_132", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_132", OutputName = "Add_132"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc96) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc96) + } -> tensor<1x80x12x20xbf16> loc(#loc96) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc96) + } -> tensor<1x80x12x20xbf16> loc(#loc330) + %240 = xten_nn.subgraph (%arg5 = %239: tensor<1x80x12x20xbf16>, %arg6 = %100: tensor<184x80x1x1xbf16>, %arg7 = %99: tensor<184xbf16>) attributes { + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc97) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc97) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc97) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_133", + PartOfOutputName = "Conv_133", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc97) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc97) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc97) + } -> tensor<1x184x12x20xbf16> loc(#loc97) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc97) + } -> tensor<1x184x12x20xbf16> loc(#loc97) + %241 = xten_nn.subgraph (%arg5 = %240: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_135", OutputName = "Add_135"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc98) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc98) + } -> tensor<1x184x12x20xbf16> loc(#loc98) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc98) + } -> tensor<1x184x12x20xbf16> loc(#loc98) + %242 = xten_nn.subgraph (%arg5 = %241: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_138", + OutputName = "Clip_138", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + %243 = xten_nn.subgraph (%arg5 = %242: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_140", + OutputName = "Div_140", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + %244 = xten_nn.subgraph (%arg5 = %240: tensor<1x184x12x20xbf16>, %arg6 = %243: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_141", + OutputName = "Mul_141", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + %245 = xten_nn.subgraph (%arg5 = %244: tensor<1x184x12x20xbf16>, %arg6 = %98: tensor<184x1x3x3xbf16>, %arg7 = %97: tensor<184xbf16>) attributes { + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc102) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc102) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc102) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_142", + PartOfOutputName = "Conv_142", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc102) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + %246 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_144", OutputName = "Add_144"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + %247 = xten_nn.subgraph (%arg5 = %246: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_147", + OutputName = "Clip_147", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + %248 = xten_nn.subgraph (%arg5 = %247: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_149", + OutputName = "Div_149", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + %249 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>, %arg6 = %248: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_150", + OutputName = "Mul_150", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + %250 = xten_nn.subgraph (%arg5 = %249: tensor<1x184x12x20xbf16>, %arg6 = %96: tensor<80x184x1x1xbf16>, %arg7 = %95: tensor<80xbf16>, %arg8 = %239: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_151", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_151", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_151", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc107) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc107) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc107) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_151", + PartOfOutputName = "Conv_151", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc107) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc107) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc107) + } -> tensor<1x80x12x20xbf16> loc(#loc107) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_152", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_152", OutputName = "Add_152"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc108) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc108) + } -> tensor<1x80x12x20xbf16> loc(#loc108) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc108) + } -> tensor<1x80x12x20xbf16> loc(#loc331) + %251 = xten_nn.subgraph (%arg5 = %250: tensor<1x80x12x20xbf16>, %arg6 = %94: tensor<184x80x1x1xbf16>, %arg7 = %93: tensor<184xbf16>) attributes { + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc109) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc109) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc109) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_153", + PartOfOutputName = "Conv_153", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc109) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + %252 = xten_nn.subgraph (%arg5 = %251: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_155", OutputName = "Add_155"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + %253 = xten_nn.subgraph (%arg5 = %252: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_158", + OutputName = "Clip_158", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + %254 = xten_nn.subgraph (%arg5 = %253: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_160", + OutputName = "Div_160", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + %255 = xten_nn.subgraph (%arg5 = %251: tensor<1x184x12x20xbf16>, %arg6 = %254: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_161", + OutputName = "Mul_161", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + %256 = xten_nn.subgraph (%arg5 = %255: tensor<1x184x12x20xbf16>, %arg6 = %92: tensor<184x1x3x3xbf16>, %arg7 = %91: tensor<184xbf16>) attributes { + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc114) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc114) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc114) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_162", + PartOfOutputName = "Conv_162", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc114) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + %257 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_164", OutputName = "Add_164"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + %258 = xten_nn.subgraph (%arg5 = %257: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_167", + OutputName = "Clip_167", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + %259 = xten_nn.subgraph (%arg5 = %258: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_169", + OutputName = "Div_169", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + %260 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>, %arg6 = %259: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_170", + OutputName = "Mul_170", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + %261 = xten_nn.subgraph (%arg5 = %260: tensor<1x184x12x20xbf16>, %arg6 = %90: tensor<80x184x1x1xbf16>, %arg7 = %89: tensor<80xbf16>, %arg8 = %250: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_171", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_171", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_171", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc119) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc119) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc119) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_171", + PartOfOutputName = "Conv_171", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc119) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc119) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc119) + } -> tensor<1x80x12x20xbf16> loc(#loc119) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_172", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_172", OutputName = "Add_172"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc120) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc120) + } -> tensor<1x80x12x20xbf16> loc(#loc120) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc120) + } -> tensor<1x80x12x20xbf16> loc(#loc332) + %262 = xten_nn.subgraph (%arg5 = %261: tensor<1x80x12x20xbf16>, %arg6 = %88: tensor<480x80x1x1xbf16>, %arg7 = %87: tensor<480xbf16>) attributes { + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<480x80x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc121) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x80x1x1xbf16>) -> tensor<480x1x1x80xbf16> loc(#loc121) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc121) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_173", + PartOfOutputName = "Conv_173", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<480x1x1x80xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc121) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc121) + xten_nn.output %467 : tensor<1x480x12x20xbf16> loc(#loc121) + } -> tensor<1x480x12x20xbf16> loc(#loc121) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc121) + } -> tensor<1x480x12x20xbf16> loc(#loc121) + %263 = xten_nn.subgraph (%arg5 = %262: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_175", OutputName = "Add_175"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc122) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc122) + } -> tensor<1x480x12x20xbf16> loc(#loc122) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc122) + } -> tensor<1x480x12x20xbf16> loc(#loc122) + %264 = xten_nn.subgraph (%arg5 = %263: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_178", + OutputName = "Clip_178", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + %265 = xten_nn.subgraph (%arg5 = %264: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_180", + OutputName = "Div_180", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + %266 = xten_nn.subgraph (%arg5 = %262: tensor<1x480x12x20xbf16>, %arg6 = %265: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_181", + OutputName = "Mul_181", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + %267 = xten_nn.subgraph (%arg5 = %266: tensor<1x480x12x20xbf16>, %arg6 = %86: tensor<480x1x3x3xbf16>, %arg7 = %85: tensor<480xbf16>) attributes { + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<480x1x3x3xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc126) + %465 = tosa.transpose %arg9, %464 : (tensor<480x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x480x1xbf16> loc(#loc126) + %466 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc126) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_182", + PartOfOutputName = "Conv_182", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<3x3x480x1xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc126) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %468 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + %268 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_184", OutputName = "Add_184"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + %269 = xten_nn.subgraph (%arg5 = %268: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_187", + OutputName = "Clip_187", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + %270 = xten_nn.subgraph (%arg5 = %269: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_189", + OutputName = "Div_189", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + %271 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>, %arg6 = %270: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_190", + OutputName = "Mul_190", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + %272 = xten_nn.subgraph (%arg5 = %271: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Generated-#24", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#25", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 60 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x1x240xbf16> loc(#loc333) + xten_nn.output %461 : tensor<1x480x1x240xbf16> loc(#loc333) + } -> tensor<1x480x1x240xbf16> loc(#loc333) + %273 = xten_nn.subgraph (%arg5 = %272: tensor<1x480x1x240xbf16>) attributes { + LayerName = "Generated-#26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x240xbf16>) attributes { + LayerName = "Generated-#26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#27", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 480 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x480x1x240xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc131) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc131) + } -> tensor<1x480x1x1xbf16> loc(#loc131) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc131) + } -> tensor<1x480x1x1xbf16> loc(#loc131) + %274 = xten_nn.subgraph (%arg5 = %273: tensor<1x480x1x1xbf16>, %arg6 = %84: tensor<120x480x1x1xbf16>, %arg7 = %83: tensor<120xbf16>) attributes { + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x1x1xbf16>, %arg9 = %arg6: tensor<120x480x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x480x1x1xbf16>) -> tensor<120x1x1x480xbf16> loc(#loc334) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc334) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_192", + PartOfOutputName = "Conv_192", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x480xbf16>, tensor<120x1x1x480xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc132) + %465 = tosa.clamp %464 { + LayerName = "Relu_193", + OutputName = "Relu_193", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc133) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc334) + xten_nn.output %466 : tensor<1x120x1x1xbf16> loc(#loc133) + } -> tensor<1x120x1x1xbf16> loc(#loc334) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc334) + } -> tensor<1x120x1x1xbf16> loc(#loc334) + %275 = xten_nn.subgraph (%arg5 = %274: tensor<1x120x1x1xbf16>, %arg6 = %82: tensor<480x120x1x1xbf16>, %arg7 = %81: tensor<480xbf16>) attributes { + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<480x120x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x120x1x1xbf16>) -> tensor<480x1x1x120xbf16> loc(#loc134) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc134) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_194", + PartOfOutputName = "Conv_194", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<480x1x1x120xbf16>, tensor<480xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc134) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x480xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc134) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc134) + } -> tensor<1x480x1x1xbf16> loc(#loc134) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc134) + } -> tensor<1x480x1x1xbf16> loc(#loc134) + %276 = xten_nn.subgraph (%arg5 = %275: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_196", OutputName = "Add_196"} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc135) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc135) + } -> tensor<1x480x1x1xbf16> loc(#loc135) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc135) + } -> tensor<1x480x1x1xbf16> loc(#loc135) + %277 = xten_nn.subgraph (%arg5 = %276: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_199", + OutputName = "Clip_199", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + %278 = xten_nn.subgraph (%arg5 = %277: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_201", + OutputName = "Div_201", + shift = 0 : i8} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + %279 = xten_nn.subgraph (%arg5 = %278: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Generated-#28", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#29", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 480 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + %280 = xten_nn.subgraph (%arg5 = %279: tensor<1x480x12x20xbf16>, %arg6 = %271: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_202", + OutputName = "Mul_202", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc138) + } -> tensor<1x480x12x20xbf16> loc(#loc138) + %281 = xten_nn.subgraph (%arg5 = %280: tensor<1x480x12x20xbf16>, %arg6 = %80: tensor<112x480x1x1xbf16>, %arg7 = %79: tensor<112xbf16>) attributes { + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<112x480x1x1xbf16>, %arg10 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc139) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<112x480x1x1xbf16>) -> tensor<112x1x1x480xbf16> loc(#loc139) + %465 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc139) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_203", + PartOfOutputName = "Conv_203", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<112x1x1x480xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc139) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc139) + xten_nn.output %467 : tensor<1x112x12x20xbf16> loc(#loc139) + } -> tensor<1x112x12x20xbf16> loc(#loc139) + xten_nn.output %461 : tensor<1x112x12x20xbf16> loc(#loc139) + } -> tensor<1x112x12x20xbf16> loc(#loc139) + %282 = xten_nn.subgraph (%arg5 = %281: tensor<1x112x12x20xbf16>, %arg6 = %78: tensor<672x112x1x1xbf16>, %arg7 = %77: tensor<672xbf16>) attributes { + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc140) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc140) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc140) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_204", + PartOfOutputName = "Conv_204", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc140) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc140) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc140) + } -> tensor<1x672x12x20xbf16> loc(#loc140) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc140) + } -> tensor<1x672x12x20xbf16> loc(#loc140) + %283 = xten_nn.subgraph (%arg5 = %282: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_206", OutputName = "Add_206"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc141) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc141) + } -> tensor<1x672x12x20xbf16> loc(#loc141) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc141) + } -> tensor<1x672x12x20xbf16> loc(#loc141) + %284 = xten_nn.subgraph (%arg5 = %283: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_209", + OutputName = "Clip_209", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + %285 = xten_nn.subgraph (%arg5 = %284: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_211", + OutputName = "Div_211", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + %286 = xten_nn.subgraph (%arg5 = %282: tensor<1x672x12x20xbf16>, %arg6 = %285: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_212", + OutputName = "Mul_212", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + %287 = xten_nn.subgraph (%arg5 = %286: tensor<1x672x12x20xbf16>, %arg6 = %76: tensor<672x1x3x3xbf16>, %arg7 = %75: tensor<672xbf16>) attributes { + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x3x3xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc145) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x672x1xbf16> loc(#loc145) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc145) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_213", + PartOfOutputName = "Conv_213", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<3x3x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc145) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + %288 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_215", OutputName = "Add_215"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + %289 = xten_nn.subgraph (%arg5 = %288: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_218", + OutputName = "Clip_218", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + %290 = xten_nn.subgraph (%arg5 = %289: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_220", + OutputName = "Div_220", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + %291 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>, %arg6 = %290: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_221", + OutputName = "Mul_221", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + %292 = xten_nn.subgraph (%arg5 = %291: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Generated-#30", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#31", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc335) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc335) + } -> tensor<1x672x1x240xbf16> loc(#loc335) + %293 = xten_nn.subgraph (%arg5 = %292: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#33", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc150) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc150) + } -> tensor<1x672x1x1xbf16> loc(#loc150) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc150) + } -> tensor<1x672x1x1xbf16> loc(#loc150) + %294 = xten_nn.subgraph (%arg5 = %293: tensor<1x672x1x1xbf16>, %arg6 = %74: tensor<168x672x1x1xbf16>, %arg7 = %73: tensor<168xbf16>) attributes { + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc336) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc336) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_223", + PartOfOutputName = "Conv_223", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc151) + %465 = tosa.clamp %464 { + LayerName = "Relu_224", + OutputName = "Relu_224", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc152) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc336) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc152) + } -> tensor<1x168x1x1xbf16> loc(#loc336) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc336) + } -> tensor<1x168x1x1xbf16> loc(#loc336) + %295 = xten_nn.subgraph (%arg5 = %294: tensor<1x168x1x1xbf16>, %arg6 = %72: tensor<672x168x1x1xbf16>, %arg7 = %71: tensor<672xbf16>) attributes { + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc153) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc153) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_225", + PartOfOutputName = "Conv_225", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc153) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc153) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc153) + } -> tensor<1x672x1x1xbf16> loc(#loc153) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc153) + } -> tensor<1x672x1x1xbf16> loc(#loc153) + %296 = xten_nn.subgraph (%arg5 = %295: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_227", OutputName = "Add_227"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc154) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc154) + } -> tensor<1x672x1x1xbf16> loc(#loc154) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc154) + } -> tensor<1x672x1x1xbf16> loc(#loc154) + %297 = xten_nn.subgraph (%arg5 = %296: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_230", + OutputName = "Clip_230", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + %298 = xten_nn.subgraph (%arg5 = %297: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_232", + OutputName = "Div_232", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + %299 = xten_nn.subgraph (%arg5 = %298: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Generated-#34", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#35", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + %300 = xten_nn.subgraph (%arg5 = %299: tensor<1x672x12x20xbf16>, %arg6 = %291: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_233", + OutputName = "Mul_233", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc157) + } -> tensor<1x672x12x20xbf16> loc(#loc157) + %301 = xten_nn.subgraph (%arg5 = %300: tensor<1x672x12x20xbf16>, %arg6 = %70: tensor<112x672x1x1xbf16>, %arg7 = %69: tensor<112xbf16>, %arg8 = %281: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Conv_234", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x672x12x20xbf16>, %arg10 = %arg6: tensor<112x672x1x1xbf16>, %arg11 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_234", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_234", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc158) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<112x672x1x1xbf16>) -> tensor<112x1x1x672xbf16> loc(#loc158) + %466 = tosa.transpose %arg9, %464 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc158) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_234", + PartOfOutputName = "Conv_234", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<112x1x1x672xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc158) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc158) + xten_nn.output %468 : tensor<1x112x12x20xbf16> loc(#loc158) + } -> tensor<1x112x12x20xbf16> loc(#loc158) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x112x12x20xbf16>, %arg10 = %arg8: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Add_235", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_235", OutputName = "Add_235"} : (tensor<1x112x12x20xbf16>, tensor<1x112x12x20xbf16>) -> tensor<1x112x12x20xbf16> loc(#loc159) + xten_nn.output %463 : tensor<1x112x12x20xbf16> loc(#loc159) + } -> tensor<1x112x12x20xbf16> loc(#loc159) + xten_nn.output %462 : tensor<1x112x12x20xbf16> loc(#loc159) + } -> tensor<1x112x12x20xbf16> loc(#loc337) + %302 = xten_nn.subgraph (%arg5 = %301: tensor<1x112x12x20xbf16>, %arg6 = %68: tensor<672x112x1x1xbf16>, %arg7 = %67: tensor<672xbf16>) attributes { + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc160) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc160) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc160) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_236", + PartOfOutputName = "Conv_236", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc160) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc160) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc160) + } -> tensor<1x672x12x20xbf16> loc(#loc160) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc160) + } -> tensor<1x672x12x20xbf16> loc(#loc160) + %303 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_238", OutputName = "Add_238"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc161) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc161) + } -> tensor<1x672x12x20xbf16> loc(#loc161) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc161) + } -> tensor<1x672x12x20xbf16> loc(#loc161) + %304 = xten_nn.subgraph (%arg5 = %303: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_241", + OutputName = "Clip_241", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + %305 = xten_nn.subgraph (%arg5 = %304: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_243", + OutputName = "Div_243", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + %306 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>, %arg6 = %305: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_244", + OutputName = "Mul_244", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + %307 = xten_nn.subgraph (%arg5 = %306: tensor<1x672x12x20xbf16>, %arg6 = %66: tensor<672x1x9x9xbf16>, %arg7 = %65: tensor<672xbf16>) attributes { + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x9x9xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc165) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x672x1xbf16> loc(#loc165) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc165) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_245", + PartOfOutputName = "Conv_245", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<9x9x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc165) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + %308 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_247", OutputName = "Add_247"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + %309 = xten_nn.subgraph (%arg5 = %308: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_250", + OutputName = "Clip_250", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + %310 = xten_nn.subgraph (%arg5 = %309: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_252", + OutputName = "Div_252", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + %311 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>, %arg6 = %310: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_253", + OutputName = "Mul_253", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + %312 = xten_nn.subgraph (%arg5 = %311: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Generated-#36", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#37", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc338) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc338) + } -> tensor<1x672x1x240xbf16> loc(#loc338) + %313 = xten_nn.subgraph (%arg5 = %312: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#38", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#39", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "Generated-#38", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#39", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc170) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc170) + } -> tensor<1x672x1x1xbf16> loc(#loc170) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc170) + } -> tensor<1x672x1x1xbf16> loc(#loc170) + %314 = xten_nn.subgraph (%arg5 = %313: tensor<1x672x1x1xbf16>, %arg6 = %64: tensor<168x672x1x1xbf16>, %arg7 = %63: tensor<168xbf16>) attributes { + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc339) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc339) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_255", + PartOfOutputName = "Conv_255", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc171) + %465 = tosa.clamp %464 { + LayerName = "Relu_256", + OutputName = "Relu_256", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc172) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc339) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc172) + } -> tensor<1x168x1x1xbf16> loc(#loc339) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc339) + } -> tensor<1x168x1x1xbf16> loc(#loc339) + %315 = xten_nn.subgraph (%arg5 = %314: tensor<1x168x1x1xbf16>, %arg6 = %62: tensor<672x168x1x1xbf16>, %arg7 = %61: tensor<672xbf16>) attributes { + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc173) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc173) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_257", + PartOfOutputName = "Conv_257", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc173) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc173) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc173) + } -> tensor<1x672x1x1xbf16> loc(#loc173) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc173) + } -> tensor<1x672x1x1xbf16> loc(#loc173) + %316 = xten_nn.subgraph (%arg5 = %315: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_259", OutputName = "Add_259"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc174) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc174) + } -> tensor<1x672x1x1xbf16> loc(#loc174) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc174) + } -> tensor<1x672x1x1xbf16> loc(#loc174) + %317 = xten_nn.subgraph (%arg5 = %316: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_262", + OutputName = "Clip_262", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + %318 = xten_nn.subgraph (%arg5 = %317: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_264", + OutputName = "Div_264", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + %319 = xten_nn.subgraph (%arg5 = %318: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Generated-#40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#41", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + %320 = xten_nn.subgraph (%arg5 = %319: tensor<1x672x12x20xbf16>, %arg6 = %311: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_265", + OutputName = "Mul_265", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc177) + } -> tensor<1x672x12x20xbf16> loc(#loc177) + %321 = xten_nn.subgraph (%arg5 = %320: tensor<1x672x12x20xbf16>, %arg6 = %60: tensor<160x672x1x1xbf16>, %arg7 = %59: tensor<160xbf16>) attributes { + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<160x672x1x1xbf16>, %arg10 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc178) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<160x672x1x1xbf16>) -> tensor<160x1x1x672xbf16> loc(#loc178) + %465 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc178) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_266", + PartOfOutputName = "Conv_266", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<160x1x1x672xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc178) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc178) + xten_nn.output %467 : tensor<1x160x12x20xbf16> loc(#loc178) + } -> tensor<1x160x12x20xbf16> loc(#loc178) + xten_nn.output %461 : tensor<1x160x12x20xbf16> loc(#loc178) + } -> tensor<1x160x12x20xbf16> loc(#loc178) + %322 = xten_nn.subgraph (%arg5 = %321: tensor<1x160x12x20xbf16>, %arg6 = %58: tensor<960x160x1x1xbf16>, %arg7 = %57: tensor<960xbf16>) attributes { + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc179) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc179) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc179) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_267", + PartOfOutputName = "Conv_267", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc179) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc179) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc179) + } -> tensor<1x960x12x20xbf16> loc(#loc179) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc179) + } -> tensor<1x960x12x20xbf16> loc(#loc179) + %323 = xten_nn.subgraph (%arg5 = %322: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_269", OutputName = "Add_269"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc180) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc180) + } -> tensor<1x960x12x20xbf16> loc(#loc180) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc180) + } -> tensor<1x960x12x20xbf16> loc(#loc180) + %324 = xten_nn.subgraph (%arg5 = %323: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_272", + OutputName = "Clip_272", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + %325 = xten_nn.subgraph (%arg5 = %324: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_274", + OutputName = "Div_274", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + %326 = xten_nn.subgraph (%arg5 = %322: tensor<1x960x12x20xbf16>, %arg6 = %325: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_275", + OutputName = "Mul_275", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + %327 = xten_nn.subgraph (%arg5 = %326: tensor<1x960x12x20xbf16>, %arg6 = %56: tensor<960x1x9x9xbf16>, %arg7 = %55: tensor<960xbf16>) attributes { + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc184) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc184) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc184) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_276", + PartOfOutputName = "Conv_276", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc184) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + %328 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_278", OutputName = "Add_278"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + %329 = xten_nn.subgraph (%arg5 = %328: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_281", + OutputName = "Clip_281", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + %330 = xten_nn.subgraph (%arg5 = %329: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_283", + OutputName = "Div_283", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + %331 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>, %arg6 = %330: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_284", + OutputName = "Mul_284", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + %332 = xten_nn.subgraph (%arg5 = %331: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Generated-#42", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#43", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc340) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc340) + } -> tensor<1x960x1x240xbf16> loc(#loc340) + %333 = xten_nn.subgraph (%arg5 = %332: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#44", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#45", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#44", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#45", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc189) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc189) + } -> tensor<1x960x1x1xbf16> loc(#loc189) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc189) + } -> tensor<1x960x1x1xbf16> loc(#loc189) + %334 = xten_nn.subgraph (%arg5 = %333: tensor<1x960x1x1xbf16>, %arg6 = %54: tensor<240x960x1x1xbf16>, %arg7 = %53: tensor<240xbf16>) attributes { + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc341) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc341) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_286", + PartOfOutputName = "Conv_286", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc190) + %465 = tosa.clamp %464 { + LayerName = "Relu_287", + OutputName = "Relu_287", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc191) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc341) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc191) + } -> tensor<1x240x1x1xbf16> loc(#loc341) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc341) + } -> tensor<1x240x1x1xbf16> loc(#loc341) + %335 = xten_nn.subgraph (%arg5 = %334: tensor<1x240x1x1xbf16>, %arg6 = %52: tensor<960x240x1x1xbf16>, %arg7 = %51: tensor<960xbf16>) attributes { + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc192) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc192) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_288", + PartOfOutputName = "Conv_288", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc192) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc192) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc192) + } -> tensor<1x960x1x1xbf16> loc(#loc192) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc192) + } -> tensor<1x960x1x1xbf16> loc(#loc192) + %336 = xten_nn.subgraph (%arg5 = %335: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_290", OutputName = "Add_290"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc193) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc193) + } -> tensor<1x960x1x1xbf16> loc(#loc193) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc193) + } -> tensor<1x960x1x1xbf16> loc(#loc193) + %337 = xten_nn.subgraph (%arg5 = %336: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_293", + OutputName = "Clip_293", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + %338 = xten_nn.subgraph (%arg5 = %337: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_295", + OutputName = "Div_295", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + %339 = xten_nn.subgraph (%arg5 = %338: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Generated-#46", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#47", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + %340 = xten_nn.subgraph (%arg5 = %339: tensor<1x960x12x20xbf16>, %arg6 = %331: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_296", + OutputName = "Mul_296", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc196) + } -> tensor<1x960x12x20xbf16> loc(#loc196) + %341 = xten_nn.subgraph (%arg5 = %340: tensor<1x960x12x20xbf16>, %arg6 = %50: tensor<160x960x1x1xbf16>, %arg7 = %49: tensor<160xbf16>, %arg8 = %321: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_297", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_297", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_297", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc197) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc197) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc197) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_297", + PartOfOutputName = "Conv_297", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc197) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc197) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc197) + } -> tensor<1x160x12x20xbf16> loc(#loc197) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_298", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_298", OutputName = "Add_298"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc198) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc198) + } -> tensor<1x160x12x20xbf16> loc(#loc198) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc198) + } -> tensor<1x160x12x20xbf16> loc(#loc342) + %342 = xten_nn.subgraph (%arg5 = %341: tensor<1x160x12x20xbf16>, %arg6 = %48: tensor<960x160x1x1xbf16>, %arg7 = %47: tensor<960xbf16>) attributes { + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc199) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc199) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc199) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_299", + PartOfOutputName = "Conv_299", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc199) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc199) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc199) + } -> tensor<1x960x12x20xbf16> loc(#loc199) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc199) + } -> tensor<1x960x12x20xbf16> loc(#loc199) + %343 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_301", OutputName = "Add_301"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc200) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc200) + } -> tensor<1x960x12x20xbf16> loc(#loc200) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc200) + } -> tensor<1x960x12x20xbf16> loc(#loc200) + %344 = xten_nn.subgraph (%arg5 = %343: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_304", + OutputName = "Clip_304", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + %345 = xten_nn.subgraph (%arg5 = %344: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_306", + OutputName = "Div_306", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + %346 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>, %arg6 = %345: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_307", + OutputName = "Mul_307", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + %347 = xten_nn.subgraph (%arg5 = %346: tensor<1x960x12x20xbf16>, %arg6 = %46: tensor<960x1x9x9xbf16>, %arg7 = %45: tensor<960xbf16>) attributes { + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc204) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc204) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc204) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_308", + PartOfOutputName = "Conv_308", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc204) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + %348 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_310", OutputName = "Add_310"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + %349 = xten_nn.subgraph (%arg5 = %348: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_313", + OutputName = "Clip_313", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + %350 = xten_nn.subgraph (%arg5 = %349: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_315", + OutputName = "Div_315", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + %351 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>, %arg6 = %350: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_316", + OutputName = "Mul_316", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + %352 = xten_nn.subgraph (%arg5 = %351: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Generated-#48", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#49", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc343) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc343) + } -> tensor<1x960x1x240xbf16> loc(#loc343) + %353 = xten_nn.subgraph (%arg5 = %352: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#51", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#51", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc209) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc209) + } -> tensor<1x960x1x1xbf16> loc(#loc209) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc209) + } -> tensor<1x960x1x1xbf16> loc(#loc209) + %354 = xten_nn.subgraph (%arg5 = %353: tensor<1x960x1x1xbf16>, %arg6 = %44: tensor<240x960x1x1xbf16>, %arg7 = %43: tensor<240xbf16>) attributes { + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc344) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc344) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_318", + PartOfOutputName = "Conv_318", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc210) + %465 = tosa.clamp %464 { + LayerName = "Relu_319", + OutputName = "Relu_319", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc211) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc344) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc211) + } -> tensor<1x240x1x1xbf16> loc(#loc344) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc344) + } -> tensor<1x240x1x1xbf16> loc(#loc344) + %355 = xten_nn.subgraph (%arg5 = %354: tensor<1x240x1x1xbf16>, %arg6 = %42: tensor<960x240x1x1xbf16>, %arg7 = %41: tensor<960xbf16>) attributes { + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc212) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc212) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_320", + PartOfOutputName = "Conv_320", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc212) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc212) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc212) + } -> tensor<1x960x1x1xbf16> loc(#loc212) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc212) + } -> tensor<1x960x1x1xbf16> loc(#loc212) + %356 = xten_nn.subgraph (%arg5 = %355: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_322", OutputName = "Add_322"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc213) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc213) + } -> tensor<1x960x1x1xbf16> loc(#loc213) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc213) + } -> tensor<1x960x1x1xbf16> loc(#loc213) + %357 = xten_nn.subgraph (%arg5 = %356: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_325", + OutputName = "Clip_325", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + %358 = xten_nn.subgraph (%arg5 = %357: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_327", + OutputName = "Div_327", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + %359 = xten_nn.subgraph (%arg5 = %358: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Generated-#52", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#53", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + %360 = xten_nn.subgraph (%arg5 = %359: tensor<1x960x12x20xbf16>, %arg6 = %351: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_328", + OutputName = "Mul_328", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc216) + } -> tensor<1x960x12x20xbf16> loc(#loc216) + %361 = xten_nn.subgraph (%arg5 = %360: tensor<1x960x12x20xbf16>, %arg6 = %40: tensor<160x960x1x1xbf16>, %arg7 = %39: tensor<160xbf16>, %arg8 = %341: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_329", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_329", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_329", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc217) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc217) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc217) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_329", + PartOfOutputName = "Conv_329", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc217) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc217) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc217) + } -> tensor<1x160x12x20xbf16> loc(#loc217) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_330", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_330", OutputName = "Add_330"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc218) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc218) + } -> tensor<1x160x12x20xbf16> loc(#loc218) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc218) + } -> tensor<1x160x12x20xbf16> loc(#loc345) + %362 = xten_nn.subgraph (%arg5 = %361: tensor<1x160x12x20xbf16>, %arg6 = %38: tensor<960x160x1x1xbf16>, %arg7 = %37: tensor<960xbf16>) attributes { + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc219) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc219) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc219) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_331", + PartOfOutputName = "Conv_331", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc219) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc219) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc219) + } -> tensor<1x960x12x20xbf16> loc(#loc219) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc219) + } -> tensor<1x960x12x20xbf16> loc(#loc219) + %363 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_333", OutputName = "Add_333"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc220) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc220) + } -> tensor<1x960x12x20xbf16> loc(#loc220) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc220) + } -> tensor<1x960x12x20xbf16> loc(#loc220) + %364 = xten_nn.subgraph (%arg5 = %363: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_336", + OutputName = "Clip_336", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + %365 = xten_nn.subgraph (%arg5 = %364: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_338", + OutputName = "Div_338", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + %366 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>, %arg6 = %365: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_339", + OutputName = "Mul_339", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + %367 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Generated-#54", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Generated-#55", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc346) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc346) + } -> tensor<1x960x1x240xbf16> loc(#loc346) + %368 = xten_nn.subgraph (%arg5 = %367: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "Generated-#56", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "Generated-#57", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc224) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc224) + } -> tensor<1x960x1x1xbf16> loc(#loc224) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc224) + } -> tensor<1x960x1x1xbf16> loc(#loc224) + %369 = xten_nn.subgraph (%arg5 = %368: tensor<1x960x1x1xbf16>, %arg6 = %36: tensor<128x960x1x1xbf16>, %arg7 = %35: tensor<128xbf16>) attributes { + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<128x960x1x1xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc225) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc225) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_343", + PartOfOutputName = "Conv_343", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x1x1x128xbf16> loc(#loc225) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x128xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc225) + xten_nn.output %465 : tensor<1x128x1x1xbf16> loc(#loc225) + } -> tensor<1x128x1x1xbf16> loc(#loc225) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc225) + } -> tensor<1x128x1x1xbf16> loc(#loc225) + %370 = xten_nn.subgraph (%arg5 = %369: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_344", OutputName = "Sigmoid_344"} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc226) + xten_nn.output %462 : tensor<1x128x1x1xbf16> loc(#loc226) + } -> tensor<1x128x1x1xbf16> loc(#loc226) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc226) + } -> tensor<1x128x1x1xbf16> loc(#loc226) + %371 = xten_nn.subgraph (%arg5 = %370: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Generated-#58", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Generated-#59", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 128 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc227) + %372 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>, %arg6 = %34: tensor<128x960x1x1xbf16>, %arg7 = %33: tensor<128xbf16>, %arg8 = %371: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Conv_340", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<128x960x1x1xbf16>, %arg11 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_340", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_341", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc348) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc348) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc348) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_340", + PartOfOutputName = "Conv_340", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc228) + %468 = tosa.clamp %467 { + LayerName = "Relu_341", + OutputName = "Relu_341", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x12x20x128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc229) + %469 = tosa.transpose %468, %463 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc348) + xten_nn.output %469 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc348) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x128x12x20xbf16>, %arg10 = %arg8: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Mul_345", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.mul %arg9, %arg10 { + LayerName = "Mul_345", + OutputName = "Mul_345", + shift = 0 : i8} : (tensor<1x128x12x20xbf16>, tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %463 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc227) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc227) + } -> tensor<1x128x12x20xbf16> loc(#loc347) + %373 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc230) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc230) + } -> tensor<1x64x12x20xbf16> loc(#loc230) + %374 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc230) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc230) + } -> tensor<1x64x12x20xbf16> loc(#loc230) + %375 = xten_nn.subgraph (%arg5 = %374: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_350", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_350", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_350", + OutputName = "Concat_350", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc231) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc231) + } -> tensor<1x128x12x20xbf16> loc(#loc231) + %376 = xten_nn.subgraph (%arg5 = %375: tensor<1x128x12x20xbf16>, %arg6 = %32: tensor<128x128x3x3xbf16>, %arg7 = %31: tensor<128xbf16>) attributes { + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<128x128x3x3xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<128x128x3x3xbf16>, tensor<4xi32>) -> tensor<128x3x3x128xbf16> loc(#loc232) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc232) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_351", + PartOfOutputName = "Conv_351", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<128x3x3x128xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc232) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc232) + xten_nn.output %467 : tensor<1x128x12x20xbf16> loc(#loc232) + } -> tensor<1x128x12x20xbf16> loc(#loc232) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc232) + } -> tensor<1x128x12x20xbf16> loc(#loc232) + %377 = xten_nn.subgraph (%arg5 = %376: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_352", OutputName = "Sigmoid_352"} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + %378 = xten_nn.subgraph (%arg5 = %377: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc234) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc234) + } -> tensor<1x64x12x20xbf16> loc(#loc234) + %379 = xten_nn.subgraph (%arg5 = %377: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc234) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc234) + } -> tensor<1x64x12x20xbf16> loc(#loc234) + %380 = xten_nn.subgraph (%arg5 = %378: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_354", + OutputName = "Mul_354", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc235) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc235) + } -> tensor<1x64x12x20xbf16> loc(#loc235) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc235) + } -> tensor<1x64x12x20xbf16> loc(#loc235) + %381 = xten_nn.subgraph (%arg5 = %374: tensor<1x64x12x20xbf16>, %arg6 = %380: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_355", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_355", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_355", + OutputName = "Concat_355", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc236) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc236) + } -> tensor<1x128x12x20xbf16> loc(#loc236) + %382 = xten_nn.subgraph (%arg5 = %381: tensor<1x128x12x20xbf16>, %arg6 = %30: tensor<64x128x3x3xbf16>, %arg7 = %29: tensor<64xbf16>) attributes { + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<64x128x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<64x128x3x3xbf16>, tensor<4xi32>) -> tensor<64x3x3x128xbf16> loc(#loc237) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc237) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_356", + PartOfOutputName = "Conv_356", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<64x3x3x128xbf16>, tensor<64xbf16>) -> tensor<1x12x20x64xbf16> loc(#loc237) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x64xbf16>, tensor<4xi32>) -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %467 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + %383 = xten_nn.subgraph (%arg5 = %382: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_357", OutputName = "Tanh_357"} : (tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + %384 = xten_nn.subgraph (%arg5 = %379: tensor<1x64x12x20xbf16>, %arg6 = %383: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_361", + OutputName = "Mul_361", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc239) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc239) + } -> tensor<1x64x12x20xbf16> loc(#loc239) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc239) + } -> tensor<1x64x12x20xbf16> loc(#loc239) + %385 = xten_nn.subgraph (%arg5 = %28: tensor<1x64x12x20xbf16>, %arg6 = %379: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_359", OutputName = "Sub_359"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + %386 = xten_nn.subgraph (%arg5 = %385: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_360", + OutputName = "Mul_360", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + %387 = xten_nn.subgraph (%arg5 = %386: tensor<1x64x12x20xbf16>, %arg6 = %384: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_362", OutputName = "Add_362"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + %388 = xten_nn.subgraph (%arg5 = %373: tensor<1x64x12x20xbf16>, %arg6 = %387: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_363", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_363", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_363", + OutputName = "Concat_363", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc242) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc242) + } -> tensor<1x128x12x20xbf16> loc(#loc242) + %389 = xten_nn.subgraph (%arg5 = %388: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Resize_365", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Resize_365", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 128 : ui32, + config.dim_2 = 12 : ui32, + config.dim_3 = 20 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 24 : ui32, + config.output_W = 40 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_365", + OutputName = "Resize_365", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x24x40xbf16> loc(#loc243) + xten_nn.output %461 : tensor<1x128x24x40xbf16> loc(#loc243) + } -> tensor<1x128x24x40xbf16> loc(#loc243) + %390 = xten_nn.subgraph (%arg5 = %389: tensor<1x128x24x40xbf16>) attributes { + LayerName = "Slice_371", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + OutputName = "Slice_371", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 128 : ui32, + config.dim_h = 24 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 23 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_371", + OutputName = "Slice_371", + size = array, + start = array} : (tensor<1x128x24x40xbf16>) -> tensor<1x128x23x40xbf16> loc(#loc244) + xten_nn.output %461 : tensor<1x128x23x40xbf16> loc(#loc244) + } -> tensor<1x128x23x40xbf16> loc(#loc244) + %391 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>) attributes { + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc12) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc12) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_346", + PartOfOutputName = "AveragePool_346", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>) -> tensor<1x90x160x3xbf16> loc(#loc12) + %466 = tosa.transpose %465, %462 : (tensor<1x90x160x3xbf16>, tensor<4xi32>) -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %466 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %461 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + %392 = xten_nn.subgraph (%arg5 = %391: tensor<1x3x90x160xbf16>) attributes { + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x90x160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc245) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x3xbf16> loc(#loc245) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_347", + PartOfOutputName = "AveragePool_347", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x90x160x3xbf16>) -> tensor<1x45x80x3xbf16> loc(#loc245) + %466 = tosa.transpose %465, %462 : (tensor<1x45x80x3xbf16>, tensor<4xi32>) -> tensor<1x3x45x80xbf16> loc(#loc245) + xten_nn.output %466 : tensor<1x3x45x80xbf16> loc(#loc245) + } -> tensor<1x3x45x80xbf16> loc(#loc245) + xten_nn.output %461 : tensor<1x3x45x80xbf16> loc(#loc245) + } -> tensor<1x3x45x80xbf16> loc(#loc245) + %393 = xten_nn.subgraph (%arg5 = %392: tensor<1x3x45x80xbf16>) attributes { + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x45x80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 1], [0, 0]], + HWPaddingNotCounted = [[0, 1], [0, 0]], + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc246) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x3xbf16> loc(#loc246) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_348", + PartOfOutputName = "AveragePool_348", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x45x80x3xbf16>) -> tensor<1x23x40x3xbf16> loc(#loc246) + %466 = tosa.transpose %465, %462 : (tensor<1x23x40x3xbf16>, tensor<4xi32>) -> tensor<1x3x23x40xbf16> loc(#loc246) + xten_nn.output %466 : tensor<1x3x23x40xbf16> loc(#loc246) + } -> tensor<1x3x23x40xbf16> loc(#loc246) + xten_nn.output %461 : tensor<1x3x23x40xbf16> loc(#loc246) + } -> tensor<1x3x23x40xbf16> loc(#loc246) + %394 = xten_nn.subgraph (%arg5 = %390: tensor<1x128x23x40xbf16>, %arg6 = %217: tensor<1x40x23x40xbf16>, %arg7 = %393: tensor<1x3x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_372", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_372", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_372", + OutputName = "Concat_372", + axis = 1 : i32} : (tensor<1x128x23x40xbf16>, tensor<1x40x23x40xbf16>, tensor<1x3x23x40xbf16>) -> tensor<1x171x23x40xbf16> loc(#loc247) + xten_nn.output %461 : tensor<1x171x23x40xbf16> loc(#loc247) + } -> tensor<1x171x23x40xbf16> loc(#loc247) + %395 = xten_nn.subgraph (%arg5 = %394: tensor<1x171x23x40xbf16>, %arg6 = %27: tensor<80x171x3x3xbf16>, %arg7 = %26: tensor<80xbf16>) attributes { + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x171x23x40xbf16>, %arg9 = %arg6: tensor<80x171x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x171x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x171xbf16> loc(#loc349) + %465 = tosa.transpose %arg8, %463 : (tensor<1x171x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x171xbf16> loc(#loc349) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_373", + PartOfOutputName = "Conv_373", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x171xbf16>, tensor<80x3x3x171xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc248) + %467 = tosa.clamp %466 { + LayerName = "Relu_374", + OutputName = "Relu_374", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc249) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc349) + xten_nn.output %468 : tensor<1x80x23x40xbf16> loc(#loc249) + } -> tensor<1x80x23x40xbf16> loc(#loc349) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc349) + } -> tensor<1x80x23x40xbf16> loc(#loc349) + %396 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %397 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %398 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_376", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_376", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_376", + OutputName = "Concat_376", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc251) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc251) + } -> tensor<1x80x23x40xbf16> loc(#loc251) + %399 = xten_nn.subgraph (%arg5 = %398: tensor<1x80x23x40xbf16>, %arg6 = %25: tensor<80x80x3x3xbf16>, %arg7 = %24: tensor<80xbf16>) attributes { + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<80x80x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x80x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x80xbf16> loc(#loc252) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_377", + PartOfOutputName = "Conv_377", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<80x3x3x80xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %467 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + %400 = xten_nn.subgraph (%arg5 = %399: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_378", OutputName = "Sigmoid_378"} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %462 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + %401 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %402 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %403 = xten_nn.subgraph (%arg5 = %401: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_380", + OutputName = "Mul_380", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + %404 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %403: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_381", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_381", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_381", + OutputName = "Concat_381", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc256) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc256) + } -> tensor<1x80x23x40xbf16> loc(#loc256) + %405 = xten_nn.subgraph (%arg5 = %404: tensor<1x80x23x40xbf16>, %arg6 = %23: tensor<40x80x3x3xbf16>, %arg7 = %22: tensor<40xbf16>) attributes { + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<40x80x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x80x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x80xbf16> loc(#loc257) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc257) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_382", + PartOfOutputName = "Conv_382", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<40x3x3x80xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc257) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc257) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc257) + } -> tensor<1x40x23x40xbf16> loc(#loc257) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc257) + } -> tensor<1x40x23x40xbf16> loc(#loc257) + %406 = xten_nn.subgraph (%arg5 = %405: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_383", OutputName = "Tanh_383"} : (tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + %407 = xten_nn.subgraph (%arg5 = %402: tensor<1x40x23x40xbf16>, %arg6 = %406: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_387", + OutputName = "Mul_387", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + %408 = xten_nn.subgraph (%arg5 = %21: tensor<1x40x23x40xbf16>, %arg6 = %402: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_385", OutputName = "Sub_385"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + %409 = xten_nn.subgraph (%arg5 = %408: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_386", + OutputName = "Mul_386", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + %410 = xten_nn.subgraph (%arg5 = %409: tensor<1x40x23x40xbf16>, %arg6 = %407: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_388", OutputName = "Add_388"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + %411 = xten_nn.subgraph (%arg5 = %396: tensor<1x40x23x40xbf16>, %arg6 = %410: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_389", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_389", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_389", + OutputName = "Concat_389", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc262) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc262) + } -> tensor<1x80x23x40xbf16> loc(#loc262) + %412 = xten_nn.subgraph (%arg5 = %411: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Resize_391", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Resize_391", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 80 : ui32, + config.dim_2 = 23 : ui32, + config.dim_3 = 40 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 46 : ui32, + config.output_W = 80 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_391", + OutputName = "Resize_391", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x46x80xbf16> loc(#loc263) + xten_nn.output %461 : tensor<1x80x46x80xbf16> loc(#loc263) + } -> tensor<1x80x46x80xbf16> loc(#loc263) + %413 = xten_nn.subgraph (%arg5 = %412: tensor<1x80x46x80xbf16>) attributes { + LayerName = "Slice_397", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + OutputName = "Slice_397", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 80 : ui32, + config.dim_h = 46 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 45 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_397", + OutputName = "Slice_397", + size = array, + start = array} : (tensor<1x80x46x80xbf16>) -> tensor<1x80x45x80xbf16> loc(#loc264) + xten_nn.output %461 : tensor<1x80x45x80xbf16> loc(#loc264) + } -> tensor<1x80x45x80xbf16> loc(#loc264) + %414 = xten_nn.subgraph (%arg5 = %413: tensor<1x80x45x80xbf16>, %arg6 = %181: tensor<1x24x45x80xbf16>, %arg7 = %392: tensor<1x3x45x80xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_398", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_398", + OutputName = "Concat_398", + axis = 1 : i32} : (tensor<1x80x45x80xbf16>, tensor<1x24x45x80xbf16>, tensor<1x3x45x80xbf16>) -> tensor<1x107x45x80xbf16> loc(#loc265) + xten_nn.output %461 : tensor<1x107x45x80xbf16> loc(#loc265) + } -> tensor<1x107x45x80xbf16> loc(#loc265) + %415 = xten_nn.subgraph (%arg5 = %414: tensor<1x107x45x80xbf16>, %arg6 = %20: tensor<40x107x3x3xbf16>, %arg7 = %19: tensor<40xbf16>) attributes { + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x107x45x80xbf16>, %arg9 = %arg6: tensor<40x107x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x107x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x107xbf16> loc(#loc350) + %465 = tosa.transpose %arg8, %463 : (tensor<1x107x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x107xbf16> loc(#loc350) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_399", + PartOfOutputName = "Conv_399", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x107xbf16>, tensor<40x3x3x107xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc266) + %467 = tosa.clamp %466 { + LayerName = "Relu_400", + OutputName = "Relu_400", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc267) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc350) + xten_nn.output %468 : tensor<1x40x45x80xbf16> loc(#loc267) + } -> tensor<1x40x45x80xbf16> loc(#loc350) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc350) + } -> tensor<1x40x45x80xbf16> loc(#loc350) + %416 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %417 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %418 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_402", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_402", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_402", + OutputName = "Concat_402", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc269) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc269) + } -> tensor<1x40x45x80xbf16> loc(#loc269) + %419 = xten_nn.subgraph (%arg5 = %418: tensor<1x40x45x80xbf16>, %arg6 = %18: tensor<40x40x3x3xbf16>, %arg7 = %17: tensor<40xbf16>) attributes { + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<40x40x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x40x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x40xbf16> loc(#loc270) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_403", + PartOfOutputName = "Conv_403", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<40x3x3x40xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %467 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + %420 = xten_nn.subgraph (%arg5 = %419: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_404", OutputName = "Sigmoid_404"} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %462 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + %421 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %422 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %423 = xten_nn.subgraph (%arg5 = %421: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_406", + OutputName = "Mul_406", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + %424 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %423: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_407", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_407", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_407", + OutputName = "Concat_407", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc274) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc274) + } -> tensor<1x40x45x80xbf16> loc(#loc274) + %425 = xten_nn.subgraph (%arg5 = %424: tensor<1x40x45x80xbf16>, %arg6 = %16: tensor<20x40x3x3xbf16>, %arg7 = %15: tensor<20xbf16>) attributes { + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<20x40x3x3xbf16>, %arg10 = %arg7: tensor<20xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<20x40x3x3xbf16>, tensor<4xi32>) -> tensor<20x3x3x40xbf16> loc(#loc275) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc275) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_408", + PartOfOutputName = "Conv_408", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<20x3x3x40xbf16>, tensor<20xbf16>) -> tensor<1x45x80x20xbf16> loc(#loc275) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x20xbf16>, tensor<4xi32>) -> tensor<1x20x45x80xbf16> loc(#loc275) + xten_nn.output %467 : tensor<1x20x45x80xbf16> loc(#loc275) + } -> tensor<1x20x45x80xbf16> loc(#loc275) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc275) + } -> tensor<1x20x45x80xbf16> loc(#loc275) + %426 = xten_nn.subgraph (%arg5 = %425: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_409", OutputName = "Tanh_409"} : (tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + %427 = xten_nn.subgraph (%arg5 = %422: tensor<1x20x45x80xbf16>, %arg6 = %426: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_413", + OutputName = "Mul_413", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + %428 = xten_nn.subgraph (%arg5 = %14: tensor<1x20x45x80xbf16>, %arg6 = %422: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_411", OutputName = "Sub_411"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + %429 = xten_nn.subgraph (%arg5 = %428: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_412", + OutputName = "Mul_412", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + %430 = xten_nn.subgraph (%arg5 = %429: tensor<1x20x45x80xbf16>, %arg6 = %427: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_414", OutputName = "Add_414"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + %431 = xten_nn.subgraph (%arg5 = %416: tensor<1x20x45x80xbf16>, %arg6 = %430: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_415", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_415", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_415", + OutputName = "Concat_415", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc280) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc280) + } -> tensor<1x40x45x80xbf16> loc(#loc280) + %432 = xten_nn.subgraph (%arg5 = %431: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Resize_417", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Resize_417", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 40 : ui32, + config.dim_2 = 45 : ui32, + config.dim_3 = 80 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 90 : ui32, + config.output_W = 160 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_417", + OutputName = "Resize_417", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x90x160xbf16> loc(#loc281) + xten_nn.output %461 : tensor<1x40x90x160xbf16> loc(#loc281) + } -> tensor<1x40x90x160xbf16> loc(#loc281) + %433 = xten_nn.subgraph (%arg5 = %432: tensor<1x40x90x160xbf16>, %arg6 = %175: tensor<1x16x90x160xbf16>, %arg7 = %391: tensor<1x3x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_418", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_418", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_418", + OutputName = "Concat_418", + axis = 1 : i32} : (tensor<1x40x90x160xbf16>, tensor<1x16x90x160xbf16>, tensor<1x3x90x160xbf16>) -> tensor<1x59x90x160xbf16> loc(#loc282) + xten_nn.output %461 : tensor<1x59x90x160xbf16> loc(#loc282) + } -> tensor<1x59x90x160xbf16> loc(#loc282) + %434 = xten_nn.subgraph (%arg5 = %433: tensor<1x59x90x160xbf16>, %arg6 = %13: tensor<32x59x3x3xbf16>, %arg7 = %12: tensor<32xbf16>) attributes { + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x59x90x160xbf16>, %arg9 = %arg6: tensor<32x59x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x59x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x59xbf16> loc(#loc351) + %465 = tosa.transpose %arg8, %463 : (tensor<1x59x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x59xbf16> loc(#loc351) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_419", + PartOfOutputName = "Conv_419", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x59xbf16>, tensor<32x3x3x59xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc283) + %467 = tosa.clamp %466 { + LayerName = "Relu_420", + OutputName = "Relu_420", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc284) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc351) + xten_nn.output %468 : tensor<1x32x90x160xbf16> loc(#loc284) + } -> tensor<1x32x90x160xbf16> loc(#loc351) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc351) + } -> tensor<1x32x90x160xbf16> loc(#loc351) + %435 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %436 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %437 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_422", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_422", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_422", + OutputName = "Concat_422", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc286) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc286) + } -> tensor<1x32x90x160xbf16> loc(#loc286) + %438 = xten_nn.subgraph (%arg5 = %437: tensor<1x32x90x160xbf16>, %arg6 = %11: tensor<32x32x3x3xbf16>, %arg7 = %10: tensor<32xbf16>) attributes { + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<32x32x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x32x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x32xbf16> loc(#loc287) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_423", + PartOfOutputName = "Conv_423", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<32x3x3x32xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %467 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + %439 = xten_nn.subgraph (%arg5 = %438: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_424", OutputName = "Sigmoid_424"} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %462 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + %440 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %441 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %442 = xten_nn.subgraph (%arg5 = %440: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_426", + OutputName = "Mul_426", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + %443 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %442: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_427", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_427", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_427", + OutputName = "Concat_427", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc291) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc291) + } -> tensor<1x32x90x160xbf16> loc(#loc291) + %444 = xten_nn.subgraph (%arg5 = %443: tensor<1x32x90x160xbf16>, %arg6 = %9: tensor<16x32x3x3xbf16>, %arg7 = %8: tensor<16xbf16>) attributes { + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<16x32x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x32x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x32xbf16> loc(#loc292) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc292) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_428", + PartOfOutputName = "Conv_428", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<16x3x3x32xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc292) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc292) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc292) + } -> tensor<1x16x90x160xbf16> loc(#loc292) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc292) + } -> tensor<1x16x90x160xbf16> loc(#loc292) + %445 = xten_nn.subgraph (%arg5 = %444: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_429", OutputName = "Tanh_429"} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + %446 = xten_nn.subgraph (%arg5 = %441: tensor<1x16x90x160xbf16>, %arg6 = %445: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_433", + OutputName = "Mul_433", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + %447 = xten_nn.subgraph (%arg5 = %7: tensor<1x16x90x160xbf16>, %arg6 = %441: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_431", OutputName = "Sub_431"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + %448 = xten_nn.subgraph (%arg5 = %447: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_432", + OutputName = "Mul_432", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + %449 = xten_nn.subgraph (%arg5 = %448: tensor<1x16x90x160xbf16>, %arg6 = %446: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_434", OutputName = "Add_434"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + %450 = xten_nn.subgraph (%arg5 = %435: tensor<1x16x90x160xbf16>, %arg6 = %449: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_435", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_435", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_435", + OutputName = "Concat_435", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc297) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc297) + } -> tensor<1x32x90x160xbf16> loc(#loc297) + %451 = xten_nn.subgraph (%arg5 = %450: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Resize_437", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Resize_437", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 32 : ui32, + config.dim_2 = 90 : ui32, + config.dim_3 = 160 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 180 : ui32, + config.output_W = 320 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_437", + OutputName = "Resize_437", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x180x320xbf16> loc(#loc298) + xten_nn.output %461 : tensor<1x32x180x320xbf16> loc(#loc298) + } -> tensor<1x32x180x320xbf16> loc(#loc298) + %452 = xten_nn.subgraph (%arg5 = %451: tensor<1x32x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_438", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Concat_438", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_438", + OutputName = "Concat_438", + axis = 1 : i32} : (tensor<1x32x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x35x180x320xbf16> loc(#loc299) + xten_nn.output %461 : tensor<1x35x180x320xbf16> loc(#loc299) + } -> tensor<1x35x180x320xbf16> loc(#loc299) + %453 = xten_nn.subgraph (%arg5 = %452: tensor<1x35x180x320xbf16>, %arg6 = %6: tensor<16x35x3x3xbf16>, %arg7 = %5: tensor<16xbf16>) attributes { + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x35x180x320xbf16>, %arg9 = %arg6: tensor<16x35x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x35x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x35xbf16> loc(#loc352) + %465 = tosa.transpose %arg8, %463 : (tensor<1x35x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x35xbf16> loc(#loc352) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_439", + PartOfOutputName = "Conv_439", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x35xbf16>, tensor<16x3x3x35xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc300) + %467 = tosa.clamp %466 { + LayerName = "Relu_440", + OutputName = "Relu_440", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc301) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc352) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc301) + } -> tensor<1x16x180x320xbf16> loc(#loc352) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc352) + } -> tensor<1x16x180x320xbf16> loc(#loc352) + %454 = xten_nn.subgraph (%arg5 = %453: tensor<1x16x180x320xbf16>, %arg6 = %4: tensor<16x16x3x3xbf16>, %arg7 = %3: tensor<16xbf16>) attributes { + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<16x16x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x16x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x16xbf16> loc(#loc353) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc353) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_441", + PartOfOutputName = "Conv_441", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<16x3x3x16xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc302) + %467 = tosa.clamp %466 { + LayerName = "Relu_442", + OutputName = "Relu_442", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc303) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc353) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc303) + } -> tensor<1x16x180x320xbf16> loc(#loc353) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc353) + } -> tensor<1x16x180x320xbf16> loc(#loc353) + %455 = xten_nn.subgraph (%arg5 = %454: tensor<1x16x180x320xbf16>, %arg6 = %2: tensor<4x16x1x1xbf16>, %arg7 = %1: tensor<4xbf16>) attributes { + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<4x16x1x1xbf16>, %arg10 = %arg7: tensor<4xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc304) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<4x16x1x1xbf16>) -> tensor<4x1x1x16xbf16> loc(#loc304) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc304) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_443", + PartOfOutputName = "Conv_443", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<4x1x1x16xbf16>, tensor<4xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc304) + %467 = tosa.transpose %466, %462 : (tensor<1x180x320x4xbf16>, tensor<4xi32>) -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %467 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %461 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + %456 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc305) + } -> tensor<1x3x180x320xbf16> loc(#loc305) + %457 = xten_nn.subgraph (%arg5 = %456: tensor<1x3x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_445", OutputName = "Add_445"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + %458 = xten_nn.subgraph (%arg5 = %457: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_446", + OutputName = "Clip_446", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + %459 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 4 : ui32, + config.start = 3 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc305) + } -> tensor<1x1x180x320xbf16> loc(#loc305) + %460 = xten_nn.subgraph (%arg5 = %459: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_447", + OutputName = "Clip_447", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %462 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + return %449, %430, %410, %387, %458, %460 : tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16> loc(#loc308) + } loc(#loc308) + func.func @forward(%arg0: tensor<1x180x320x4xbf16> {onnx.name = "385"} loc(unknown), %arg1: tensor<1x16x90x160xbf16> {onnx.name = "394"} loc(unknown), %arg2: tensor<1x20x45x80xbf16> {onnx.name = "395"} loc(unknown), %arg3: tensor<1x40x23x40xbf16> {onnx.name = "396"} loc(unknown), %arg4: tensor<1x64x12x20xbf16> {onnx.name = "397"} loc(unknown)) -> (tensor<1x1x180x320xbf16> {onnx.name = "921"}, tensor<1x16x90x160xbf16> {onnx.name = "894"}, tensor<1x20x45x80xbf16> {onnx.name = "868"}, tensor<1x40x23x40xbf16> {onnx.name = "832"}, tensor<1x64x12x20xbf16> {onnx.name = "796"}, tensor<1x3x180x320xbf16> {onnx.name = "916"}) { + %0:6 = call @forward_outlined_part_0(%arg0, %arg1, %arg2, %arg3, %arg4) : (tensor<1x180x320x4xbf16>, tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) loc(#loc308) + return %0#5, %0#0, %0#1, %0#2, %0#3, %0#4 : tensor<1x1x180x320xbf16>, tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16> loc(#loc) + } loc(#loc) +} loc(#loc) +#loc1 = loc("Div_2") +#loc2 = loc("Sub_431") +#loc3 = loc("Sub_411") +#loc4 = loc("Sub_385") +#loc5 = loc("Sub_359") +#loc6 = loc("Div_16") +#loc7 = loc("Sub_14") +#loc8 = loc("Initializer_398") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("Conv_17") +#loc14 = loc("Add_19") +#loc15 = loc("Clip_22") +#loc16 = loc("Div_24") +#loc17 = loc("Mul_25") +#loc18 = loc("Conv_26") +#loc19 = loc("Relu_27") +#loc20 = loc("Conv_28") +#loc21 = loc("Add_29") +#loc22 = loc("Conv_30") +#loc23 = loc("Relu_31") +#loc24 = loc("Conv_32") +#loc25 = loc("Relu_33") +#loc26 = loc("Conv_34") +#loc27 = loc("Conv_35") +#loc28 = loc("Relu_36") +#loc29 = loc("Conv_37") +#loc30 = loc("Relu_38") +#loc31 = loc("Conv_39") +#loc32 = loc("Add_40") +#loc33 = loc("Conv_41") +#loc34 = loc("Relu_42") +#loc35 = loc("Conv_43") +#loc36 = loc("Relu_44") +#loc37 = loc("GlobalAveragePool_45") +#loc38 = loc("Conv_46") +#loc39 = loc("Relu_47") +#loc40 = loc("Conv_48") +#loc41 = loc("Add_50") +#loc42 = loc("Clip_53") +#loc43 = loc("Div_55") +#loc44 = loc("Mul_56") +#loc45 = loc("Conv_57") +#loc46 = loc("Conv_58") +#loc47 = loc("Relu_59") +#loc48 = loc("Conv_60") +#loc49 = loc("Relu_61") +#loc50 = loc("GlobalAveragePool_62") +#loc51 = loc("Conv_63") +#loc52 = loc("Relu_64") +#loc53 = loc("Conv_65") +#loc54 = loc("Add_67") +#loc55 = loc("Clip_70") +#loc56 = loc("Div_72") +#loc57 = loc("Mul_73") +#loc58 = loc("Conv_74") +#loc59 = loc("Add_75") +#loc60 = loc("Conv_76") +#loc61 = loc("Relu_77") +#loc62 = loc("Conv_78") +#loc63 = loc("Relu_79") +#loc64 = loc("GlobalAveragePool_80") +#loc65 = loc("Conv_81") +#loc66 = loc("Relu_82") +#loc67 = loc("Conv_83") +#loc68 = loc("Add_85") +#loc69 = loc("Clip_88") +#loc70 = loc("Div_90") +#loc71 = loc("Mul_91") +#loc72 = loc("Conv_92") +#loc73 = loc("Add_93") +#loc74 = loc("Conv_94") +#loc75 = loc("Add_96") +#loc76 = loc("Clip_99") +#loc77 = loc("Div_101") +#loc78 = loc("Mul_102") +#loc79 = loc("Conv_103") +#loc80 = loc("Add_105") +#loc81 = loc("Clip_108") +#loc82 = loc("Div_110") +#loc83 = loc("Mul_111") +#loc84 = loc("Conv_112") +#loc85 = loc("Conv_113") +#loc86 = loc("Add_115") +#loc87 = loc("Clip_118") +#loc88 = loc("Div_120") +#loc89 = loc("Mul_121") +#loc90 = loc("Conv_122") +#loc91 = loc("Add_124") +#loc92 = loc("Clip_127") +#loc93 = loc("Div_129") +#loc94 = loc("Mul_130") +#loc95 = loc("Conv_131") +#loc96 = loc("Add_132") +#loc97 = loc("Conv_133") +#loc98 = loc("Add_135") +#loc99 = loc("Clip_138") +#loc100 = loc("Div_140") +#loc101 = loc("Mul_141") +#loc102 = loc("Conv_142") +#loc103 = loc("Add_144") +#loc104 = loc("Clip_147") +#loc105 = loc("Div_149") +#loc106 = loc("Mul_150") +#loc107 = loc("Conv_151") +#loc108 = loc("Add_152") +#loc109 = loc("Conv_153") +#loc110 = loc("Add_155") +#loc111 = loc("Clip_158") +#loc112 = loc("Div_160") +#loc113 = loc("Mul_161") +#loc114 = loc("Conv_162") +#loc115 = loc("Add_164") +#loc116 = loc("Clip_167") +#loc117 = loc("Div_169") +#loc118 = loc("Mul_170") +#loc119 = loc("Conv_171") +#loc120 = loc("Add_172") +#loc121 = loc("Conv_173") +#loc122 = loc("Add_175") +#loc123 = loc("Clip_178") +#loc124 = loc("Div_180") +#loc125 = loc("Mul_181") +#loc126 = loc("Conv_182") +#loc127 = loc("Add_184") +#loc128 = loc("Clip_187") +#loc129 = loc("Div_189") +#loc130 = loc("Mul_190") +#loc131 = loc("GlobalAveragePool_191") +#loc132 = loc("Conv_192") +#loc133 = loc("Relu_193") +#loc134 = loc("Conv_194") +#loc135 = loc("Add_196") +#loc136 = loc("Clip_199") +#loc137 = loc("Div_201") +#loc138 = loc("Mul_202") +#loc139 = loc("Conv_203") +#loc140 = loc("Conv_204") +#loc141 = loc("Add_206") +#loc142 = loc("Clip_209") +#loc143 = loc("Div_211") +#loc144 = loc("Mul_212") +#loc145 = loc("Conv_213") +#loc146 = loc("Add_215") +#loc147 = loc("Clip_218") +#loc148 = loc("Div_220") +#loc149 = loc("Mul_221") +#loc150 = loc("GlobalAveragePool_222") +#loc151 = loc("Conv_223") +#loc152 = loc("Relu_224") +#loc153 = loc("Conv_225") +#loc154 = loc("Add_227") +#loc155 = loc("Clip_230") +#loc156 = loc("Div_232") +#loc157 = loc("Mul_233") +#loc158 = loc("Conv_234") +#loc159 = loc("Add_235") +#loc160 = loc("Conv_236") +#loc161 = loc("Add_238") +#loc162 = loc("Clip_241") +#loc163 = loc("Div_243") +#loc164 = loc("Mul_244") +#loc165 = loc("Conv_245") +#loc166 = loc("Add_247") +#loc167 = loc("Clip_250") +#loc168 = loc("Div_252") +#loc169 = loc("Mul_253") +#loc170 = loc("GlobalAveragePool_254") +#loc171 = loc("Conv_255") +#loc172 = loc("Relu_256") +#loc173 = loc("Conv_257") +#loc174 = loc("Add_259") +#loc175 = loc("Clip_262") +#loc176 = loc("Div_264") +#loc177 = loc("Mul_265") +#loc178 = loc("Conv_266") +#loc179 = loc("Conv_267") +#loc180 = loc("Add_269") +#loc181 = loc("Clip_272") +#loc182 = loc("Div_274") +#loc183 = loc("Mul_275") +#loc184 = loc("Conv_276") +#loc185 = loc("Add_278") +#loc186 = loc("Clip_281") +#loc187 = loc("Div_283") +#loc188 = loc("Mul_284") +#loc189 = loc("GlobalAveragePool_285") +#loc190 = loc("Conv_286") +#loc191 = loc("Relu_287") +#loc192 = loc("Conv_288") +#loc193 = loc("Add_290") +#loc194 = loc("Clip_293") +#loc195 = loc("Div_295") +#loc196 = loc("Mul_296") +#loc197 = loc("Conv_297") +#loc198 = loc("Add_298") +#loc199 = loc("Conv_299") +#loc200 = loc("Add_301") +#loc201 = loc("Clip_304") +#loc202 = loc("Div_306") +#loc203 = loc("Mul_307") +#loc204 = loc("Conv_308") +#loc205 = loc("Add_310") +#loc206 = loc("Clip_313") +#loc207 = loc("Div_315") +#loc208 = loc("Mul_316") +#loc209 = loc("GlobalAveragePool_317") +#loc210 = loc("Conv_318") +#loc211 = loc("Relu_319") +#loc212 = loc("Conv_320") +#loc213 = loc("Add_322") +#loc214 = loc("Clip_325") +#loc215 = loc("Div_327") +#loc216 = loc("Mul_328") +#loc217 = loc("Conv_329") +#loc218 = loc("Add_330") +#loc219 = loc("Conv_331") +#loc220 = loc("Add_333") +#loc221 = loc("Clip_336") +#loc222 = loc("Div_338") +#loc223 = loc("Mul_339") +#loc224 = loc("GlobalAveragePool_342") +#loc225 = loc("Conv_343") +#loc226 = loc("Sigmoid_344") +#loc227 = loc("Mul_345") +#loc228 = loc("Conv_340") +#loc229 = loc("Relu_341") +#loc230 = loc("Split_349") +#loc231 = loc("Concat_350") +#loc232 = loc("Conv_351") +#loc233 = loc("Sigmoid_352") +#loc234 = loc("Split_353") +#loc235 = loc("Mul_354") +#loc236 = loc("Concat_355") +#loc237 = loc("Conv_356") +#loc238 = loc("Tanh_357") +#loc239 = loc("Mul_361") +#loc240 = loc("Mul_360") +#loc241 = loc("Add_362") +#loc242 = loc("Concat_363") +#loc243 = loc("Resize_365") +#loc244 = loc("Slice_371") +#loc245 = loc("AveragePool_347") +#loc246 = loc("AveragePool_348") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_380") +#loc256 = loc("Concat_381") +#loc257 = loc("Conv_382") +#loc258 = loc("Tanh_383") +#loc259 = loc("Mul_387") +#loc260 = loc("Mul_386") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_406") +#loc274 = loc("Concat_407") +#loc275 = loc("Conv_408") +#loc276 = loc("Tanh_409") +#loc277 = loc("Mul_413") +#loc278 = loc("Mul_412") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_426") +#loc291 = loc("Concat_427") +#loc292 = loc("Conv_428") +#loc293 = loc("Tanh_429") +#loc294 = loc("Mul_433") +#loc295 = loc("Mul_432") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_446") +#loc307 = loc("Clip_447") +#loc308 = loc(fused[#loc1, #loc2, #loc3, #loc4, #loc5, #loc6, #loc7, #loc8, #loc9, #loc10, #loc11, #loc12, #loc13, #loc14, #loc15, #loc16, #loc17, #loc18, #loc19, #loc20, #loc21, #loc22, #loc23, #loc24, #loc25, #loc26, #loc27, #loc28, #loc29, #loc30, #loc31, #loc32, #loc33, #loc34, #loc35, #loc36, #loc37, #loc38, #loc39, #loc40, #loc41, #loc42, #loc43, #loc44, #loc45, #loc46, #loc47, #loc48, #loc49, #loc50, #loc51, #loc52, #loc53, #loc54, #loc55, #loc56, #loc57, #loc58, #loc59, #loc60, #loc61, #loc62, #loc63, #loc64, #loc65, #loc66, #loc67, #loc68, #loc69, #loc70, #loc71, #loc72, #loc73, #loc74, #loc75, #loc76, #loc77, #loc78, #loc79, #loc80, #loc81, #loc82, #loc83, #loc84, #loc85, #loc86, #loc87, #loc88, #loc89, #loc90, #loc91, #loc92, #loc93, #loc94, #loc95, #loc96, #loc97, #loc98, #loc99, #loc100, #loc101, #loc102, #loc103, #loc104, #loc105, #loc106, #loc107, #loc108, #loc109, #loc110, #loc111, #loc112, #loc113, #loc114, #loc115, #loc116, #loc117, #loc118, #loc119, #loc120, #loc121, #loc122, #loc123, #loc124, #loc125, #loc126, #loc127, #loc128, #loc129, #loc130, #loc131, #loc132, #loc133, #loc134, #loc135, #loc136, #loc137, #loc138, #loc139, #loc140, #loc141, #loc142, #loc143, #loc144, #loc145, #loc146, #loc147, #loc148, #loc149, #loc150, #loc151, #loc152, #loc153, #loc154, #loc155, #loc156, #loc157, #loc158, #loc159, #loc160, #loc161, #loc162, #loc163, #loc164, #loc165, #loc166, #loc167, #loc168, #loc169, #loc170, #loc171, #loc172, #loc173, #loc174, #loc175, #loc176, #loc177, #loc178, #loc179, #loc180, #loc181, #loc182, #loc183, #loc184, #loc185, #loc186, #loc187, #loc188, #loc189, #loc190, #loc191, #loc192, #loc193, #loc194, #loc195, #loc196, #loc197, #loc198, #loc199, #loc200, #loc201, #loc202, #loc203, #loc204, #loc205, #loc206, #loc207, #loc208, #loc209, #loc210, #loc211, #loc212, #loc213, #loc214, #loc215, #loc216, #loc217, #loc218, #loc219, #loc220, #loc221, #loc222, #loc223, #loc224, #loc225, #loc226, #loc227, #loc228, #loc229, #loc230, #loc231, #loc232, #loc233, #loc234, #loc235, #loc236, #loc237, #loc238, #loc239, #loc240, #loc241, #loc242, #loc243, #loc244, #loc245, #loc246, #loc247, #loc248, #loc249, #loc250, #loc251, #loc252, #loc253, #loc254, #loc255, #loc256, #loc257, #loc258, #loc259, #loc260, #loc261, #loc262, #loc263, #loc264, #loc265, #loc266, #loc267, #loc268, #loc269, #loc270, #loc271, #loc272, #loc273, #loc274, #loc275, #loc276, #loc277, #loc278, #loc279, #loc280, #loc281, #loc282, #loc283, #loc284, #loc285, #loc286, #loc287, #loc288, #loc289, #loc290, #loc291, #loc292, #loc293, #loc294, #loc295, #loc296, #loc297, #loc298, #loc299, #loc300, #loc301, #loc302, #loc303, #loc304, #loc305, #loc306, #loc307]) +#loc309 = loc(fused[#loc7, #loc8]) +#loc310 = loc(fused[#loc11, #loc9, #loc12]) +#loc311 = loc(fused[#loc9, #loc12, #loc11]) +#loc312 = loc(fused[#loc18, #loc19]) +#loc313 = loc(fused[#loc20, #loc21]) +#loc314 = loc(fused[#loc22, #loc23]) +#loc315 = loc(fused[#loc24, #loc25]) +#loc316 = loc(fused[#loc27, #loc28]) +#loc317 = loc(fused[#loc29, #loc30]) +#loc318 = loc(fused[#loc31, #loc32]) +#loc319 = loc(fused[#loc33, #loc34]) +#loc320 = loc(fused[#loc35, #loc36]) +#loc321 = loc(fused[#loc38, #loc39]) +#loc322 = loc(fused[#loc46, #loc47]) +#loc323 = loc(fused[#loc48, #loc49]) +#loc324 = loc(fused[#loc51, #loc52]) +#loc325 = loc(fused[#loc58, #loc59]) +#loc326 = loc(fused[#loc60, #loc61]) +#loc327 = loc(fused[#loc62, #loc63]) +#loc328 = loc(fused[#loc65, #loc66]) +#loc329 = loc(fused[#loc72, #loc73]) +#loc330 = loc(fused[#loc95, #loc96]) +#loc331 = loc(fused[#loc107, #loc108]) +#loc332 = loc(fused[#loc119, #loc120]) +#loc333 = loc(fused[#loc130, #loc131]) +#loc334 = loc(fused[#loc132, #loc133]) +#loc335 = loc(fused[#loc149, #loc150]) +#loc336 = loc(fused[#loc151, #loc152]) +#loc337 = loc(fused[#loc158, #loc159]) +#loc338 = loc(fused[#loc169, #loc170]) +#loc339 = loc(fused[#loc171, #loc172]) +#loc340 = loc(fused[#loc188, #loc189]) +#loc341 = loc(fused[#loc190, #loc191]) +#loc342 = loc(fused[#loc197, #loc198]) +#loc343 = loc(fused[#loc208, #loc209]) +#loc344 = loc(fused[#loc210, #loc211]) +#loc345 = loc(fused[#loc217, #loc218]) +#loc346 = loc(fused[#loc223, #loc224]) +#loc347 = loc(fused[#loc228, #loc229, #loc227]) +#loc348 = loc(fused[#loc228, #loc229]) +#loc349 = loc(fused[#loc248, #loc249]) +#loc350 = loc(fused[#loc266, #loc267]) +#loc351 = loc(fused[#loc283, #loc284]) +#loc352 = loc(fused[#loc300, #loc301]) +#loc353 = loc(fused[#loc302, #loc303]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.pre-llvm.mlir b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.pre-llvm.mlir new file mode 100644 index 0000000000000000000000000000000000000000..85232e8169783f5452d410f8a3df313078bbb990 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/OnnxModel.pre-llvm.mlir @@ -0,0 +1,324 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func private @forward_outlined_part_0(tensor<1x180x320x4xf32>, tensor<1x16x90x160xf32>, tensor<1x20x45x80xf32>, tensor<1x40x23x40xf32>, tensor<1x64x12x20xf32>) -> (tensor<1x16x90x160xf32>, tensor<1x20x45x80xf32>, tensor<1x40x23x40xf32>, tensor<1x64x12x20xf32>, tensor<1x3x180x320xf32>, tensor<1x1x180x320xf32>) attributes {aie_partition = 0 : i32, kernel} loc(#loc308) + func.func @forward(%arg0: tensor<1x180x320x4xf32> {onnx.name = "385"} loc(unknown), %arg1: tensor<1x16x90x160xf32> {onnx.name = "394"} loc(unknown), %arg2: tensor<1x20x45x80xf32> {onnx.name = "395"} loc(unknown), %arg3: tensor<1x40x23x40xf32> {onnx.name = "396"} loc(unknown), %arg4: tensor<1x64x12x20xf32> {onnx.name = "397"} loc(unknown)) -> (tensor<1x1x180x320xf32> {onnx.name = "921"}, tensor<1x16x90x160xf32> {onnx.name = "894"}, tensor<1x20x45x80xf32> {onnx.name = "868"}, tensor<1x40x23x40xf32> {onnx.name = "832"}, tensor<1x64x12x20xf32> {onnx.name = "796"}, tensor<1x3x180x320xf32> {onnx.name = "916"}) { + %0:6 = call @forward_outlined_part_0(%arg0, %arg1, %arg2, %arg3, %arg4) : (tensor<1x180x320x4xf32>, tensor<1x16x90x160xf32>, tensor<1x20x45x80xf32>, tensor<1x40x23x40xf32>, tensor<1x64x12x20xf32>) -> (tensor<1x16x90x160xf32>, tensor<1x20x45x80xf32>, tensor<1x40x23x40xf32>, tensor<1x64x12x20xf32>, tensor<1x3x180x320xf32>, tensor<1x1x180x320xf32>) loc(#loc308) + return %0#5, %0#0, %0#1, %0#2, %0#3, %0#4 : tensor<1x1x180x320xf32>, tensor<1x16x90x160xf32>, tensor<1x20x45x80xf32>, tensor<1x40x23x40xf32>, tensor<1x64x12x20xf32>, tensor<1x3x180x320xf32> loc(#loc) + } loc(#loc) +} loc(#loc) +#loc1 = loc("Div_2") +#loc2 = loc("Sub_431") +#loc3 = loc("Sub_411") +#loc4 = loc("Sub_385") +#loc5 = loc("Sub_359") +#loc6 = loc("Div_16") +#loc7 = loc("Sub_14") +#loc8 = loc("Initializer_398") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("Conv_17") +#loc14 = loc("Add_19") +#loc15 = loc("Clip_22") +#loc16 = loc("Div_24") +#loc17 = loc("Mul_25") +#loc18 = loc("Conv_26") +#loc19 = loc("Relu_27") +#loc20 = loc("Conv_28") +#loc21 = loc("Add_29") +#loc22 = loc("Conv_30") +#loc23 = loc("Relu_31") +#loc24 = loc("Conv_32") +#loc25 = loc("Relu_33") +#loc26 = loc("Conv_34") +#loc27 = loc("Conv_35") +#loc28 = loc("Relu_36") +#loc29 = loc("Conv_37") +#loc30 = loc("Relu_38") +#loc31 = loc("Conv_39") +#loc32 = loc("Add_40") +#loc33 = loc("Conv_41") +#loc34 = loc("Relu_42") +#loc35 = loc("Conv_43") +#loc36 = loc("Relu_44") +#loc37 = loc("GlobalAveragePool_45") +#loc38 = loc("Conv_46") +#loc39 = loc("Relu_47") +#loc40 = loc("Conv_48") +#loc41 = loc("Add_50") +#loc42 = loc("Clip_53") +#loc43 = loc("Div_55") +#loc44 = loc("Mul_56") +#loc45 = loc("Conv_57") +#loc46 = loc("Conv_58") +#loc47 = loc("Relu_59") +#loc48 = loc("Conv_60") +#loc49 = loc("Relu_61") +#loc50 = loc("GlobalAveragePool_62") +#loc51 = loc("Conv_63") +#loc52 = loc("Relu_64") +#loc53 = loc("Conv_65") +#loc54 = loc("Add_67") +#loc55 = loc("Clip_70") +#loc56 = loc("Div_72") +#loc57 = loc("Mul_73") +#loc58 = loc("Conv_74") +#loc59 = loc("Add_75") +#loc60 = loc("Conv_76") +#loc61 = loc("Relu_77") +#loc62 = loc("Conv_78") +#loc63 = loc("Relu_79") +#loc64 = loc("GlobalAveragePool_80") +#loc65 = loc("Conv_81") +#loc66 = loc("Relu_82") +#loc67 = loc("Conv_83") +#loc68 = loc("Add_85") +#loc69 = loc("Clip_88") +#loc70 = loc("Div_90") +#loc71 = loc("Mul_91") +#loc72 = loc("Conv_92") +#loc73 = loc("Add_93") +#loc74 = loc("Conv_94") +#loc75 = loc("Add_96") +#loc76 = loc("Clip_99") +#loc77 = loc("Div_101") +#loc78 = loc("Mul_102") +#loc79 = loc("Conv_103") +#loc80 = loc("Add_105") +#loc81 = loc("Clip_108") +#loc82 = loc("Div_110") +#loc83 = loc("Mul_111") +#loc84 = loc("Conv_112") +#loc85 = loc("Conv_113") +#loc86 = loc("Add_115") +#loc87 = loc("Clip_118") +#loc88 = loc("Div_120") +#loc89 = loc("Mul_121") +#loc90 = loc("Conv_122") +#loc91 = loc("Add_124") +#loc92 = loc("Clip_127") +#loc93 = loc("Div_129") +#loc94 = loc("Mul_130") +#loc95 = loc("Conv_131") +#loc96 = loc("Add_132") +#loc97 = loc("Conv_133") +#loc98 = loc("Add_135") +#loc99 = loc("Clip_138") +#loc100 = loc("Div_140") +#loc101 = loc("Mul_141") +#loc102 = loc("Conv_142") +#loc103 = loc("Add_144") +#loc104 = loc("Clip_147") +#loc105 = loc("Div_149") +#loc106 = loc("Mul_150") +#loc107 = loc("Conv_151") +#loc108 = loc("Add_152") +#loc109 = loc("Conv_153") +#loc110 = loc("Add_155") +#loc111 = loc("Clip_158") +#loc112 = loc("Div_160") +#loc113 = loc("Mul_161") +#loc114 = loc("Conv_162") +#loc115 = loc("Add_164") +#loc116 = loc("Clip_167") +#loc117 = loc("Div_169") +#loc118 = loc("Mul_170") +#loc119 = loc("Conv_171") +#loc120 = loc("Add_172") +#loc121 = loc("Conv_173") +#loc122 = loc("Add_175") +#loc123 = loc("Clip_178") +#loc124 = loc("Div_180") +#loc125 = loc("Mul_181") +#loc126 = loc("Conv_182") +#loc127 = loc("Add_184") +#loc128 = loc("Clip_187") +#loc129 = loc("Div_189") +#loc130 = loc("Mul_190") +#loc131 = loc("GlobalAveragePool_191") +#loc132 = loc("Conv_192") +#loc133 = loc("Relu_193") +#loc134 = loc("Conv_194") +#loc135 = loc("Add_196") +#loc136 = loc("Clip_199") +#loc137 = loc("Div_201") +#loc138 = loc("Mul_202") +#loc139 = loc("Conv_203") +#loc140 = loc("Conv_204") +#loc141 = loc("Add_206") +#loc142 = loc("Clip_209") +#loc143 = loc("Div_211") +#loc144 = loc("Mul_212") +#loc145 = loc("Conv_213") +#loc146 = loc("Add_215") +#loc147 = loc("Clip_218") +#loc148 = loc("Div_220") +#loc149 = loc("Mul_221") +#loc150 = loc("GlobalAveragePool_222") +#loc151 = loc("Conv_223") +#loc152 = loc("Relu_224") +#loc153 = loc("Conv_225") +#loc154 = loc("Add_227") +#loc155 = loc("Clip_230") +#loc156 = loc("Div_232") +#loc157 = loc("Mul_233") +#loc158 = loc("Conv_234") +#loc159 = loc("Add_235") +#loc160 = loc("Conv_236") +#loc161 = loc("Add_238") +#loc162 = loc("Clip_241") +#loc163 = loc("Div_243") +#loc164 = loc("Mul_244") +#loc165 = loc("Conv_245") +#loc166 = loc("Add_247") +#loc167 = loc("Clip_250") +#loc168 = loc("Div_252") +#loc169 = loc("Mul_253") +#loc170 = loc("GlobalAveragePool_254") +#loc171 = loc("Conv_255") +#loc172 = loc("Relu_256") +#loc173 = loc("Conv_257") +#loc174 = loc("Add_259") +#loc175 = loc("Clip_262") +#loc176 = loc("Div_264") +#loc177 = loc("Mul_265") +#loc178 = loc("Conv_266") +#loc179 = loc("Conv_267") +#loc180 = loc("Add_269") +#loc181 = loc("Clip_272") +#loc182 = loc("Div_274") +#loc183 = loc("Mul_275") +#loc184 = loc("Conv_276") +#loc185 = loc("Add_278") +#loc186 = loc("Clip_281") +#loc187 = loc("Div_283") +#loc188 = loc("Mul_284") +#loc189 = loc("GlobalAveragePool_285") +#loc190 = loc("Conv_286") +#loc191 = loc("Relu_287") +#loc192 = loc("Conv_288") +#loc193 = loc("Add_290") +#loc194 = loc("Clip_293") +#loc195 = loc("Div_295") +#loc196 = loc("Mul_296") +#loc197 = loc("Conv_297") +#loc198 = loc("Add_298") +#loc199 = loc("Conv_299") +#loc200 = loc("Add_301") +#loc201 = loc("Clip_304") +#loc202 = loc("Div_306") +#loc203 = loc("Mul_307") +#loc204 = loc("Conv_308") +#loc205 = loc("Add_310") +#loc206 = loc("Clip_313") +#loc207 = loc("Div_315") +#loc208 = loc("Mul_316") +#loc209 = loc("GlobalAveragePool_317") +#loc210 = loc("Conv_318") +#loc211 = loc("Relu_319") +#loc212 = loc("Conv_320") +#loc213 = loc("Add_322") +#loc214 = loc("Clip_325") +#loc215 = loc("Div_327") +#loc216 = loc("Mul_328") +#loc217 = loc("Conv_329") +#loc218 = loc("Add_330") +#loc219 = loc("Conv_331") +#loc220 = loc("Add_333") +#loc221 = loc("Clip_336") +#loc222 = loc("Div_338") +#loc223 = loc("Mul_339") +#loc224 = loc("GlobalAveragePool_342") +#loc225 = loc("Conv_343") +#loc226 = loc("Sigmoid_344") +#loc227 = loc("Mul_345") +#loc228 = loc("Conv_340") +#loc229 = loc("Relu_341") +#loc230 = loc("Split_349") +#loc231 = loc("Concat_350") +#loc232 = loc("Conv_351") +#loc233 = loc("Sigmoid_352") +#loc234 = loc("Split_353") +#loc235 = loc("Mul_354") +#loc236 = loc("Concat_355") +#loc237 = loc("Conv_356") +#loc238 = loc("Tanh_357") +#loc239 = loc("Mul_361") +#loc240 = loc("Mul_360") +#loc241 = loc("Add_362") +#loc242 = loc("Concat_363") +#loc243 = loc("Resize_365") +#loc244 = loc("Slice_371") +#loc245 = loc("AveragePool_347") +#loc246 = loc("AveragePool_348") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_380") +#loc256 = loc("Concat_381") +#loc257 = loc("Conv_382") +#loc258 = loc("Tanh_383") +#loc259 = loc("Mul_387") +#loc260 = loc("Mul_386") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_406") +#loc274 = loc("Concat_407") +#loc275 = loc("Conv_408") +#loc276 = loc("Tanh_409") +#loc277 = loc("Mul_413") +#loc278 = loc("Mul_412") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_426") +#loc291 = loc("Concat_427") +#loc292 = loc("Conv_428") +#loc293 = loc("Tanh_429") +#loc294 = loc("Mul_433") +#loc295 = loc("Mul_432") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_446") +#loc307 = loc("Clip_447") +#loc308 = loc(fused[#loc1, #loc2, #loc3, #loc4, #loc5, #loc6, #loc7, #loc8, #loc9, #loc10, #loc11, #loc12, #loc13, #loc14, #loc15, #loc16, #loc17, #loc18, #loc19, #loc20, #loc21, #loc22, #loc23, #loc24, #loc25, #loc26, #loc27, #loc28, #loc29, #loc30, #loc31, #loc32, #loc33, #loc34, #loc35, #loc36, #loc37, #loc38, #loc39, #loc40, #loc41, #loc42, #loc43, #loc44, #loc45, #loc46, #loc47, #loc48, #loc49, #loc50, #loc51, #loc52, #loc53, #loc54, #loc55, #loc56, #loc57, #loc58, #loc59, #loc60, #loc61, #loc62, #loc63, #loc64, #loc65, #loc66, #loc67, #loc68, #loc69, #loc70, #loc71, #loc72, #loc73, #loc74, #loc75, #loc76, #loc77, #loc78, #loc79, #loc80, #loc81, #loc82, #loc83, #loc84, #loc85, #loc86, #loc87, #loc88, #loc89, #loc90, #loc91, #loc92, #loc93, #loc94, #loc95, #loc96, #loc97, #loc98, #loc99, #loc100, #loc101, #loc102, #loc103, #loc104, #loc105, #loc106, #loc107, #loc108, #loc109, #loc110, #loc111, #loc112, #loc113, #loc114, #loc115, #loc116, #loc117, #loc118, #loc119, #loc120, #loc121, #loc122, #loc123, #loc124, #loc125, #loc126, #loc127, #loc128, #loc129, #loc130, #loc131, #loc132, #loc133, #loc134, #loc135, #loc136, #loc137, #loc138, #loc139, #loc140, #loc141, #loc142, #loc143, #loc144, #loc145, #loc146, #loc147, #loc148, #loc149, #loc150, #loc151, #loc152, #loc153, #loc154, #loc155, #loc156, #loc157, #loc158, #loc159, #loc160, #loc161, #loc162, #loc163, #loc164, #loc165, #loc166, #loc167, #loc168, #loc169, #loc170, #loc171, #loc172, #loc173, #loc174, #loc175, #loc176, #loc177, #loc178, #loc179, #loc180, #loc181, #loc182, #loc183, #loc184, #loc185, #loc186, #loc187, #loc188, #loc189, #loc190, #loc191, #loc192, #loc193, #loc194, #loc195, #loc196, #loc197, #loc198, #loc199, #loc200, #loc201, #loc202, #loc203, #loc204, #loc205, #loc206, #loc207, #loc208, #loc209, #loc210, #loc211, #loc212, #loc213, #loc214, #loc215, #loc216, #loc217, #loc218, #loc219, #loc220, #loc221, #loc222, #loc223, #loc224, #loc225, #loc226, #loc227, #loc228, #loc229, #loc230, #loc231, #loc232, #loc233, #loc234, #loc235, #loc236, #loc237, #loc238, #loc239, #loc240, #loc241, #loc242, #loc243, #loc244, #loc245, #loc246, #loc247, #loc248, #loc249, #loc250, #loc251, #loc252, #loc253, #loc254, #loc255, #loc256, #loc257, #loc258, #loc259, #loc260, #loc261, #loc262, #loc263, #loc264, #loc265, #loc266, #loc267, #loc268, #loc269, #loc270, #loc271, #loc272, #loc273, #loc274, #loc275, #loc276, #loc277, #loc278, #loc279, #loc280, #loc281, #loc282, #loc283, #loc284, #loc285, #loc286, #loc287, #loc288, #loc289, #loc290, #loc291, #loc292, #loc293, #loc294, #loc295, #loc296, #loc297, #loc298, #loc299, #loc300, #loc301, #loc302, #loc303, #loc304, #loc305, #loc306, #loc307]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/aie_unsupported_original_ops.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/aie_unsupported_original_ops.json new file mode 100644 index 0000000000000000000000000000000000000000..0d4f101c7a37a4c875e6999bee1a287fdb733380 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/aie_unsupported_original_ops.json @@ -0,0 +1,2 @@ +[ +] diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/aie_unsupported_original_ops_with_reasons.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/aie_unsupported_original_ops_with_reasons.json new file mode 100644 index 0000000000000000000000000000000000000000..0d4f101c7a37a4c875e6999bee1a287fdb733380 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/aie_unsupported_original_ops_with_reasons.json @@ -0,0 +1,2 @@ +[ +] diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/chained_kernel.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/chained_kernel.json new file mode 100644 index 0000000000000000000000000000000000000000..736a6050334c4aefe3065d860c57fdd50ded509a --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/chained_kernel.json @@ -0,0 +1,241 @@ +{ + "mapping": [ + {"InCoreChain_LayerName": "Div_2", "OutputName": "Div_2"}, + {"InCoreChain_LayerName": "Sub_14", "OutputName": "Initializer_398"}, + {"InCoreChain_LayerName": "Div_16", "OutputName": "Div_16"}, + {"InCoreChain_LayerName": "Conv_17", "OutputName": "Conv_17"}, + {"InCoreChain_LayerName": "Add_19", "OutputName": "Add_19"}, + {"InCoreChain_LayerName": "Clip_22", "OutputName": "Clip_22"}, + {"InCoreChain_LayerName": "Div_24", "OutputName": "Div_24"}, + {"InCoreChain_LayerName": "Mul_25", "OutputName": "Mul_25"}, + {"InCoreChain_LayerName": "Conv_26", "OutputName": "Relu_27"}, + {"InCoreChain_LayerName": "Conv_28", "OutputName": "Add_29"}, + {"InCoreChain_LayerName": "Conv_30", "OutputName": "Relu_31"}, + {"InCoreChain_LayerName": "Conv_32", "OutputName": "Relu_33"}, + {"InCoreChain_LayerName": "Conv_34", "OutputName": "Conv_34"}, + {"InCoreChain_LayerName": "Conv_35", "OutputName": "Relu_36"}, + {"InCoreChain_LayerName": "Conv_37", "OutputName": "Relu_38"}, + {"InCoreChain_LayerName": "Conv_39", "OutputName": "Add_40"}, + {"InCoreChain_LayerName": "Conv_41", "OutputName": "Relu_42"}, + {"InCoreChain_LayerName": "Conv_43", "OutputName": "Relu_44"}, + {"InCoreChain_LayerName": "Generated-#8", "OutputName": "Generated-#9"}, + {"InCoreChain_LayerName": "Conv_46", "OutputName": "Relu_47"}, + {"InCoreChain_LayerName": "Conv_48", "OutputName": "Conv_48"}, + {"InCoreChain_LayerName": "Add_50", "OutputName": "Add_50"}, + {"InCoreChain_LayerName": "Clip_53", "OutputName": "Clip_53"}, + {"InCoreChain_LayerName": "Div_55", "OutputName": "Div_55"}, + {"InCoreChain_LayerName": "Mul_56", "OutputName": "Mul_56"}, + {"InCoreChain_LayerName": "Conv_57", "OutputName": "Conv_57"}, + {"InCoreChain_LayerName": "Conv_58", "OutputName": "Relu_59"}, + {"InCoreChain_LayerName": "Conv_60", "OutputName": "Relu_61"}, + {"InCoreChain_LayerName": "Generated-#14", "OutputName": "Generated-#15"}, + {"InCoreChain_LayerName": "Conv_63", "OutputName": "Relu_64"}, + {"InCoreChain_LayerName": "Conv_65", "OutputName": "Conv_65"}, + {"InCoreChain_LayerName": "Add_67", "OutputName": "Add_67"}, + {"InCoreChain_LayerName": "Clip_70", "OutputName": "Clip_70"}, + {"InCoreChain_LayerName": "Div_72", "OutputName": "Div_72"}, + {"InCoreChain_LayerName": "Mul_73", "OutputName": "Mul_73"}, + {"InCoreChain_LayerName": "Conv_74", "OutputName": "Add_75"}, + {"InCoreChain_LayerName": "Conv_76", "OutputName": "Relu_77"}, + {"InCoreChain_LayerName": "Conv_78", "OutputName": "Relu_79"}, + {"InCoreChain_LayerName": "Generated-#20", "OutputName": "Generated-#21"}, + {"InCoreChain_LayerName": "Conv_81", "OutputName": "Relu_82"}, + {"InCoreChain_LayerName": "Conv_83", "OutputName": "Conv_83"}, + {"InCoreChain_LayerName": "Add_85", "OutputName": "Add_85"}, + {"InCoreChain_LayerName": "Clip_88", "OutputName": "Clip_88"}, + {"InCoreChain_LayerName": "Div_90", "OutputName": "Div_90"}, + {"InCoreChain_LayerName": "Mul_91", "OutputName": "Mul_91"}, + {"InCoreChain_LayerName": "Conv_92", "OutputName": "Add_93"}, + {"InCoreChain_LayerName": "Conv_94", "OutputName": "Conv_94"}, + {"InCoreChain_LayerName": "Add_96", "OutputName": "Add_96"}, + {"InCoreChain_LayerName": "Clip_99", "OutputName": "Clip_99"}, + {"InCoreChain_LayerName": "Div_101", "OutputName": "Div_101"}, + {"InCoreChain_LayerName": "Mul_102", "OutputName": "Mul_102"}, + {"InCoreChain_LayerName": "Conv_103", "OutputName": "Conv_103"}, + {"InCoreChain_LayerName": "Add_105", "OutputName": "Add_105"}, + {"InCoreChain_LayerName": "Clip_108", "OutputName": "Clip_108"}, + {"InCoreChain_LayerName": "Div_110", "OutputName": "Div_110"}, + {"InCoreChain_LayerName": "Mul_111", "OutputName": "Mul_111"}, + {"InCoreChain_LayerName": "Conv_112", "OutputName": "Conv_112"}, + {"InCoreChain_LayerName": "Conv_113", "OutputName": "Conv_113"}, + {"InCoreChain_LayerName": "Add_115", "OutputName": "Add_115"}, + {"InCoreChain_LayerName": "Clip_118", "OutputName": "Clip_118"}, + {"InCoreChain_LayerName": "Div_120", "OutputName": "Div_120"}, + {"InCoreChain_LayerName": "Mul_121", "OutputName": "Mul_121"}, + {"InCoreChain_LayerName": "Conv_122", "OutputName": "Conv_122"}, + {"InCoreChain_LayerName": "Add_124", "OutputName": "Add_124"}, + {"InCoreChain_LayerName": "Clip_127", "OutputName": "Clip_127"}, + {"InCoreChain_LayerName": "Div_129", "OutputName": "Div_129"}, + {"InCoreChain_LayerName": "Mul_130", "OutputName": "Mul_130"}, + {"InCoreChain_LayerName": "Conv_131", "OutputName": "Add_132"}, + {"InCoreChain_LayerName": "Conv_133", "OutputName": "Conv_133"}, + {"InCoreChain_LayerName": "Add_135", "OutputName": "Add_135"}, + {"InCoreChain_LayerName": "Clip_138", "OutputName": "Clip_138"}, + {"InCoreChain_LayerName": "Div_140", "OutputName": "Div_140"}, + {"InCoreChain_LayerName": "Mul_141", "OutputName": "Mul_141"}, + {"InCoreChain_LayerName": "Conv_142", "OutputName": "Conv_142"}, + {"InCoreChain_LayerName": "Add_144", "OutputName": "Add_144"}, + {"InCoreChain_LayerName": "Clip_147", "OutputName": "Clip_147"}, + {"InCoreChain_LayerName": "Div_149", "OutputName": "Div_149"}, + {"InCoreChain_LayerName": "Mul_150", "OutputName": "Mul_150"}, + {"InCoreChain_LayerName": "Conv_151", "OutputName": "Add_152"}, + {"InCoreChain_LayerName": "Conv_153", "OutputName": "Conv_153"}, + {"InCoreChain_LayerName": "Add_155", "OutputName": "Add_155"}, + {"InCoreChain_LayerName": "Clip_158", "OutputName": "Clip_158"}, + {"InCoreChain_LayerName": "Div_160", "OutputName": "Div_160"}, + {"InCoreChain_LayerName": "Mul_161", "OutputName": "Mul_161"}, + {"InCoreChain_LayerName": "Conv_162", "OutputName": "Conv_162"}, + {"InCoreChain_LayerName": "Add_164", "OutputName": "Add_164"}, + {"InCoreChain_LayerName": "Clip_167", "OutputName": "Clip_167"}, + {"InCoreChain_LayerName": "Div_169", "OutputName": "Div_169"}, + {"InCoreChain_LayerName": "Mul_170", "OutputName": "Mul_170"}, + {"InCoreChain_LayerName": "Conv_171", "OutputName": "Add_172"}, + {"InCoreChain_LayerName": "Conv_173", "OutputName": "Conv_173"}, + {"InCoreChain_LayerName": "Add_175", "OutputName": "Add_175"}, + {"InCoreChain_LayerName": "Clip_178", "OutputName": "Clip_178"}, + {"InCoreChain_LayerName": "Div_180", "OutputName": "Div_180"}, + {"InCoreChain_LayerName": "Mul_181", "OutputName": "Mul_181"}, + {"InCoreChain_LayerName": "Conv_182", "OutputName": "Conv_182"}, + {"InCoreChain_LayerName": "Add_184", "OutputName": "Add_184"}, + {"InCoreChain_LayerName": "Clip_187", "OutputName": "Clip_187"}, + {"InCoreChain_LayerName": "Div_189", "OutputName": "Div_189"}, + {"InCoreChain_LayerName": "Mul_190", "OutputName": "Mul_190"}, + {"InCoreChain_LayerName": "Generated-#26", "OutputName": "Generated-#27"}, + {"InCoreChain_LayerName": "Conv_192", "OutputName": "Relu_193"}, + {"InCoreChain_LayerName": "Conv_194", "OutputName": "Conv_194"}, + {"InCoreChain_LayerName": "Add_196", "OutputName": "Add_196"}, + {"InCoreChain_LayerName": "Clip_199", "OutputName": "Clip_199"}, + {"InCoreChain_LayerName": "Div_201", "OutputName": "Div_201"}, + {"InCoreChain_LayerName": "Mul_202", "OutputName": "Mul_202"}, + {"InCoreChain_LayerName": "Conv_203", "OutputName": "Conv_203"}, + {"InCoreChain_LayerName": "Conv_204", "OutputName": "Conv_204"}, + {"InCoreChain_LayerName": "Add_206", "OutputName": "Add_206"}, + {"InCoreChain_LayerName": "Clip_209", "OutputName": "Clip_209"}, + {"InCoreChain_LayerName": "Div_211", "OutputName": "Div_211"}, + {"InCoreChain_LayerName": "Mul_212", "OutputName": "Mul_212"}, + {"InCoreChain_LayerName": "Conv_213", "OutputName": "Conv_213"}, + {"InCoreChain_LayerName": "Add_215", "OutputName": "Add_215"}, + {"InCoreChain_LayerName": "Clip_218", "OutputName": "Clip_218"}, + {"InCoreChain_LayerName": "Div_220", "OutputName": "Div_220"}, + {"InCoreChain_LayerName": "Mul_221", "OutputName": "Mul_221"}, + {"InCoreChain_LayerName": "Generated-#32", "OutputName": "Generated-#33"}, + {"InCoreChain_LayerName": "Conv_223", "OutputName": "Relu_224"}, + {"InCoreChain_LayerName": "Conv_225", "OutputName": "Conv_225"}, + {"InCoreChain_LayerName": "Add_227", "OutputName": "Add_227"}, + {"InCoreChain_LayerName": "Clip_230", "OutputName": "Clip_230"}, + {"InCoreChain_LayerName": "Div_232", "OutputName": "Div_232"}, + {"InCoreChain_LayerName": "Mul_233", "OutputName": "Mul_233"}, + {"InCoreChain_LayerName": "Conv_234", "OutputName": "Add_235"}, + {"InCoreChain_LayerName": "Conv_236", "OutputName": "Conv_236"}, + {"InCoreChain_LayerName": "Add_238", "OutputName": "Add_238"}, + {"InCoreChain_LayerName": "Clip_241", "OutputName": "Clip_241"}, + {"InCoreChain_LayerName": "Div_243", "OutputName": "Div_243"}, + {"InCoreChain_LayerName": "Mul_244", "OutputName": "Mul_244"}, + {"InCoreChain_LayerName": "Conv_245", "OutputName": "Conv_245"}, + {"InCoreChain_LayerName": "Add_247", "OutputName": "Add_247"}, + {"InCoreChain_LayerName": "Clip_250", "OutputName": "Clip_250"}, + {"InCoreChain_LayerName": "Div_252", "OutputName": "Div_252"}, + {"InCoreChain_LayerName": "Mul_253", "OutputName": "Mul_253"}, + {"InCoreChain_LayerName": "Generated-#38", "OutputName": "Generated-#39"}, + {"InCoreChain_LayerName": "Conv_255", "OutputName": "Relu_256"}, + {"InCoreChain_LayerName": "Conv_257", "OutputName": "Conv_257"}, + {"InCoreChain_LayerName": "Add_259", "OutputName": "Add_259"}, + {"InCoreChain_LayerName": "Clip_262", "OutputName": "Clip_262"}, + {"InCoreChain_LayerName": "Div_264", "OutputName": "Div_264"}, + {"InCoreChain_LayerName": "Mul_265", "OutputName": "Mul_265"}, + {"InCoreChain_LayerName": "Conv_266", "OutputName": "Conv_266"}, + {"InCoreChain_LayerName": "Conv_267", "OutputName": "Conv_267"}, + {"InCoreChain_LayerName": "Add_269", "OutputName": "Add_269"}, + {"InCoreChain_LayerName": "Clip_272", "OutputName": "Clip_272"}, + {"InCoreChain_LayerName": "Div_274", "OutputName": "Div_274"}, + {"InCoreChain_LayerName": "Mul_275", "OutputName": "Mul_275"}, + {"InCoreChain_LayerName": "Conv_276", "OutputName": "Conv_276"}, + {"InCoreChain_LayerName": "Add_278", "OutputName": "Add_278"}, + {"InCoreChain_LayerName": "Clip_281", "OutputName": "Clip_281"}, + {"InCoreChain_LayerName": "Div_283", "OutputName": "Div_283"}, + {"InCoreChain_LayerName": "Mul_284", "OutputName": "Mul_284"}, + {"InCoreChain_LayerName": "Generated-#44", "OutputName": "Generated-#45"}, + {"InCoreChain_LayerName": "Conv_286", "OutputName": "Relu_287"}, + {"InCoreChain_LayerName": "Conv_288", "OutputName": "Conv_288"}, + {"InCoreChain_LayerName": "Add_290", "OutputName": "Add_290"}, + {"InCoreChain_LayerName": "Clip_293", "OutputName": "Clip_293"}, + {"InCoreChain_LayerName": "Div_295", "OutputName": "Div_295"}, + {"InCoreChain_LayerName": "Mul_296", "OutputName": "Mul_296"}, + {"InCoreChain_LayerName": "Conv_297", "OutputName": "Add_298"}, + {"InCoreChain_LayerName": "Conv_299", "OutputName": "Conv_299"}, + {"InCoreChain_LayerName": "Add_301", "OutputName": "Add_301"}, + {"InCoreChain_LayerName": "Clip_304", "OutputName": "Clip_304"}, + {"InCoreChain_LayerName": "Div_306", "OutputName": "Div_306"}, + {"InCoreChain_LayerName": "Mul_307", "OutputName": "Mul_307"}, + {"InCoreChain_LayerName": "Conv_308", "OutputName": "Conv_308"}, + {"InCoreChain_LayerName": "Add_310", "OutputName": "Add_310"}, + {"InCoreChain_LayerName": "Clip_313", "OutputName": "Clip_313"}, + {"InCoreChain_LayerName": "Div_315", "OutputName": "Div_315"}, + {"InCoreChain_LayerName": "Mul_316", "OutputName": "Mul_316"}, + {"InCoreChain_LayerName": "Generated-#50", "OutputName": "Generated-#51"}, + {"InCoreChain_LayerName": "Conv_318", "OutputName": "Relu_319"}, + {"InCoreChain_LayerName": "Conv_320", "OutputName": "Conv_320"}, + {"InCoreChain_LayerName": "Add_322", "OutputName": "Add_322"}, + {"InCoreChain_LayerName": "Clip_325", "OutputName": "Clip_325"}, + {"InCoreChain_LayerName": "Div_327", "OutputName": "Div_327"}, + {"InCoreChain_LayerName": "Mul_328", "OutputName": "Mul_328"}, + {"InCoreChain_LayerName": "Conv_329", "OutputName": "Add_330"}, + {"InCoreChain_LayerName": "Conv_331", "OutputName": "Conv_331"}, + {"InCoreChain_LayerName": "Add_333", "OutputName": "Add_333"}, + {"InCoreChain_LayerName": "Clip_336", "OutputName": "Clip_336"}, + {"InCoreChain_LayerName": "Div_338", "OutputName": "Div_338"}, + {"InCoreChain_LayerName": "Mul_339", "OutputName": "Mul_339"}, + {"InCoreChain_LayerName": "Generated-#56", "OutputName": "Generated-#57"}, + {"InCoreChain_LayerName": "Conv_343", "OutputName": "Conv_343"}, + {"InCoreChain_LayerName": "Sigmoid_344", "OutputName": "Sigmoid_344"}, + {"InCoreChain_LayerName": "Conv_340", "OutputName": "Mul_345"}, + {"InCoreChain_LayerName": "Conv_351", "OutputName": "Conv_351"}, + {"InCoreChain_LayerName": "Sigmoid_352", "OutputName": "Sigmoid_352"}, + {"InCoreChain_LayerName": "Sub_359", "OutputName": "Sub_359"}, + {"InCoreChain_LayerName": "Mul_360", "OutputName": "Mul_360"}, + {"InCoreChain_LayerName": "Mul_354", "OutputName": "Mul_354"}, + {"InCoreChain_LayerName": "Conv_356", "OutputName": "Conv_356"}, + {"InCoreChain_LayerName": "Tanh_357", "OutputName": "Tanh_357"}, + {"InCoreChain_LayerName": "Mul_361", "OutputName": "Mul_361"}, + {"InCoreChain_LayerName": "Add_362", "OutputName": "Add_362"}, + {"InCoreChain_LayerName": "AveragePool_346", "OutputName": "AveragePool_346"}, + {"InCoreChain_LayerName": "AveragePool_347", "OutputName": "AveragePool_347"}, + {"InCoreChain_LayerName": "AveragePool_348", "OutputName": "AveragePool_348"}, + {"InCoreChain_LayerName": "Conv_373", "OutputName": "Relu_374"}, + {"InCoreChain_LayerName": "Conv_377", "OutputName": "Conv_377"}, + {"InCoreChain_LayerName": "Sigmoid_378", "OutputName": "Sigmoid_378"}, + {"InCoreChain_LayerName": "Sub_385", "OutputName": "Sub_385"}, + {"InCoreChain_LayerName": "Mul_386", "OutputName": "Mul_386"}, + {"InCoreChain_LayerName": "Mul_380", "OutputName": "Mul_380"}, + {"InCoreChain_LayerName": "Conv_382", "OutputName": "Conv_382"}, + {"InCoreChain_LayerName": "Tanh_383", "OutputName": "Tanh_383"}, + {"InCoreChain_LayerName": "Mul_387", "OutputName": "Mul_387"}, + {"InCoreChain_LayerName": "Add_388", "OutputName": "Add_388"}, + {"InCoreChain_LayerName": "Conv_399", "OutputName": "Relu_400"}, + {"InCoreChain_LayerName": "Conv_403", "OutputName": "Conv_403"}, + {"InCoreChain_LayerName": "Sigmoid_404", "OutputName": "Sigmoid_404"}, + {"InCoreChain_LayerName": "Sub_411", "OutputName": "Sub_411"}, + {"InCoreChain_LayerName": "Mul_412", "OutputName": "Mul_412"}, + {"InCoreChain_LayerName": "Mul_406", "OutputName": "Mul_406"}, + {"InCoreChain_LayerName": "Conv_408", "OutputName": "Conv_408"}, + {"InCoreChain_LayerName": "Tanh_409", "OutputName": "Tanh_409"}, + {"InCoreChain_LayerName": "Mul_413", "OutputName": "Mul_413"}, + {"InCoreChain_LayerName": "Add_414", "OutputName": "Add_414"}, + {"InCoreChain_LayerName": "Conv_419", "OutputName": "Relu_420"}, + {"InCoreChain_LayerName": "Conv_423", "OutputName": "Conv_423"}, + {"InCoreChain_LayerName": "Sigmoid_424", "OutputName": "Sigmoid_424"}, + {"InCoreChain_LayerName": "Sub_431", "OutputName": "Sub_431"}, + {"InCoreChain_LayerName": "Mul_432", "OutputName": "Mul_432"}, + {"InCoreChain_LayerName": "Mul_426", "OutputName": "Mul_426"}, + {"InCoreChain_LayerName": "Conv_428", "OutputName": "Conv_428"}, + {"InCoreChain_LayerName": "Tanh_429", "OutputName": "Tanh_429"}, + {"InCoreChain_LayerName": "Mul_433", "OutputName": "Mul_433"}, + {"InCoreChain_LayerName": "Add_434", "OutputName": "Add_434"}, + {"InCoreChain_LayerName": "Conv_439", "OutputName": "Relu_440"}, + {"InCoreChain_LayerName": "Conv_441", "OutputName": "Relu_442"}, + {"InCoreChain_LayerName": "Conv_443", "OutputName": "Conv_443"}, + {"InCoreChain_LayerName": "Clip_447", "OutputName": "Clip_447"}, + {"InCoreChain_LayerName": "Add_445", "OutputName": "Add_445"}, + {"InCoreChain_LayerName": "Clip_446", "OutputName": "Clip_446"} + ] +} diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/fail_safe_summary.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/fail_safe_summary.json new file mode 100644 index 0000000000000000000000000000000000000000..c683d07c15505bd5b0cc3a23089623f4071452e1 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/fail_safe_summary.json @@ -0,0 +1,6 @@ +{ + "offload_map": { + "AIE": 100, + "CPU": 0 + } +} \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/input_output_shape.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/input_output_shape.json new file mode 100644 index 0000000000000000000000000000000000000000..2789bdc631c2aed8d1d733e908c3ef8a0552d924 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/input_output_shape.json @@ -0,0 +1,17 @@ +{ + "input_shapes": { + "385": "1 180 320 4 ", + "394": "1 16 90 160 ", + "395": "1 20 45 80 ", + "396": "1 40 23 40 ", + "397": "1 64 12 20 " + }, + "output_shapes": { + "796": "1 64 12 20 ", + "832": "1 40 23 40 ", + "868": "1 20 45 80 ", + "894": "1 16 90 160 ", + "916": "1 3 180 320 ", + "921": "1 1 180 320 " + } +} \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/libflexml_usermodel.so b/segmentation_1_4_0_fp32_combined/vaiml_par_0/libflexml_usermodel.so new file mode 100644 index 0000000000000000000000000000000000000000..674c9567fb5c07bc15dc1b11905bbda645b22f7e Binary files /dev/null and b/segmentation_1_4_0_fp32_combined/vaiml_par_0/libflexml_usermodel.so differ diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/original-info-signature.txt b/segmentation_1_4_0_fp32_combined/vaiml_par_0/original-info-signature.txt new file mode 100644 index 0000000000000000000000000000000000000000..4262119974eb805ab03f9c534be3592fbbdd1a54 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/original-info-signature.txt @@ -0,0 +1 @@ +4f00fb244983f7c2158dc9333522f122 \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/original-model-signature.txt b/segmentation_1_4_0_fp32_combined/vaiml_par_0/original-model-signature.txt new file mode 100644 index 0000000000000000000000000000000000000000..c662815d6b6f6f7d554bf95938c117b912e549a3 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/original-model-signature.txt @@ -0,0 +1 @@ +6bbb891ab96ca9362e0e61024cd02778 \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_par_0/partition-info.json b/segmentation_1_4_0_fp32_combined/vaiml_par_0/partition-info.json new file mode 100644 index 0000000000000000000000000000000000000000..f92d06a31fe30cf466af136b7ab72d185c070d5d --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_par_0/partition-info.json @@ -0,0 +1,19 @@ +{ + "compiler": "flexml.compile", + "dry_run": false, + "sha1sum": "49fbd10bbe83af92f72f18b3ecc2be9c4b74f1f4", + "device": "stx", + "status": "completed", + "num_aie_partitions": 1, + "xrt": "na", + "flexml": "1.5.0.dev20250320125000+g1d4c9c71", + "os": "Ubuntu 22.04", + "target_os": "linux", + "modelname": "OnnxModel", + "ai_analyzer_profiling_enabled": false, + "runner_type": "hw", + "output_type": "aie-exe", + "aie_partition_call_order": [ + "forward_outlined_part_0" + ] +} \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/0/par.mlopslib.tosa.mlir b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/0/par.mlopslib.tosa.mlir new file mode 100644 index 0000000000000000000000000000000000000000..e4452b711aa005972567eef191a2458379f17fc0 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/0/par.mlopslib.tosa.mlir @@ -0,0 +1,26418 @@ +#loc308 = loc("Cast_0") +#loc309 = loc("Transpose_9") +#loc310 = loc("Transpose_10") +#loc311 = loc("Transpose_11") +#loc312 = loc("Transpose_12") +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func private @forward(%arg0: tensor<1x180x320x4xbf16> loc("Cast_0"), %arg1: tensor<1x16x90x160xbf16> loc("Transpose_9"), %arg2: tensor<1x20x45x80xbf16> loc("Transpose_10"), %arg3: tensor<1x40x23x40xbf16> loc("Transpose_11"), %arg4: tensor<1x64x12x20xbf16> loc("Transpose_12")) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) { + %0 = xten_nn.subgraph (%arg5 = %arg0: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "386", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "386", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.906250e-03 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.906250e-03> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_2", + OutputName = "Div_2", + shift = 0 : i8} : (tensor<1x180x320x4xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %463 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %461 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + %1 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/biases"} -> tensor<4xbf16> loc(#loc) + %2 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/weights"} -> tensor<4x16x1x1xbf16> loc(#loc) + %3 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/biases"} -> tensor<16xbf16> loc(#loc) + %4 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/weights"} -> tensor<16x16x3x3xbf16> loc(#loc) + %5 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/biases"} -> tensor<16xbf16> loc(#loc) + %6 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/weights"} -> tensor<16x35x3x3xbf16> loc(#loc) + %7 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/biases"} -> tensor<16xbf16> loc(#loc) + %8 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/weights"} -> tensor<16x32x3x3xbf16> loc(#loc) + %9 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_431/Constant_0_0"} -> tensor<1x16x90x160xbf16> loc(#loc2) + %10 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/biases"} -> tensor<32xbf16> loc(#loc) + %11 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/weights"} -> tensor<32x32x3x3xbf16> loc(#loc) + %12 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/biases"} -> tensor<32xbf16> loc(#loc) + %13 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/weights"} -> tensor<32x59x3x3xbf16> loc(#loc) + %14 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/biases"} -> tensor<20xbf16> loc(#loc) + %15 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/weights"} -> tensor<20x40x3x3xbf16> loc(#loc) + %16 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_411/Constant_0_0"} -> tensor<1x20x45x80xbf16> loc(#loc3) + %17 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/biases"} -> tensor<40xbf16> loc(#loc) + %18 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/weights"} -> tensor<40x40x3x3xbf16> loc(#loc) + %19 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/biases"} -> tensor<40xbf16> loc(#loc) + %20 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/weights"} -> tensor<40x107x3x3xbf16> loc(#loc) + %21 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/biases"} -> tensor<40xbf16> loc(#loc) + %22 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/weights"} -> tensor<40x80x3x3xbf16> loc(#loc) + %23 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_385/Constant_0_0"} -> tensor<1x40x23x40xbf16> loc(#loc4) + %24 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/biases"} -> tensor<80xbf16> loc(#loc) + %25 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/weights"} -> tensor<80x80x3x3xbf16> loc(#loc) + %26 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/biases"} -> tensor<80xbf16> loc(#loc) + %27 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/weights"} -> tensor<80x171x3x3xbf16> loc(#loc) + %28 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/biases"} -> tensor<64xbf16> loc(#loc) + %29 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/weights"} -> tensor<64x128x3x3xbf16> loc(#loc) + %30 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_359/Constant_0_0"} -> tensor<1x64x12x20xbf16> loc(#loc5) + %31 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/biases"} -> tensor<128xbf16> loc(#loc) + %32 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/weights"} -> tensor<128x128x3x3xbf16> loc(#loc) + %33 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/biases"} -> tensor<128xbf16> loc(#loc) + %34 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %35 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/biases"} -> tensor<128xbf16> loc(#loc) + %36 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %37 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/biases"} -> tensor<960xbf16> loc(#loc) + %38 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %39 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/biases"} -> tensor<160xbf16> loc(#loc) + %40 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %41 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/biases"} -> tensor<960xbf16> loc(#loc) + %42 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %43 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/biases"} -> tensor<240xbf16> loc(#loc) + %44 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %45 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/biases"} -> tensor<960xbf16> loc(#loc) + %46 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %47 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/biases"} -> tensor<960xbf16> loc(#loc) + %48 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %49 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/biases"} -> tensor<160xbf16> loc(#loc) + %50 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %51 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/biases"} -> tensor<960xbf16> loc(#loc) + %52 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %53 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/biases"} -> tensor<240xbf16> loc(#loc) + %54 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %55 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/biases"} -> tensor<960xbf16> loc(#loc) + %56 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %57 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/biases"} -> tensor<960xbf16> loc(#loc) + %58 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %59 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/biases"} -> tensor<160xbf16> loc(#loc) + %60 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/weights"} -> tensor<160x672x1x1xbf16> loc(#loc) + %61 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/biases"} -> tensor<672xbf16> loc(#loc) + %62 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %63 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/biases"} -> tensor<168xbf16> loc(#loc) + %64 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %65 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/biases"} -> tensor<672xbf16> loc(#loc) + %66 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/weights"} -> tensor<672x1x9x9xbf16> loc(#loc) + %67 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/biases"} -> tensor<672xbf16> loc(#loc) + %68 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %69 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/biases"} -> tensor<112xbf16> loc(#loc) + %70 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/weights"} -> tensor<112x672x1x1xbf16> loc(#loc) + %71 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/biases"} -> tensor<672xbf16> loc(#loc) + %72 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %73 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/biases"} -> tensor<168xbf16> loc(#loc) + %74 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %75 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/biases"} -> tensor<672xbf16> loc(#loc) + %76 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/weights"} -> tensor<672x1x3x3xbf16> loc(#loc) + %77 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/biases"} -> tensor<672xbf16> loc(#loc) + %78 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %79 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/biases"} -> tensor<112xbf16> loc(#loc) + %80 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/weights"} -> tensor<112x480x1x1xbf16> loc(#loc) + %81 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/biases"} -> tensor<480xbf16> loc(#loc) + %82 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/weights"} -> tensor<480x120x1x1xbf16> loc(#loc) + %83 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/biases"} -> tensor<120xbf16> loc(#loc) + %84 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/weights"} -> tensor<120x480x1x1xbf16> loc(#loc) + %85 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/biases"} -> tensor<480xbf16> loc(#loc) + %86 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/weights"} -> tensor<480x1x3x3xbf16> loc(#loc) + %87 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/biases"} -> tensor<480xbf16> loc(#loc) + %88 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/weights"} -> tensor<480x80x1x1xbf16> loc(#loc) + %89 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/biases"} -> tensor<80xbf16> loc(#loc) + %90 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %91 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/biases"} -> tensor<184xbf16> loc(#loc) + %92 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %93 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/biases"} -> tensor<184xbf16> loc(#loc) + %94 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %95 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/biases"} -> tensor<80xbf16> loc(#loc) + %96 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %97 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/biases"} -> tensor<184xbf16> loc(#loc) + %98 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %99 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/biases"} -> tensor<184xbf16> loc(#loc) + %100 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %101 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/biases"} -> tensor<80xbf16> loc(#loc) + %102 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/weights"} -> tensor<80x200x1x1xbf16> loc(#loc) + %103 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/biases"} -> tensor<200xbf16> loc(#loc) + %104 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/weights"} -> tensor<200x1x3x3xbf16> loc(#loc) + %105 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/biases"} -> tensor<200xbf16> loc(#loc) + %106 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/weights"} -> tensor<200x80x1x1xbf16> loc(#loc) + %107 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/biases"} -> tensor<80xbf16> loc(#loc) + %108 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/weights"} -> tensor<80x240x1x1xbf16> loc(#loc) + %109 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/biases"} -> tensor<240xbf16> loc(#loc) + %110 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/weights"} -> tensor<240x1x3x3xbf16> loc(#loc) + %111 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/biases"} -> tensor<240xbf16> loc(#loc) + %112 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/weights"} -> tensor<240x40x1x1xbf16> loc(#loc) + %113 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/biases"} -> tensor<40xbf16> loc(#loc) + %114 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %115 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/biases"} -> tensor<120xbf16> loc(#loc) + %116 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %117 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/biases"} -> tensor<32xbf16> loc(#loc) + %118 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %119 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/biases"} -> tensor<120xbf16> loc(#loc) + %120 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %121 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/biases"} -> tensor<120xbf16> loc(#loc) + %122 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %123 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/biases"} -> tensor<40xbf16> loc(#loc) + %124 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %125 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/biases"} -> tensor<120xbf16> loc(#loc) + %126 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %127 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/biases"} -> tensor<32xbf16> loc(#loc) + %128 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %129 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/biases"} -> tensor<120xbf16> loc(#loc) + %130 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %131 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/biases"} -> tensor<120xbf16> loc(#loc) + %132 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %133 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/biases"} -> tensor<40xbf16> loc(#loc) + %134 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/weights"} -> tensor<40x72x1x1xbf16> loc(#loc) + %135 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/biases"} -> tensor<72xbf16> loc(#loc) + %136 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %137 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/biases"} -> tensor<24xbf16> loc(#loc) + %138 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %139 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/biases"} -> tensor<72xbf16> loc(#loc) + %140 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/weights"} -> tensor<72x1x5x5xbf16> loc(#loc) + %141 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/biases"} -> tensor<72xbf16> loc(#loc) + %142 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %143 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/biases"} -> tensor<24xbf16> loc(#loc) + %144 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %145 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/biases"} -> tensor<72xbf16> loc(#loc) + %146 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/weights"} -> tensor<72x1x3x3xbf16> loc(#loc) + %147 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/biases"} -> tensor<72xbf16> loc(#loc) + %148 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %149 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/biases"} -> tensor<24xbf16> loc(#loc) + %150 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/weights"} -> tensor<24x64x1x1xbf16> loc(#loc) + %151 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/biases"} -> tensor<64xbf16> loc(#loc) + %152 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/weights"} -> tensor<64x1x3x3xbf16> loc(#loc) + %153 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/biases"} -> tensor<64xbf16> loc(#loc) + %154 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/weights"} -> tensor<64x16x1x1xbf16> loc(#loc) + %155 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/biases"} -> tensor<16xbf16> loc(#loc) + %156 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/weights"} -> tensor<16x16x1x1xbf16> loc(#loc) + %157 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/biases"} -> tensor<16xbf16> loc(#loc) + %158 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/weights"} -> tensor<16x1x3x3xbf16> loc(#loc) + %159 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/biases"} -> tensor<16xbf16> loc(#loc) + %160 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/weights"} -> tensor<16x3x3x3xbf16> loc(#loc) + %161 = xten_nn.load_external_const {file = "constants.h5", key = "Div_16/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc6) + %162 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_14/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc314) + %163 = xten_nn.subgraph (%arg5 = %0: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Slice_7_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Slice_7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "392", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "W", + config.dim_c = 184 : ui32, + config.dim_h = 320 : ui32, + config.dim_w = 4 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_7", + OutputName = "Slice_7", + size = array, + start = array} : (tensor<1x180x320x4xbf16>) -> tensor<1x180x320x3xbf16> loc(#loc9) + xten_nn.output %461 : tensor<1x180x320x3xbf16> loc(#loc9) + } -> tensor<1x180x320x3xbf16> loc(#loc9) + %164 = xten_nn.subgraph (%arg5 = %163: tensor<1x180x320x3xbf16>) attributes { + LayerName = "CompilerGenerated_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "CompilerGenerated_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "BufferPadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_0_padded = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_1_padded = 23 : ui32, + config.dim_2 = 3 : ui32, + config.dim_2_padded = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_padded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x180x320x3xbf16> loc(#loc10) + } -> tensor<1x180x320x3xbf16> loc(#loc10) + %165 = xten_nn.subgraph (%arg5 = %164: tensor<1x180x320x3xbf16>) attributes { + LayerName = "Slice_7_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Add_445_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_2 = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 10 : ui32 + }} { + %461 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc11) + %462 = tosa.transpose %arg5, %461 : (tensor<1x180x320x3xbf16>, tensor<4xi32>) -> tensor<1x3x180x320xbf16> loc(#loc316) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc316) + } -> tensor<1x3x180x320xbf16> loc(#loc315) + %166 = xten_nn.subgraph (%arg5 = %165: tensor<1x3x180x320xbf16>) attributes { + LayerName = "CompilerGenerated_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "CompilerGenerated_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 184 : ui32, + config.dim_0_unpadded = 180 : ui32, + config.dim_1 = 1 : ui32, + config.dim_1_unpadded = 1 : ui32, + config.dim_2 = 320 : ui32, + config.dim_2_unpadded = 320 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x3x180x320xbf16> loc(#loc10) + } -> tensor<1x3x180x320xbf16> loc(#loc10) + %167 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>) attributes { + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc12) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc12) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_346", + PartOfOutputName = "AveragePool_346", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>) -> tensor<1x90x160x3xbf16> loc(#loc12) + %466 = tosa.transpose %465, %462 : (tensor<1x90x160x3xbf16>, tensor<4xi32>) -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %466 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %461 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + %168 = xten_nn.subgraph (%arg5 = %167: tensor<1x3x90x160xbf16>) attributes { + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x90x160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc13) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x3xbf16> loc(#loc13) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_347", + PartOfOutputName = "AveragePool_347", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x90x160x3xbf16>) -> tensor<1x45x80x3xbf16> loc(#loc13) + %466 = tosa.transpose %465, %462 : (tensor<1x45x80x3xbf16>, tensor<4xi32>) -> tensor<1x3x45x80xbf16> loc(#loc13) + xten_nn.output %466 : tensor<1x3x45x80xbf16> loc(#loc13) + } -> tensor<1x3x45x80xbf16> loc(#loc13) + xten_nn.output %461 : tensor<1x3x45x80xbf16> loc(#loc13) + } -> tensor<1x3x45x80xbf16> loc(#loc13) + %169 = xten_nn.subgraph (%arg5 = %168: tensor<1x3x45x80xbf16>) attributes { + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "780", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x45x80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 1], [0, 0]], + HWPaddingNotCounted = [[0, 1], [0, 0]], + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "780", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc14) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x3xbf16> loc(#loc14) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_348", + PartOfOutputName = "AveragePool_348", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x45x80x3xbf16>) -> tensor<1x23x40x3xbf16> loc(#loc14) + %466 = tosa.transpose %465, %462 : (tensor<1x23x40x3xbf16>, tensor<4xi32>) -> tensor<1x3x23x40xbf16> loc(#loc14) + xten_nn.output %466 : tensor<1x3x23x40xbf16> loc(#loc14) + } -> tensor<1x3x23x40xbf16> loc(#loc14) + xten_nn.output %461 : tensor<1x3x23x40xbf16> loc(#loc14) + } -> tensor<1x3x23x40xbf16> loc(#loc14) + %170 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>, %arg6 = %162: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "392", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Sub_14", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "392", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Sub_14", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Sub_14", OutputName = "Initializer_398"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc314) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc314) + } -> tensor<1x3x180x320xbf16> loc(#loc314) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc314) + } -> tensor<1x3x180x320xbf16> loc(#loc314) + %171 = xten_nn.subgraph (%arg5 = %170: tensor<1x3x180x320xbf16>, %arg6 = %161: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + OutputName = "Div_16", + PartOfLayerName = "Div_16", + shift = 0 : i8} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + %172 = xten_nn.subgraph (%arg5 = %171: tensor<1x3x180x320xbf16>, %arg6 = %160: tensor<16x3x3x3xbf16>, %arg7 = %159: tensor<16xbf16>) attributes { + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Name = "399", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Name = "929", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x3x180x320xbf16>, %arg9 = %arg6: tensor<16x3x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Name = "399", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Name = "929", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 2 : ui8, + config.stride_w = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x3x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x3xbf16> loc(#loc15) + %465 = tosa.transpose %arg8, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc15) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_17", + PartOfOutputName = "Conv_17", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>, tensor<16x3x3x3xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc15) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + %173 = xten_nn.subgraph (%arg5 = %172: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_19", OutputName = "Add_19"} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + %174 = xten_nn.subgraph (%arg5 = %173: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_22", + OutputName = "Clip_22", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + %175 = xten_nn.subgraph (%arg5 = %174: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "410", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "410", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_24", + OutputName = "Div_24", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc18) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc18) + } -> tensor<1x16x90x160xbf16> loc(#loc18) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc18) + } -> tensor<1x16x90x160xbf16> loc(#loc18) + %176 = xten_nn.subgraph (%arg5 = %172: tensor<1x16x90x160xbf16>, %arg6 = %175: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_25", + OutputName = "Mul_25", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc19) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc19) + } -> tensor<1x16x90x160xbf16> loc(#loc19) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc19) + } -> tensor<1x16x90x160xbf16> loc(#loc19) + %177 = xten_nn.subgraph (%arg5 = %176: tensor<1x16x90x160xbf16>, %arg6 = %158: tensor<16x1x3x3xbf16>, %arg7 = %157: tensor<16xbf16>) attributes { + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "411", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<16x1x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "411", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc317) + %465 = tosa.transpose %arg9, %464 : (tensor<16x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x16x1xbf16> loc(#loc317) + %466 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc317) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_26", + PartOfOutputName = "Conv_26", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<3x3x16x1xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %468 = tosa.clamp %467 { + LayerName = "Relu_27", + OutputName = "Relu_27", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc21) + %469 = tosa.transpose %468, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc317) + xten_nn.output %469 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc317) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc317) + } -> tensor<1x16x90x160xbf16> loc(#loc317) + %178 = xten_nn.subgraph (%arg5 = %177: tensor<1x16x90x160xbf16>, %arg6 = %156: tensor<16x16x1x1xbf16>, %arg7 = %155: tensor<16xbf16>, %arg8 = %176: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Conv_28", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "931", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "935", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x16x90x160xbf16>, %arg10 = %arg6: tensor<16x16x1x1xbf16>, %arg11 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_28", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "931", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "935", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_28", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "934", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc22) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<16x16x1x1xbf16>) -> tensor<16x1x1x16xbf16> loc(#loc22) + %466 = tosa.transpose %arg9, %464 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc22) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_28", + PartOfOutputName = "Conv_28", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<16x1x1x16xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc22) + %468 = tosa.transpose %467, %463 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc22) + xten_nn.output %468 : tensor<1x16x90x160xbf16> loc(#loc22) + } -> tensor<1x16x90x160xbf16> loc(#loc22) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x16x90x160xbf16>, %arg10 = %arg8: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_29", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "934", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_29", OutputName = "Add_29"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc23) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc23) + } -> tensor<1x16x90x160xbf16> loc(#loc23) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc23) + } -> tensor<1x16x90x160xbf16> loc(#loc318) + %179 = xten_nn.subgraph (%arg5 = %178: tensor<1x16x90x160xbf16>, %arg6 = %154: tensor<64x16x1x1xbf16>, %arg7 = %153: tensor<64xbf16>) attributes { + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "934", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "417", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<64x16x1x1xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "934", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "417", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc319) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<64x16x1x1xbf16>) -> tensor<64x1x1x16xbf16> loc(#loc319) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc319) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_30", + PartOfOutputName = "Conv_30", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<64x1x1x16xbf16>, tensor<64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc24) + %467 = tosa.clamp %466 { + LayerName = "Relu_31", + OutputName = "Relu_31", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc25) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x64xbf16>, tensor<4xi32>) -> tensor<1x64x90x160xbf16> loc(#loc319) + xten_nn.output %468 : tensor<1x64x90x160xbf16> loc(#loc25) + } -> tensor<1x64x90x160xbf16> loc(#loc319) + xten_nn.output %461 : tensor<1x64x90x160xbf16> loc(#loc319) + } -> tensor<1x64x90x160xbf16> loc(#loc319) + %180 = xten_nn.subgraph (%arg5 = %179: tensor<1x64x90x160xbf16>, %arg6 = %152: tensor<64x1x3x3xbf16>, %arg7 = %151: tensor<64xbf16>) attributes { + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "937", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "941", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x90x160xbf16>, %arg9 = %arg6: tensor<64x1x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "937", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "941", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc320) + %465 = tosa.transpose %arg9, %464 : (tensor<64x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x64x1xbf16> loc(#loc320) + %466 = tosa.transpose %arg8, %463 : (tensor<1x64x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x64xbf16> loc(#loc320) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_32", + PartOfOutputName = "Conv_32", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x64xbf16>, tensor<3x3x64x1xbf16>, tensor<64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc26) + %468 = tosa.clamp %467 { + LayerName = "Relu_33", + OutputName = "Relu_33", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc27) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x64xbf16>, tensor<4xi32>) -> tensor<1x64x45x80xbf16> loc(#loc320) + xten_nn.output %469 : tensor<1x64x45x80xbf16> loc(#loc27) + } -> tensor<1x64x45x80xbf16> loc(#loc320) + xten_nn.output %461 : tensor<1x64x45x80xbf16> loc(#loc320) + } -> tensor<1x64x45x80xbf16> loc(#loc320) + %181 = xten_nn.subgraph (%arg5 = %180: tensor<1x64x45x80xbf16>, %arg6 = %150: tensor<24x64x1x1xbf16>, %arg7 = %149: tensor<24xbf16>) attributes { + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Name = "940", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Name = "944", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x45x80xbf16>, %arg9 = %arg6: tensor<24x64x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Name = "940", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Name = "944", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc28) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x64x1x1xbf16>) -> tensor<24x1x1x64xbf16> loc(#loc28) + %465 = tosa.transpose %arg8, %463 : (tensor<1x64x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x64xbf16> loc(#loc28) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_34", + PartOfOutputName = "Conv_34", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x64xbf16>, tensor<24x1x1x64xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc28) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc28) + xten_nn.output %467 : tensor<1x24x45x80xbf16> loc(#loc28) + } -> tensor<1x24x45x80xbf16> loc(#loc28) + xten_nn.output %461 : tensor<1x24x45x80xbf16> loc(#loc28) + } -> tensor<1x24x45x80xbf16> loc(#loc28) + %182 = xten_nn.subgraph (%arg5 = %181: tensor<1x24x45x80xbf16>, %arg6 = %148: tensor<72x24x1x1xbf16>, %arg7 = %147: tensor<72xbf16>) attributes { + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "423", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "947", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "423", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "947", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc321) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc321) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc321) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_35", + PartOfOutputName = "Conv_35", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc29) + %467 = tosa.clamp %466 { + LayerName = "Relu_36", + OutputName = "Relu_36", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc30) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc321) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc30) + } -> tensor<1x72x45x80xbf16> loc(#loc321) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc321) + } -> tensor<1x72x45x80xbf16> loc(#loc321) + %183 = xten_nn.subgraph (%arg5 = %182: tensor<1x72x45x80xbf16>, %arg6 = %146: tensor<72x1x3x3xbf16>, %arg7 = %145: tensor<72xbf16>) attributes { + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "946", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "950", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x3x3xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "946", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "950", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc322) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x72x1xbf16> loc(#loc322) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc322) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_37", + PartOfOutputName = "Conv_37", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<3x3x72x1xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc31) + %468 = tosa.clamp %467 { + LayerName = "Relu_38", + OutputName = "Relu_38", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc32) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc322) + xten_nn.output %469 : tensor<1x72x45x80xbf16> loc(#loc32) + } -> tensor<1x72x45x80xbf16> loc(#loc322) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc322) + } -> tensor<1x72x45x80xbf16> loc(#loc322) + %184 = xten_nn.subgraph (%arg5 = %183: tensor<1x72x45x80xbf16>, %arg6 = %144: tensor<24x72x1x1xbf16>, %arg7 = %143: tensor<24xbf16>, %arg8 = %181: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Conv_39", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Name = "949", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "953", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x72x45x80xbf16>, %arg10 = %arg6: tensor<24x72x1x1xbf16>, %arg11 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_39", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Name = "949", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "953", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_39", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "952", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc33) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc33) + %466 = tosa.transpose %arg9, %464 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc33) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_39", + PartOfOutputName = "Conv_39", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc33) + %468 = tosa.transpose %467, %463 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc33) + xten_nn.output %468 : tensor<1x24x45x80xbf16> loc(#loc33) + } -> tensor<1x24x45x80xbf16> loc(#loc33) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x24x45x80xbf16>, %arg10 = %arg8: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Add_40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "952", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_40", OutputName = "Add_40"} : (tensor<1x24x45x80xbf16>, tensor<1x24x45x80xbf16>) -> tensor<1x24x45x80xbf16> loc(#loc34) + xten_nn.output %463 : tensor<1x24x45x80xbf16> loc(#loc34) + } -> tensor<1x24x45x80xbf16> loc(#loc34) + xten_nn.output %462 : tensor<1x24x45x80xbf16> loc(#loc34) + } -> tensor<1x24x45x80xbf16> loc(#loc323) + %185 = xten_nn.subgraph (%arg5 = %184: tensor<1x24x45x80xbf16>, %arg6 = %142: tensor<72x24x1x1xbf16>, %arg7 = %141: tensor<72xbf16>) attributes { + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "952", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "434", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "952", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "434", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc324) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc324) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc324) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_41", + PartOfOutputName = "Conv_41", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc35) + %467 = tosa.clamp %466 { + LayerName = "Relu_42", + OutputName = "Relu_42", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc36) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc324) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc36) + } -> tensor<1x72x45x80xbf16> loc(#loc324) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc324) + } -> tensor<1x72x45x80xbf16> loc(#loc324) + %186 = xten_nn.subgraph (%arg5 = %185: tensor<1x72x45x80xbf16>, %arg6 = %140: tensor<72x1x5x5xbf16>, %arg7 = %139: tensor<72xbf16>) attributes { + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "955", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "959", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x5x5xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 1]], + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "955", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "959", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc325) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x72x1xbf16> loc(#loc325) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc325) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_43", + PartOfOutputName = "Conv_43", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<5x5x72x1xbf16>, tensor<72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc37) + %468 = tosa.clamp %467 { + LayerName = "Relu_44", + OutputName = "Relu_44", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc38) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x72xbf16>, tensor<4xi32>) -> tensor<1x72x23x40xbf16> loc(#loc325) + xten_nn.output %469 : tensor<1x72x23x40xbf16> loc(#loc38) + } -> tensor<1x72x23x40xbf16> loc(#loc325) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc325) + } -> tensor<1x72x23x40xbf16> loc(#loc325) + %187 = xten_nn.subgraph (%arg5 = %186: tensor<1x72x23x40xbf16>) attributes { + LayerName = "GlobalAveragePool_45_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_45_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 9 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x72x23x40xbf16>) -> tensor<1x72x1x920xbf16> loc(#loc39) + xten_nn.output %461 : tensor<1x72x1x920xbf16> loc(#loc39) + } -> tensor<1x72x1x920xbf16> loc(#loc39) + %188 = xten_nn.subgraph (%arg5 = %187: tensor<1x72x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_45_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_45_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_45_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_45_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 72 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x72x1x920xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc39) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc39) + } -> tensor<1x72x1x1xbf16> loc(#loc39) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc39) + } -> tensor<1x72x1x1xbf16> loc(#loc39) + %189 = xten_nn.subgraph (%arg5 = %188: tensor<1x72x1x1xbf16>, %arg6 = %138: tensor<24x72x1x1xbf16>, %arg7 = %137: tensor<24xbf16>) attributes { + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "440", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x1x1xbf16>, %arg9 = %arg6: tensor<24x72x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "440", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc326) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc326) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_46", + PartOfOutputName = "Conv_46", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc40) + %465 = tosa.clamp %464 { + LayerName = "Relu_47", + OutputName = "Relu_47", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc41) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x24xbf16>) -> tensor<1x24x1x1xbf16> loc(#loc326) + xten_nn.output %466 : tensor<1x24x1x1xbf16> loc(#loc41) + } -> tensor<1x24x1x1xbf16> loc(#loc326) + xten_nn.output %461 : tensor<1x24x1x1xbf16> loc(#loc326) + } -> tensor<1x24x1x1xbf16> loc(#loc326) + %190 = xten_nn.subgraph (%arg5 = %189: tensor<1x24x1x1xbf16>, %arg6 = %136: tensor<72x24x1x1xbf16>, %arg7 = %135: tensor<72xbf16>) attributes { + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "442", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x1x1xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "442", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc42) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x24x1x1xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc42) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_48", + PartOfOutputName = "Conv_48", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc42) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x72xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + %191 = xten_nn.subgraph (%arg5 = %190: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_50", OutputName = "Add_50"} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + %192 = xten_nn.subgraph (%arg5 = %191: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_53", + OutputName = "Clip_53", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc44) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc44) + } -> tensor<1x72x1x1xbf16> loc(#loc44) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc44) + } -> tensor<1x72x1x1xbf16> loc(#loc44) + %193 = xten_nn.subgraph (%arg5 = %192: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_55", + OutputName = "Div_55", + shift = 0 : i8} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc45) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc45) + } -> tensor<1x72x1x1xbf16> loc(#loc45) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc45) + } -> tensor<1x72x1x1xbf16> loc(#loc45) + %194 = xten_nn.subgraph (%arg5 = %193: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Mul_56_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_56_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 72 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc46) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc46) + } -> tensor<1x72x23x40xbf16> loc(#loc46) + %195 = xten_nn.subgraph (%arg5 = %194: tensor<1x72x23x40xbf16>, %arg6 = %186: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x72x23x40xbf16>, %arg8 = %arg6: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_56", + OutputName = "Mul_56", + shift = 0 : i8} : (tensor<1x72x23x40xbf16>, tensor<1x72x23x40xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc46) + xten_nn.output %462 : tensor<1x72x23x40xbf16> loc(#loc46) + } -> tensor<1x72x23x40xbf16> loc(#loc46) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc46) + } -> tensor<1x72x23x40xbf16> loc(#loc46) + %196 = xten_nn.subgraph (%arg5 = %195: tensor<1x72x23x40xbf16>, %arg6 = %134: tensor<40x72x1x1xbf16>, %arg7 = %133: tensor<40xbf16>) attributes { + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Name = "451", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "452", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x23x40xbf16>, %arg9 = %arg6: tensor<40x72x1x1xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Name = "451", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "452", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc47) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<40x72x1x1xbf16>) -> tensor<40x1x1x72xbf16> loc(#loc47) + %465 = tosa.transpose %arg8, %463 : (tensor<1x72x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x72xbf16> loc(#loc47) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_57", + PartOfOutputName = "Conv_57", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x72xbf16>, tensor<40x1x1x72xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc47) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc47) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc47) + } -> tensor<1x40x23x40xbf16> loc(#loc47) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc47) + } -> tensor<1x40x23x40xbf16> loc(#loc47) + %197 = xten_nn.subgraph (%arg5 = %196: tensor<1x40x23x40xbf16>, %arg6 = %132: tensor<120x40x1x1xbf16>, %arg7 = %131: tensor<120xbf16>) attributes { + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "452", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "965", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "452", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "965", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc327) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc327) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc327) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_58", + PartOfOutputName = "Conv_58", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc48) + %467 = tosa.clamp %466 { + LayerName = "Relu_59", + OutputName = "Relu_59", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc49) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc327) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc49) + } -> tensor<1x120x23x40xbf16> loc(#loc327) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc327) + } -> tensor<1x120x23x40xbf16> loc(#loc327) + %198 = xten_nn.subgraph (%arg5 = %197: tensor<1x120x23x40xbf16>, %arg6 = %130: tensor<120x1x5x5xbf16>, %arg7 = %129: tensor<120xbf16>) attributes { + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "964", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "968", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "964", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "968", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc328) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc328) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc328) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_60", + PartOfOutputName = "Conv_60", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc50) + %468 = tosa.clamp %467 { + LayerName = "Relu_61", + OutputName = "Relu_61", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc51) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc328) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc51) + } -> tensor<1x120x23x40xbf16> loc(#loc328) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc328) + } -> tensor<1x120x23x40xbf16> loc(#loc328) + %199 = xten_nn.subgraph (%arg5 = %198: tensor<1x120x23x40xbf16>) attributes { + LayerName = "GlobalAveragePool_62_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_62_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc52) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc52) + } -> tensor<1x120x1x920xbf16> loc(#loc52) + %200 = xten_nn.subgraph (%arg5 = %199: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_62_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_62_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_62_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_62_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc52) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc52) + } -> tensor<1x120x1x1xbf16> loc(#loc52) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc52) + } -> tensor<1x120x1x1xbf16> loc(#loc52) + %201 = xten_nn.subgraph (%arg5 = %200: tensor<1x120x1x1xbf16>, %arg6 = %128: tensor<32x120x1x1xbf16>, %arg7 = %127: tensor<32xbf16>) attributes { + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "460", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "460", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc329) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc329) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_63", + PartOfOutputName = "Conv_63", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc53) + %465 = tosa.clamp %464 { + LayerName = "Relu_64", + OutputName = "Relu_64", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc54) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc329) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc54) + } -> tensor<1x32x1x1xbf16> loc(#loc329) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc329) + } -> tensor<1x32x1x1xbf16> loc(#loc329) + %202 = xten_nn.subgraph (%arg5 = %201: tensor<1x32x1x1xbf16>, %arg6 = %126: tensor<120x32x1x1xbf16>, %arg7 = %125: tensor<120xbf16>) attributes { + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "462", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "462", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc55) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc55) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_65", + PartOfOutputName = "Conv_65", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc55) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + %203 = xten_nn.subgraph (%arg5 = %202: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_67", OutputName = "Add_67"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + %204 = xten_nn.subgraph (%arg5 = %203: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_70", + OutputName = "Clip_70", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc57) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc57) + } -> tensor<1x120x1x1xbf16> loc(#loc57) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc57) + } -> tensor<1x120x1x1xbf16> loc(#loc57) + %205 = xten_nn.subgraph (%arg5 = %204: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_72", + OutputName = "Div_72", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc58) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc58) + } -> tensor<1x120x1x1xbf16> loc(#loc58) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc58) + } -> tensor<1x120x1x1xbf16> loc(#loc58) + %206 = xten_nn.subgraph (%arg5 = %205: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Mul_73_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_73_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc59) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc59) + } -> tensor<1x120x23x40xbf16> loc(#loc59) + %207 = xten_nn.subgraph (%arg5 = %206: tensor<1x120x23x40xbf16>, %arg6 = %198: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_73", + OutputName = "Mul_73", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc59) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc59) + } -> tensor<1x120x23x40xbf16> loc(#loc59) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc59) + } -> tensor<1x120x23x40xbf16> loc(#loc59) + %208 = xten_nn.subgraph (%arg5 = %207: tensor<1x120x23x40xbf16>, %arg6 = %124: tensor<40x120x1x1xbf16>, %arg7 = %123: tensor<40xbf16>, %arg8 = %196: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_74", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "471", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "472", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_74", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "471", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "472", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_74", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "970", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc60) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc60) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc60) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_74", + PartOfOutputName = "Conv_74", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc60) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc60) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc60) + } -> tensor<1x40x23x40xbf16> loc(#loc60) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_75", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "970", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_75", OutputName = "Add_75"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc61) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc61) + } -> tensor<1x40x23x40xbf16> loc(#loc61) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc61) + } -> tensor<1x40x23x40xbf16> loc(#loc330) + %209 = xten_nn.subgraph (%arg5 = %208: tensor<1x40x23x40xbf16>, %arg6 = %122: tensor<120x40x1x1xbf16>, %arg7 = %121: tensor<120xbf16>) attributes { + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "970", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "475", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "970", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "475", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc331) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc331) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc331) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_76", + PartOfOutputName = "Conv_76", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc62) + %467 = tosa.clamp %466 { + LayerName = "Relu_77", + OutputName = "Relu_77", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc63) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc331) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc63) + } -> tensor<1x120x23x40xbf16> loc(#loc331) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc331) + } -> tensor<1x120x23x40xbf16> loc(#loc331) + %210 = xten_nn.subgraph (%arg5 = %209: tensor<1x120x23x40xbf16>, %arg6 = %120: tensor<120x1x5x5xbf16>, %arg7 = %119: tensor<120xbf16>) attributes { + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "973", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "977", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "973", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "977", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc332) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc332) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc332) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_78", + PartOfOutputName = "Conv_78", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc64) + %468 = tosa.clamp %467 { + LayerName = "Relu_79", + OutputName = "Relu_79", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc65) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc332) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc65) + } -> tensor<1x120x23x40xbf16> loc(#loc332) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc332) + } -> tensor<1x120x23x40xbf16> loc(#loc332) + %211 = xten_nn.subgraph (%arg5 = %210: tensor<1x120x23x40xbf16>) attributes { + LayerName = "GlobalAveragePool_80_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_80_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc66) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc66) + } -> tensor<1x120x1x920xbf16> loc(#loc66) + %212 = xten_nn.subgraph (%arg5 = %211: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_80_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_80_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_80_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_80_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc66) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc66) + } -> tensor<1x120x1x1xbf16> loc(#loc66) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc66) + } -> tensor<1x120x1x1xbf16> loc(#loc66) + %213 = xten_nn.subgraph (%arg5 = %212: tensor<1x120x1x1xbf16>, %arg6 = %118: tensor<32x120x1x1xbf16>, %arg7 = %117: tensor<32xbf16>) attributes { + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "481", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "481", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc333) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc333) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_81", + PartOfOutputName = "Conv_81", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc67) + %465 = tosa.clamp %464 { + LayerName = "Relu_82", + OutputName = "Relu_82", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc68) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc333) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc68) + } -> tensor<1x32x1x1xbf16> loc(#loc333) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc333) + } -> tensor<1x32x1x1xbf16> loc(#loc333) + %214 = xten_nn.subgraph (%arg5 = %213: tensor<1x32x1x1xbf16>, %arg6 = %116: tensor<120x32x1x1xbf16>, %arg7 = %115: tensor<120xbf16>) attributes { + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "483", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "483", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc69) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc69) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_83", + PartOfOutputName = "Conv_83", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc69) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + %215 = xten_nn.subgraph (%arg5 = %214: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_85", OutputName = "Add_85"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + %216 = xten_nn.subgraph (%arg5 = %215: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_88", + OutputName = "Clip_88", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc71) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc71) + } -> tensor<1x120x1x1xbf16> loc(#loc71) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc71) + } -> tensor<1x120x1x1xbf16> loc(#loc71) + %217 = xten_nn.subgraph (%arg5 = %216: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_90", + OutputName = "Div_90", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc72) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc72) + } -> tensor<1x120x1x1xbf16> loc(#loc72) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc72) + } -> tensor<1x120x1x1xbf16> loc(#loc72) + %218 = xten_nn.subgraph (%arg5 = %217: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Mul_91_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_91_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc73) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc73) + } -> tensor<1x120x23x40xbf16> loc(#loc73) + %219 = xten_nn.subgraph (%arg5 = %218: tensor<1x120x23x40xbf16>, %arg6 = %210: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_91", + OutputName = "Mul_91", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc73) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc73) + } -> tensor<1x120x23x40xbf16> loc(#loc73) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc73) + } -> tensor<1x120x23x40xbf16> loc(#loc73) + %220 = xten_nn.subgraph (%arg5 = %219: tensor<1x120x23x40xbf16>, %arg6 = %114: tensor<40x120x1x1xbf16>, %arg7 = %113: tensor<40xbf16>, %arg8 = %208: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_92", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "492", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "493", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_92", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "492", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "493", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_92", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "979", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc74) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc74) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc74) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_92", + PartOfOutputName = "Conv_92", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc74) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc74) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc74) + } -> tensor<1x40x23x40xbf16> loc(#loc74) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_93", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "979", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_93", OutputName = "Add_93"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc75) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc75) + } -> tensor<1x40x23x40xbf16> loc(#loc75) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc75) + } -> tensor<1x40x23x40xbf16> loc(#loc334) + %221 = xten_nn.subgraph (%arg5 = %220: tensor<1x40x23x40xbf16>, %arg6 = %112: tensor<240x40x1x1xbf16>, %arg7 = %111: tensor<240xbf16>) attributes { + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "979", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "496", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<240x40x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "979", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "496", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc76) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x40x1x1xbf16>) -> tensor<240x1x1x40xbf16> loc(#loc76) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc76) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_94", + PartOfOutputName = "Conv_94", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<240x1x1x40xbf16>, tensor<240xbf16>) -> tensor<1x23x40x240xbf16> loc(#loc76) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x240xbf16>, tensor<4xi32>) -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %467 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + %222 = xten_nn.subgraph (%arg5 = %221: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_96", OutputName = "Add_96"} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + %223 = xten_nn.subgraph (%arg5 = %222: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_99", + OutputName = "Clip_99", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + %224 = xten_nn.subgraph (%arg5 = %223: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "505", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "505", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_101", + OutputName = "Div_101", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc79) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc79) + } -> tensor<1x240x23x40xbf16> loc(#loc79) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc79) + } -> tensor<1x240x23x40xbf16> loc(#loc79) + %225 = xten_nn.subgraph (%arg5 = %221: tensor<1x240x23x40xbf16>, %arg6 = %224: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x23x40xbf16>, %arg8 = %arg6: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_102", + OutputName = "Mul_102", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc80) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc80) + } -> tensor<1x240x23x40xbf16> loc(#loc80) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc80) + } -> tensor<1x240x23x40xbf16> loc(#loc80) + %226 = xten_nn.subgraph (%arg5 = %225: tensor<1x240x23x40xbf16>, %arg6 = %110: tensor<240x1x3x3xbf16>, %arg7 = %109: tensor<240xbf16>) attributes { + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "506", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x23x40xbf16>, %arg9 = %arg6: tensor<240x1x3x3xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 0]], + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "506", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc81) + %465 = tosa.transpose %arg9, %464 : (tensor<240x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x240x1xbf16> loc(#loc81) + %466 = tosa.transpose %arg8, %463 : (tensor<1x240x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x240xbf16> loc(#loc81) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_103", + PartOfOutputName = "Conv_103", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x240xbf16>, tensor<3x3x240x1xbf16>, tensor<240xbf16>) -> tensor<1x12x20x240xbf16> loc(#loc81) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x240xbf16>, tensor<4xi32>) -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %468 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + %227 = xten_nn.subgraph (%arg5 = %226: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_105", OutputName = "Add_105"} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + %228 = xten_nn.subgraph (%arg5 = %227: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_108", + OutputName = "Clip_108", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + %229 = xten_nn.subgraph (%arg5 = %228: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "515", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "515", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_110", + OutputName = "Div_110", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc84) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc84) + } -> tensor<1x240x12x20xbf16> loc(#loc84) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc84) + } -> tensor<1x240x12x20xbf16> loc(#loc84) + %230 = xten_nn.subgraph (%arg5 = %226: tensor<1x240x12x20xbf16>, %arg6 = %229: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x12x20xbf16>, %arg8 = %arg6: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_111", + OutputName = "Mul_111", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc85) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc85) + } -> tensor<1x240x12x20xbf16> loc(#loc85) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc85) + } -> tensor<1x240x12x20xbf16> loc(#loc85) + %231 = xten_nn.subgraph (%arg5 = %230: tensor<1x240x12x20xbf16>, %arg6 = %108: tensor<80x240x1x1xbf16>, %arg7 = %107: tensor<80xbf16>) attributes { + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Name = "985", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "516", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x12x20xbf16>, %arg9 = %arg6: tensor<80x240x1x1xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Name = "985", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "516", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc86) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<80x240x1x1xbf16>) -> tensor<80x1x1x240xbf16> loc(#loc86) + %465 = tosa.transpose %arg8, %463 : (tensor<1x240x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x240xbf16> loc(#loc86) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_112", + PartOfOutputName = "Conv_112", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x240xbf16>, tensor<80x1x1x240xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc86) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc86) + xten_nn.output %467 : tensor<1x80x12x20xbf16> loc(#loc86) + } -> tensor<1x80x12x20xbf16> loc(#loc86) + xten_nn.output %461 : tensor<1x80x12x20xbf16> loc(#loc86) + } -> tensor<1x80x12x20xbf16> loc(#loc86) + %232 = xten_nn.subgraph (%arg5 = %231: tensor<1x80x12x20xbf16>, %arg6 = %106: tensor<200x80x1x1xbf16>, %arg7 = %105: tensor<200xbf16>) attributes { + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "516", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "992", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<200x80x1x1xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "516", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "992", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc87) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<200x80x1x1xbf16>) -> tensor<200x1x1x80xbf16> loc(#loc87) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc87) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_113", + PartOfOutputName = "Conv_113", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<200x1x1x80xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc87) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %467 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + %233 = xten_nn.subgraph (%arg5 = %232: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_115", OutputName = "Add_115"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + %234 = xten_nn.subgraph (%arg5 = %233: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_118", + OutputName = "Clip_118", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + %235 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "527", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "527", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_120", + OutputName = "Div_120", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + %236 = xten_nn.subgraph (%arg5 = %232: tensor<1x200x12x20xbf16>, %arg6 = %235: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_121", + OutputName = "Mul_121", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + %237 = xten_nn.subgraph (%arg5 = %236: tensor<1x200x12x20xbf16>, %arg6 = %104: tensor<200x1x3x3xbf16>, %arg7 = %103: tensor<200xbf16>) attributes { + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "528", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x200x12x20xbf16>, %arg9 = %arg6: tensor<200x1x3x3xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "528", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc92) + %465 = tosa.transpose %arg9, %464 : (tensor<200x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x200x1xbf16> loc(#loc92) + %466 = tosa.transpose %arg8, %463 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc92) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_122", + PartOfOutputName = "Conv_122", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<3x3x200x1xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc92) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %468 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + %238 = xten_nn.subgraph (%arg5 = %237: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_124", OutputName = "Add_124"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + %239 = xten_nn.subgraph (%arg5 = %238: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_127", + OutputName = "Clip_127", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + %240 = xten_nn.subgraph (%arg5 = %239: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "537", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "537", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_129", + OutputName = "Div_129", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc95) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc95) + } -> tensor<1x200x12x20xbf16> loc(#loc95) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc95) + } -> tensor<1x200x12x20xbf16> loc(#loc95) + %241 = xten_nn.subgraph (%arg5 = %237: tensor<1x200x12x20xbf16>, %arg6 = %240: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_130", + OutputName = "Mul_130", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc96) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc96) + } -> tensor<1x200x12x20xbf16> loc(#loc96) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc96) + } -> tensor<1x200x12x20xbf16> loc(#loc96) + %242 = xten_nn.subgraph (%arg5 = %241: tensor<1x200x12x20xbf16>, %arg6 = %102: tensor<80x200x1x1xbf16>, %arg7 = %101: tensor<80xbf16>, %arg8 = %231: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_131", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Name = "994", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Name = "538", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x200x12x20xbf16>, %arg10 = %arg6: tensor<80x200x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_131", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Name = "994", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Name = "538", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_131", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "997", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc97) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x200x1x1xbf16>) -> tensor<80x1x1x200xbf16> loc(#loc97) + %466 = tosa.transpose %arg9, %464 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc97) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_131", + PartOfOutputName = "Conv_131", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<80x1x1x200xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc97) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc97) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc97) + } -> tensor<1x80x12x20xbf16> loc(#loc97) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_132", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "997", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_132", OutputName = "Add_132"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc98) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc98) + } -> tensor<1x80x12x20xbf16> loc(#loc98) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc98) + } -> tensor<1x80x12x20xbf16> loc(#loc335) + %243 = xten_nn.subgraph (%arg5 = %242: tensor<1x80x12x20xbf16>, %arg6 = %100: tensor<184x80x1x1xbf16>, %arg7 = %99: tensor<184xbf16>) attributes { + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "997", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "541", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "997", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "541", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc99) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc99) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc99) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_133", + PartOfOutputName = "Conv_133", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc99) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + %244 = xten_nn.subgraph (%arg5 = %243: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_135", OutputName = "Add_135"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + %245 = xten_nn.subgraph (%arg5 = %244: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_138", + OutputName = "Clip_138", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + %246 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "550", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "550", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_140", + OutputName = "Div_140", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + %247 = xten_nn.subgraph (%arg5 = %243: tensor<1x184x12x20xbf16>, %arg6 = %246: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_141", + OutputName = "Mul_141", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + %248 = xten_nn.subgraph (%arg5 = %247: tensor<1x184x12x20xbf16>, %arg6 = %98: tensor<184x1x3x3xbf16>, %arg7 = %97: tensor<184xbf16>) attributes { + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "551", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "551", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc104) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc104) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc104) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_142", + PartOfOutputName = "Conv_142", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc104) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + %249 = xten_nn.subgraph (%arg5 = %248: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_144", OutputName = "Add_144"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + %250 = xten_nn.subgraph (%arg5 = %249: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_147", + OutputName = "Clip_147", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + %251 = xten_nn.subgraph (%arg5 = %250: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "560", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "560", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_149", + OutputName = "Div_149", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc107) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc107) + } -> tensor<1x184x12x20xbf16> loc(#loc107) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc107) + } -> tensor<1x184x12x20xbf16> loc(#loc107) + %252 = xten_nn.subgraph (%arg5 = %248: tensor<1x184x12x20xbf16>, %arg6 = %251: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_150", + OutputName = "Mul_150", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc108) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc108) + } -> tensor<1x184x12x20xbf16> loc(#loc108) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc108) + } -> tensor<1x184x12x20xbf16> loc(#loc108) + %253 = xten_nn.subgraph (%arg5 = %252: tensor<1x184x12x20xbf16>, %arg6 = %96: tensor<80x184x1x1xbf16>, %arg7 = %95: tensor<80xbf16>, %arg8 = %242: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_151", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1003", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "561", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_151", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1003", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "561", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_151", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1006", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc109) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc109) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc109) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_151", + PartOfOutputName = "Conv_151", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc109) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc109) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc109) + } -> tensor<1x80x12x20xbf16> loc(#loc109) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_152", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1006", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_152", OutputName = "Add_152"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc110) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc110) + } -> tensor<1x80x12x20xbf16> loc(#loc110) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc110) + } -> tensor<1x80x12x20xbf16> loc(#loc336) + %254 = xten_nn.subgraph (%arg5 = %253: tensor<1x80x12x20xbf16>, %arg6 = %94: tensor<184x80x1x1xbf16>, %arg7 = %93: tensor<184xbf16>) attributes { + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1006", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "564", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1006", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "564", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc111) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc111) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc111) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_153", + PartOfOutputName = "Conv_153", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc111) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + %255 = xten_nn.subgraph (%arg5 = %254: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_155", OutputName = "Add_155"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + %256 = xten_nn.subgraph (%arg5 = %255: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_158", + OutputName = "Clip_158", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + %257 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "573", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "573", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_160", + OutputName = "Div_160", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + %258 = xten_nn.subgraph (%arg5 = %254: tensor<1x184x12x20xbf16>, %arg6 = %257: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_161", + OutputName = "Mul_161", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + %259 = xten_nn.subgraph (%arg5 = %258: tensor<1x184x12x20xbf16>, %arg6 = %92: tensor<184x1x3x3xbf16>, %arg7 = %91: tensor<184xbf16>) attributes { + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "574", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "574", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc116) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc116) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc116) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_162", + PartOfOutputName = "Conv_162", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc116) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + %260 = xten_nn.subgraph (%arg5 = %259: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_164", OutputName = "Add_164"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + %261 = xten_nn.subgraph (%arg5 = %260: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_167", + OutputName = "Clip_167", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + %262 = xten_nn.subgraph (%arg5 = %261: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "583", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "583", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_169", + OutputName = "Div_169", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc119) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc119) + } -> tensor<1x184x12x20xbf16> loc(#loc119) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc119) + } -> tensor<1x184x12x20xbf16> loc(#loc119) + %263 = xten_nn.subgraph (%arg5 = %259: tensor<1x184x12x20xbf16>, %arg6 = %262: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_170", + OutputName = "Mul_170", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc120) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc120) + } -> tensor<1x184x12x20xbf16> loc(#loc120) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc120) + } -> tensor<1x184x12x20xbf16> loc(#loc120) + %264 = xten_nn.subgraph (%arg5 = %263: tensor<1x184x12x20xbf16>, %arg6 = %90: tensor<80x184x1x1xbf16>, %arg7 = %89: tensor<80xbf16>, %arg8 = %253: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_171", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1012", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "584", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_171", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1012", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "584", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_171", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1015", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc121) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc121) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc121) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_171", + PartOfOutputName = "Conv_171", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc121) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc121) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc121) + } -> tensor<1x80x12x20xbf16> loc(#loc121) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_172", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1015", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_172", OutputName = "Add_172"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc122) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc122) + } -> tensor<1x80x12x20xbf16> loc(#loc122) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc122) + } -> tensor<1x80x12x20xbf16> loc(#loc337) + %265 = xten_nn.subgraph (%arg5 = %264: tensor<1x80x12x20xbf16>, %arg6 = %88: tensor<480x80x1x1xbf16>, %arg7 = %87: tensor<480xbf16>) attributes { + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1015", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "587", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<480x80x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1015", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "587", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc123) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x80x1x1xbf16>) -> tensor<480x1x1x80xbf16> loc(#loc123) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc123) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_173", + PartOfOutputName = "Conv_173", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<480x1x1x80xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc123) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %467 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + %266 = xten_nn.subgraph (%arg5 = %265: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_175", OutputName = "Add_175"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + %267 = xten_nn.subgraph (%arg5 = %266: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_178", + OutputName = "Clip_178", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + %268 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "596", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "596", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_180", + OutputName = "Div_180", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + %269 = xten_nn.subgraph (%arg5 = %265: tensor<1x480x12x20xbf16>, %arg6 = %268: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_181", + OutputName = "Mul_181", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + %270 = xten_nn.subgraph (%arg5 = %269: tensor<1x480x12x20xbf16>, %arg6 = %86: tensor<480x1x3x3xbf16>, %arg7 = %85: tensor<480xbf16>) attributes { + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "597", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<480x1x3x3xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "597", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc128) + %465 = tosa.transpose %arg9, %464 : (tensor<480x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x480x1xbf16> loc(#loc128) + %466 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc128) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_182", + PartOfOutputName = "Conv_182", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<3x3x480x1xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc128) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %468 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + %271 = xten_nn.subgraph (%arg5 = %270: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_184", OutputName = "Add_184"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + %272 = xten_nn.subgraph (%arg5 = %271: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_187", + OutputName = "Clip_187", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + %273 = xten_nn.subgraph (%arg5 = %272: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "606", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "606", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_189", + OutputName = "Div_189", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc131) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc131) + } -> tensor<1x480x12x20xbf16> loc(#loc131) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc131) + } -> tensor<1x480x12x20xbf16> loc(#loc131) + %274 = xten_nn.subgraph (%arg5 = %270: tensor<1x480x12x20xbf16>, %arg6 = %273: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_190", + OutputName = "Mul_190", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc132) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc132) + } -> tensor<1x480x12x20xbf16> loc(#loc132) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc132) + } -> tensor<1x480x12x20xbf16> loc(#loc132) + %275 = xten_nn.subgraph (%arg5 = %274: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_191_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 60 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x1x240xbf16> loc(#loc338) + xten_nn.output %461 : tensor<1x480x1x240xbf16> loc(#loc338) + } -> tensor<1x480x1x240xbf16> loc(#loc338) + %276 = xten_nn.subgraph (%arg5 = %275: tensor<1x480x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_191", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_191_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_191", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_191_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 480 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x480x1x240xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc133) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc133) + } -> tensor<1x480x1x1xbf16> loc(#loc133) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc133) + } -> tensor<1x480x1x1xbf16> loc(#loc133) + %277 = xten_nn.subgraph (%arg5 = %276: tensor<1x480x1x1xbf16>, %arg6 = %84: tensor<120x480x1x1xbf16>, %arg7 = %83: tensor<120xbf16>) attributes { + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "607", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x1x1xbf16>, %arg9 = %arg6: tensor<120x480x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "607", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x480x1x1xbf16>) -> tensor<120x1x1x480xbf16> loc(#loc339) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc339) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_192", + PartOfOutputName = "Conv_192", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x480xbf16>, tensor<120x1x1x480xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc134) + %465 = tosa.clamp %464 { + LayerName = "Relu_193", + OutputName = "Relu_193", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc135) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc339) + xten_nn.output %466 : tensor<1x120x1x1xbf16> loc(#loc135) + } -> tensor<1x120x1x1xbf16> loc(#loc339) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc339) + } -> tensor<1x120x1x1xbf16> loc(#loc339) + %278 = xten_nn.subgraph (%arg5 = %277: tensor<1x120x1x1xbf16>, %arg6 = %82: tensor<480x120x1x1xbf16>, %arg7 = %81: tensor<480xbf16>) attributes { + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "609", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<480x120x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "609", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x120x1x1xbf16>) -> tensor<480x1x1x120xbf16> loc(#loc136) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc136) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_194", + PartOfOutputName = "Conv_194", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<480x1x1x120xbf16>, tensor<480xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc136) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x480xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + %279 = xten_nn.subgraph (%arg5 = %278: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_196", OutputName = "Add_196"} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + %280 = xten_nn.subgraph (%arg5 = %279: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_199", + OutputName = "Clip_199", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc138) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc138) + } -> tensor<1x480x1x1xbf16> loc(#loc138) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc138) + } -> tensor<1x480x1x1xbf16> loc(#loc138) + %281 = xten_nn.subgraph (%arg5 = %280: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_201", + OutputName = "Div_201", + shift = 0 : i8} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc139) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc139) + } -> tensor<1x480x1x1xbf16> loc(#loc139) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc139) + } -> tensor<1x480x1x1xbf16> loc(#loc139) + %282 = xten_nn.subgraph (%arg5 = %281: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Mul_202_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_202_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 480 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc140) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc140) + } -> tensor<1x480x12x20xbf16> loc(#loc140) + %283 = xten_nn.subgraph (%arg5 = %282: tensor<1x480x12x20xbf16>, %arg6 = %274: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_202", + OutputName = "Mul_202", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc140) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc140) + } -> tensor<1x480x12x20xbf16> loc(#loc140) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc140) + } -> tensor<1x480x12x20xbf16> loc(#loc140) + %284 = xten_nn.subgraph (%arg5 = %283: tensor<1x480x12x20xbf16>, %arg6 = %80: tensor<112x480x1x1xbf16>, %arg7 = %79: tensor<112xbf16>) attributes { + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Name = "618", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "619", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<112x480x1x1xbf16>, %arg10 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Name = "618", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "619", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc141) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<112x480x1x1xbf16>) -> tensor<112x1x1x480xbf16> loc(#loc141) + %465 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc141) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_203", + PartOfOutputName = "Conv_203", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<112x1x1x480xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc141) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc141) + xten_nn.output %467 : tensor<1x112x12x20xbf16> loc(#loc141) + } -> tensor<1x112x12x20xbf16> loc(#loc141) + xten_nn.output %461 : tensor<1x112x12x20xbf16> loc(#loc141) + } -> tensor<1x112x12x20xbf16> loc(#loc141) + %285 = xten_nn.subgraph (%arg5 = %284: tensor<1x112x12x20xbf16>, %arg6 = %78: tensor<672x112x1x1xbf16>, %arg7 = %77: tensor<672xbf16>) attributes { + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "619", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "1028", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "619", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "1028", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc142) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc142) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc142) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_204", + PartOfOutputName = "Conv_204", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc142) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + %286 = xten_nn.subgraph (%arg5 = %285: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_206", OutputName = "Add_206"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + %287 = xten_nn.subgraph (%arg5 = %286: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_209", + OutputName = "Clip_209", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + %288 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "630", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "630", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_211", + OutputName = "Div_211", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + %289 = xten_nn.subgraph (%arg5 = %285: tensor<1x672x12x20xbf16>, %arg6 = %288: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_212", + OutputName = "Mul_212", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + %290 = xten_nn.subgraph (%arg5 = %289: tensor<1x672x12x20xbf16>, %arg6 = %76: tensor<672x1x3x3xbf16>, %arg7 = %75: tensor<672xbf16>) attributes { + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "631", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x3x3xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "631", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc147) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x672x1xbf16> loc(#loc147) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc147) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_213", + PartOfOutputName = "Conv_213", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<3x3x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc147) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + %291 = xten_nn.subgraph (%arg5 = %290: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_215", OutputName = "Add_215"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + %292 = xten_nn.subgraph (%arg5 = %291: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_218", + OutputName = "Clip_218", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + %293 = xten_nn.subgraph (%arg5 = %292: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "640", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "640", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_220", + OutputName = "Div_220", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc150) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc150) + } -> tensor<1x672x12x20xbf16> loc(#loc150) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc150) + } -> tensor<1x672x12x20xbf16> loc(#loc150) + %294 = xten_nn.subgraph (%arg5 = %290: tensor<1x672x12x20xbf16>, %arg6 = %293: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_221", + OutputName = "Mul_221", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc151) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc151) + } -> tensor<1x672x12x20xbf16> loc(#loc151) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc151) + } -> tensor<1x672x12x20xbf16> loc(#loc151) + %295 = xten_nn.subgraph (%arg5 = %294: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_222_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc340) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc340) + } -> tensor<1x672x1x240xbf16> loc(#loc340) + %296 = xten_nn.subgraph (%arg5 = %295: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_222", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_222_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_222", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_222_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc152) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc152) + } -> tensor<1x672x1x1xbf16> loc(#loc152) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc152) + } -> tensor<1x672x1x1xbf16> loc(#loc152) + %297 = xten_nn.subgraph (%arg5 = %296: tensor<1x672x1x1xbf16>, %arg6 = %74: tensor<168x672x1x1xbf16>, %arg7 = %73: tensor<168xbf16>) attributes { + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "641", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "641", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc341) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc341) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_223", + PartOfOutputName = "Conv_223", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc153) + %465 = tosa.clamp %464 { + LayerName = "Relu_224", + OutputName = "Relu_224", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc154) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc341) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc154) + } -> tensor<1x168x1x1xbf16> loc(#loc341) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc341) + } -> tensor<1x168x1x1xbf16> loc(#loc341) + %298 = xten_nn.subgraph (%arg5 = %297: tensor<1x168x1x1xbf16>, %arg6 = %72: tensor<672x168x1x1xbf16>, %arg7 = %71: tensor<672xbf16>) attributes { + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "643", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "643", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc155) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc155) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_225", + PartOfOutputName = "Conv_225", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc155) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + %299 = xten_nn.subgraph (%arg5 = %298: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_227", OutputName = "Add_227"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + %300 = xten_nn.subgraph (%arg5 = %299: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_230", + OutputName = "Clip_230", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc157) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc157) + } -> tensor<1x672x1x1xbf16> loc(#loc157) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc157) + } -> tensor<1x672x1x1xbf16> loc(#loc157) + %301 = xten_nn.subgraph (%arg5 = %300: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_232", + OutputName = "Div_232", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc158) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc158) + } -> tensor<1x672x1x1xbf16> loc(#loc158) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc158) + } -> tensor<1x672x1x1xbf16> loc(#loc158) + %302 = xten_nn.subgraph (%arg5 = %301: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Mul_233_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_233_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc159) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc159) + } -> tensor<1x672x12x20xbf16> loc(#loc159) + %303 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>, %arg6 = %294: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_233", + OutputName = "Mul_233", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc159) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc159) + } -> tensor<1x672x12x20xbf16> loc(#loc159) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc159) + } -> tensor<1x672x12x20xbf16> loc(#loc159) + %304 = xten_nn.subgraph (%arg5 = %303: tensor<1x672x12x20xbf16>, %arg6 = %70: tensor<112x672x1x1xbf16>, %arg7 = %69: tensor<112xbf16>, %arg8 = %284: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Conv_234", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "652", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "653", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x672x12x20xbf16>, %arg10 = %arg6: tensor<112x672x1x1xbf16>, %arg11 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_234", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "652", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "653", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_234", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1033", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc160) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<112x672x1x1xbf16>) -> tensor<112x1x1x672xbf16> loc(#loc160) + %466 = tosa.transpose %arg9, %464 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc160) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_234", + PartOfOutputName = "Conv_234", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<112x1x1x672xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc160) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc160) + xten_nn.output %468 : tensor<1x112x12x20xbf16> loc(#loc160) + } -> tensor<1x112x12x20xbf16> loc(#loc160) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x112x12x20xbf16>, %arg10 = %arg8: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Add_235", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1033", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_235", OutputName = "Add_235"} : (tensor<1x112x12x20xbf16>, tensor<1x112x12x20xbf16>) -> tensor<1x112x12x20xbf16> loc(#loc161) + xten_nn.output %463 : tensor<1x112x12x20xbf16> loc(#loc161) + } -> tensor<1x112x12x20xbf16> loc(#loc161) + xten_nn.output %462 : tensor<1x112x12x20xbf16> loc(#loc161) + } -> tensor<1x112x12x20xbf16> loc(#loc342) + %305 = xten_nn.subgraph (%arg5 = %304: tensor<1x112x12x20xbf16>, %arg6 = %68: tensor<672x112x1x1xbf16>, %arg7 = %67: tensor<672xbf16>) attributes { + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "1033", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "656", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "1033", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "656", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc162) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc162) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc162) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_236", + PartOfOutputName = "Conv_236", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc162) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + %306 = xten_nn.subgraph (%arg5 = %305: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_238", OutputName = "Add_238"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + %307 = xten_nn.subgraph (%arg5 = %306: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_241", + OutputName = "Clip_241", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + %308 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "665", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "665", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_243", + OutputName = "Div_243", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + %309 = xten_nn.subgraph (%arg5 = %305: tensor<1x672x12x20xbf16>, %arg6 = %308: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_244", + OutputName = "Mul_244", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + %310 = xten_nn.subgraph (%arg5 = %309: tensor<1x672x12x20xbf16>, %arg6 = %66: tensor<672x1x9x9xbf16>, %arg7 = %65: tensor<672xbf16>) attributes { + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "666", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x9x9xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "666", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc167) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x672x1xbf16> loc(#loc167) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc167) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_245", + PartOfOutputName = "Conv_245", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<9x9x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc167) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + %311 = xten_nn.subgraph (%arg5 = %310: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_247", OutputName = "Add_247"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + %312 = xten_nn.subgraph (%arg5 = %311: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_250", + OutputName = "Clip_250", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + %313 = xten_nn.subgraph (%arg5 = %312: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "675", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "675", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_252", + OutputName = "Div_252", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc170) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc170) + } -> tensor<1x672x12x20xbf16> loc(#loc170) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc170) + } -> tensor<1x672x12x20xbf16> loc(#loc170) + %314 = xten_nn.subgraph (%arg5 = %310: tensor<1x672x12x20xbf16>, %arg6 = %313: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_253", + OutputName = "Mul_253", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc171) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc171) + } -> tensor<1x672x12x20xbf16> loc(#loc171) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc171) + } -> tensor<1x672x12x20xbf16> loc(#loc171) + %315 = xten_nn.subgraph (%arg5 = %314: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_254_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc343) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc343) + } -> tensor<1x672x1x240xbf16> loc(#loc343) + %316 = xten_nn.subgraph (%arg5 = %315: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_254", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_254_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_254", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_254_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc172) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc172) + } -> tensor<1x672x1x1xbf16> loc(#loc172) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc172) + } -> tensor<1x672x1x1xbf16> loc(#loc172) + %317 = xten_nn.subgraph (%arg5 = %316: tensor<1x672x1x1xbf16>, %arg6 = %64: tensor<168x672x1x1xbf16>, %arg7 = %63: tensor<168xbf16>) attributes { + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "676", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "676", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc344) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc344) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_255", + PartOfOutputName = "Conv_255", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc173) + %465 = tosa.clamp %464 { + LayerName = "Relu_256", + OutputName = "Relu_256", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc174) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc344) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc174) + } -> tensor<1x168x1x1xbf16> loc(#loc344) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc344) + } -> tensor<1x168x1x1xbf16> loc(#loc344) + %318 = xten_nn.subgraph (%arg5 = %317: tensor<1x168x1x1xbf16>, %arg6 = %62: tensor<672x168x1x1xbf16>, %arg7 = %61: tensor<672xbf16>) attributes { + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "678", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "678", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc175) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc175) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_257", + PartOfOutputName = "Conv_257", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc175) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + %319 = xten_nn.subgraph (%arg5 = %318: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_259", OutputName = "Add_259"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + %320 = xten_nn.subgraph (%arg5 = %319: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_262", + OutputName = "Clip_262", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc177) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc177) + } -> tensor<1x672x1x1xbf16> loc(#loc177) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc177) + } -> tensor<1x672x1x1xbf16> loc(#loc177) + %321 = xten_nn.subgraph (%arg5 = %320: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_264", + OutputName = "Div_264", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc178) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc178) + } -> tensor<1x672x1x1xbf16> loc(#loc178) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc178) + } -> tensor<1x672x1x1xbf16> loc(#loc178) + %322 = xten_nn.subgraph (%arg5 = %321: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Mul_265_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_265_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc179) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc179) + } -> tensor<1x672x12x20xbf16> loc(#loc179) + %323 = xten_nn.subgraph (%arg5 = %322: tensor<1x672x12x20xbf16>, %arg6 = %314: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_265", + OutputName = "Mul_265", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc179) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc179) + } -> tensor<1x672x12x20xbf16> loc(#loc179) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc179) + } -> tensor<1x672x12x20xbf16> loc(#loc179) + %324 = xten_nn.subgraph (%arg5 = %323: tensor<1x672x12x20xbf16>, %arg6 = %60: tensor<160x672x1x1xbf16>, %arg7 = %59: tensor<160xbf16>) attributes { + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "687", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "688", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<160x672x1x1xbf16>, %arg10 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "687", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "688", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc180) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<160x672x1x1xbf16>) -> tensor<160x1x1x672xbf16> loc(#loc180) + %465 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc180) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_266", + PartOfOutputName = "Conv_266", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<160x1x1x672xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc180) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc180) + xten_nn.output %467 : tensor<1x160x12x20xbf16> loc(#loc180) + } -> tensor<1x160x12x20xbf16> loc(#loc180) + xten_nn.output %461 : tensor<1x160x12x20xbf16> loc(#loc180) + } -> tensor<1x160x12x20xbf16> loc(#loc180) + %325 = xten_nn.subgraph (%arg5 = %324: tensor<1x160x12x20xbf16>, %arg6 = %58: tensor<960x160x1x1xbf16>, %arg7 = %57: tensor<960xbf16>) attributes { + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "688", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "1046", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "688", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "1046", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc181) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc181) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc181) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_267", + PartOfOutputName = "Conv_267", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc181) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + %326 = xten_nn.subgraph (%arg5 = %325: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_269", OutputName = "Add_269"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + %327 = xten_nn.subgraph (%arg5 = %326: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_272", + OutputName = "Clip_272", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + %328 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "699", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "699", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_274", + OutputName = "Div_274", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + %329 = xten_nn.subgraph (%arg5 = %325: tensor<1x960x12x20xbf16>, %arg6 = %328: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_275", + OutputName = "Mul_275", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + %330 = xten_nn.subgraph (%arg5 = %329: tensor<1x960x12x20xbf16>, %arg6 = %56: tensor<960x1x9x9xbf16>, %arg7 = %55: tensor<960xbf16>) attributes { + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "700", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "700", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc186) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc186) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc186) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_276", + PartOfOutputName = "Conv_276", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc186) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + %331 = xten_nn.subgraph (%arg5 = %330: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_278", OutputName = "Add_278"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + %332 = xten_nn.subgraph (%arg5 = %331: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_281", + OutputName = "Clip_281", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + %333 = xten_nn.subgraph (%arg5 = %332: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "709", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "709", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_283", + OutputName = "Div_283", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc189) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc189) + } -> tensor<1x960x12x20xbf16> loc(#loc189) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc189) + } -> tensor<1x960x12x20xbf16> loc(#loc189) + %334 = xten_nn.subgraph (%arg5 = %330: tensor<1x960x12x20xbf16>, %arg6 = %333: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_284", + OutputName = "Mul_284", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc190) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc190) + } -> tensor<1x960x12x20xbf16> loc(#loc190) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc190) + } -> tensor<1x960x12x20xbf16> loc(#loc190) + %335 = xten_nn.subgraph (%arg5 = %334: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_285_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc345) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc345) + } -> tensor<1x960x1x240xbf16> loc(#loc345) + %336 = xten_nn.subgraph (%arg5 = %335: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_285", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_285_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_285", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_285_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc191) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc191) + } -> tensor<1x960x1x1xbf16> loc(#loc191) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc191) + } -> tensor<1x960x1x1xbf16> loc(#loc191) + %337 = xten_nn.subgraph (%arg5 = %336: tensor<1x960x1x1xbf16>, %arg6 = %54: tensor<240x960x1x1xbf16>, %arg7 = %53: tensor<240xbf16>) attributes { + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "710", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "710", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc346) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc346) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_286", + PartOfOutputName = "Conv_286", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc192) + %465 = tosa.clamp %464 { + LayerName = "Relu_287", + OutputName = "Relu_287", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc193) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc346) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc193) + } -> tensor<1x240x1x1xbf16> loc(#loc346) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc346) + } -> tensor<1x240x1x1xbf16> loc(#loc346) + %338 = xten_nn.subgraph (%arg5 = %337: tensor<1x240x1x1xbf16>, %arg6 = %52: tensor<960x240x1x1xbf16>, %arg7 = %51: tensor<960xbf16>) attributes { + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "712", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "712", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc194) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc194) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_288", + PartOfOutputName = "Conv_288", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc194) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + %339 = xten_nn.subgraph (%arg5 = %338: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_290", OutputName = "Add_290"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + %340 = xten_nn.subgraph (%arg5 = %339: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_293", + OutputName = "Clip_293", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc196) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc196) + } -> tensor<1x960x1x1xbf16> loc(#loc196) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc196) + } -> tensor<1x960x1x1xbf16> loc(#loc196) + %341 = xten_nn.subgraph (%arg5 = %340: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_295", + OutputName = "Div_295", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc197) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc197) + } -> tensor<1x960x1x1xbf16> loc(#loc197) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc197) + } -> tensor<1x960x1x1xbf16> loc(#loc197) + %342 = xten_nn.subgraph (%arg5 = %341: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Mul_296_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_296_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc198) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc198) + } -> tensor<1x960x12x20xbf16> loc(#loc198) + %343 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>, %arg6 = %334: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_296", + OutputName = "Mul_296", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc198) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc198) + } -> tensor<1x960x12x20xbf16> loc(#loc198) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc198) + } -> tensor<1x960x12x20xbf16> loc(#loc198) + %344 = xten_nn.subgraph (%arg5 = %343: tensor<1x960x12x20xbf16>, %arg6 = %50: tensor<160x960x1x1xbf16>, %arg7 = %49: tensor<160xbf16>, %arg8 = %324: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_297", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "721", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "722", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_297", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "721", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "722", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_297", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1051", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc199) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc199) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc199) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_297", + PartOfOutputName = "Conv_297", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc199) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc199) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc199) + } -> tensor<1x160x12x20xbf16> loc(#loc199) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_298", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1051", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_298", OutputName = "Add_298"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc200) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc200) + } -> tensor<1x160x12x20xbf16> loc(#loc200) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc200) + } -> tensor<1x160x12x20xbf16> loc(#loc347) + %345 = xten_nn.subgraph (%arg5 = %344: tensor<1x160x12x20xbf16>, %arg6 = %48: tensor<960x160x1x1xbf16>, %arg7 = %47: tensor<960xbf16>) attributes { + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1051", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "725", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1051", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "725", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc201) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc201) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc201) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_299", + PartOfOutputName = "Conv_299", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc201) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + %346 = xten_nn.subgraph (%arg5 = %345: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_301", OutputName = "Add_301"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + %347 = xten_nn.subgraph (%arg5 = %346: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_304", + OutputName = "Clip_304", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + %348 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "734", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "734", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_306", + OutputName = "Div_306", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + %349 = xten_nn.subgraph (%arg5 = %345: tensor<1x960x12x20xbf16>, %arg6 = %348: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_307", + OutputName = "Mul_307", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + %350 = xten_nn.subgraph (%arg5 = %349: tensor<1x960x12x20xbf16>, %arg6 = %46: tensor<960x1x9x9xbf16>, %arg7 = %45: tensor<960xbf16>) attributes { + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "735", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "735", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc206) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc206) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc206) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_308", + PartOfOutputName = "Conv_308", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc206) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + %351 = xten_nn.subgraph (%arg5 = %350: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_310", OutputName = "Add_310"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + %352 = xten_nn.subgraph (%arg5 = %351: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_313", + OutputName = "Clip_313", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + %353 = xten_nn.subgraph (%arg5 = %352: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "744", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "744", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_315", + OutputName = "Div_315", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc209) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc209) + } -> tensor<1x960x12x20xbf16> loc(#loc209) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc209) + } -> tensor<1x960x12x20xbf16> loc(#loc209) + %354 = xten_nn.subgraph (%arg5 = %350: tensor<1x960x12x20xbf16>, %arg6 = %353: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_316", + OutputName = "Mul_316", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc210) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc210) + } -> tensor<1x960x12x20xbf16> loc(#loc210) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc210) + } -> tensor<1x960x12x20xbf16> loc(#loc210) + %355 = xten_nn.subgraph (%arg5 = %354: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_317_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc348) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc348) + } -> tensor<1x960x1x240xbf16> loc(#loc348) + %356 = xten_nn.subgraph (%arg5 = %355: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_317", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_317_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_317", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_317_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc211) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc211) + } -> tensor<1x960x1x1xbf16> loc(#loc211) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc211) + } -> tensor<1x960x1x1xbf16> loc(#loc211) + %357 = xten_nn.subgraph (%arg5 = %356: tensor<1x960x1x1xbf16>, %arg6 = %44: tensor<240x960x1x1xbf16>, %arg7 = %43: tensor<240xbf16>) attributes { + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "745", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "745", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc349) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc349) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_318", + PartOfOutputName = "Conv_318", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc212) + %465 = tosa.clamp %464 { + LayerName = "Relu_319", + OutputName = "Relu_319", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc213) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc349) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc213) + } -> tensor<1x240x1x1xbf16> loc(#loc349) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc349) + } -> tensor<1x240x1x1xbf16> loc(#loc349) + %358 = xten_nn.subgraph (%arg5 = %357: tensor<1x240x1x1xbf16>, %arg6 = %42: tensor<960x240x1x1xbf16>, %arg7 = %41: tensor<960xbf16>) attributes { + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "747", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "747", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc214) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc214) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_320", + PartOfOutputName = "Conv_320", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc214) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + %359 = xten_nn.subgraph (%arg5 = %358: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_322", OutputName = "Add_322"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + %360 = xten_nn.subgraph (%arg5 = %359: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_325", + OutputName = "Clip_325", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc216) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc216) + } -> tensor<1x960x1x1xbf16> loc(#loc216) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc216) + } -> tensor<1x960x1x1xbf16> loc(#loc216) + %361 = xten_nn.subgraph (%arg5 = %360: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_327", + OutputName = "Div_327", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc217) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc217) + } -> tensor<1x960x1x1xbf16> loc(#loc217) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc217) + } -> tensor<1x960x1x1xbf16> loc(#loc217) + %362 = xten_nn.subgraph (%arg5 = %361: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Mul_328_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_328_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc218) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc218) + } -> tensor<1x960x12x20xbf16> loc(#loc218) + %363 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>, %arg6 = %354: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_328", + OutputName = "Mul_328", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc218) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc218) + } -> tensor<1x960x12x20xbf16> loc(#loc218) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc218) + } -> tensor<1x960x12x20xbf16> loc(#loc218) + %364 = xten_nn.subgraph (%arg5 = %363: tensor<1x960x12x20xbf16>, %arg6 = %40: tensor<160x960x1x1xbf16>, %arg7 = %39: tensor<160xbf16>, %arg8 = %344: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_329", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "756", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "757", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_329", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "756", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "757", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_329", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1060", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc219) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc219) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc219) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_329", + PartOfOutputName = "Conv_329", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc219) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc219) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc219) + } -> tensor<1x160x12x20xbf16> loc(#loc219) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_330", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1060", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_330", OutputName = "Add_330"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc220) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc220) + } -> tensor<1x160x12x20xbf16> loc(#loc220) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc220) + } -> tensor<1x160x12x20xbf16> loc(#loc350) + %365 = xten_nn.subgraph (%arg5 = %364: tensor<1x160x12x20xbf16>, %arg6 = %38: tensor<960x160x1x1xbf16>, %arg7 = %37: tensor<960xbf16>) attributes { + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1060", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "760", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1060", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "760", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc221) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc221) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc221) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_331", + PartOfOutputName = "Conv_331", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc221) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + %366 = xten_nn.subgraph (%arg5 = %365: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_333", OutputName = "Add_333"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + %367 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_336", + OutputName = "Clip_336", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + %368 = xten_nn.subgraph (%arg5 = %367: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "769", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "769", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_338", + OutputName = "Div_338", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc224) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc224) + } -> tensor<1x960x12x20xbf16> loc(#loc224) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc224) + } -> tensor<1x960x12x20xbf16> loc(#loc224) + %369 = xten_nn.subgraph (%arg5 = %365: tensor<1x960x12x20xbf16>, %arg6 = %368: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_339", + OutputName = "Mul_339", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc225) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc225) + } -> tensor<1x960x12x20xbf16> loc(#loc225) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc225) + } -> tensor<1x960x12x20xbf16> loc(#loc225) + %370 = xten_nn.subgraph (%arg5 = %369: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_342_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc351) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc351) + } -> tensor<1x960x1x240xbf16> loc(#loc351) + %371 = xten_nn.subgraph (%arg5 = %370: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_342", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_342_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_342", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_342_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc226) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc226) + } -> tensor<1x960x1x1xbf16> loc(#loc226) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc226) + } -> tensor<1x960x1x1xbf16> loc(#loc226) + %372 = xten_nn.subgraph (%arg5 = %371: tensor<1x960x1x1xbf16>, %arg6 = %36: tensor<128x960x1x1xbf16>, %arg7 = %35: tensor<128xbf16>) attributes { + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "aspp.aspp2.1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<128x960x1x1xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "aspp.aspp2.1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc227) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc227) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_343", + PartOfOutputName = "Conv_343", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x1x1x128xbf16> loc(#loc227) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x128xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc227) + xten_nn.output %465 : tensor<1x128x1x1xbf16> loc(#loc227) + } -> tensor<1x128x1x1xbf16> loc(#loc227) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc227) + } -> tensor<1x128x1x1xbf16> loc(#loc227) + %373 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "776", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "776", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_344", OutputName = "Sigmoid_344"} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc228) + xten_nn.output %462 : tensor<1x128x1x1xbf16> loc(#loc228) + } -> tensor<1x128x1x1xbf16> loc(#loc228) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc228) + } -> tensor<1x128x1x1xbf16> loc(#loc228) + %374 = xten_nn.subgraph (%arg5 = %373: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Mul_345_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "773", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_345_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 128 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc229) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc229) + %375 = xten_nn.subgraph (%arg5 = %369: tensor<1x960x12x20xbf16>, %arg6 = %34: tensor<128x960x1x1xbf16>, %arg7 = %33: tensor<128xbf16>, %arg8 = %374: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Conv_340", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "1063", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1066", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<128x960x1x1xbf16>, %arg11 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_340", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "1063", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_341", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "773", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc353) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc353) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc353) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_340", + PartOfOutputName = "Conv_340", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc230) + %468 = tosa.clamp %467 { + LayerName = "Relu_341", + OutputName = "Relu_341", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x12x20x128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc231) + %469 = tosa.transpose %468, %463 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc353) + xten_nn.output %469 : tensor<1x128x12x20xbf16> loc(#loc231) + } -> tensor<1x128x12x20xbf16> loc(#loc353) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x128x12x20xbf16>, %arg10 = %arg8: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Mul_345_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "773", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1066", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.mul %arg9, %arg10 { + LayerName = "Mul_345", + OutputName = "Mul_345", + shift = 0 : i8} : (tensor<1x128x12x20xbf16>, tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc229) + xten_nn.output %463 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc229) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc352) + %376 = xten_nn.subgraph (%arg5 = %375: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "781", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc232) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc232) + } -> tensor<1x64x12x20xbf16> loc(#loc232) + %377 = xten_nn.subgraph (%arg5 = %375: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "781", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc232) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc232) + } -> tensor<1x64x12x20xbf16> loc(#loc232) + %378 = xten_nn.subgraph (%arg5 = %377: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_350", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_350", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "783", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_350", + OutputName = "Concat_350", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + %379 = xten_nn.subgraph (%arg5 = %378: tensor<1x128x12x20xbf16>, %arg6 = %32: tensor<128x128x3x3xbf16>, %arg7 = %31: tensor<128xbf16>) attributes { + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "783", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "397", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<128x128x3x3xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "783", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "397", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<128x128x3x3xbf16>, tensor<4xi32>) -> tensor<128x3x3x128xbf16> loc(#loc234) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc234) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_351", + PartOfOutputName = "Conv_351", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<128x3x3x128xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc234) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc234) + xten_nn.output %467 : tensor<1x128x12x20xbf16> loc(#loc234) + } -> tensor<1x128x12x20xbf16> loc(#loc234) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc234) + } -> tensor<1x128x12x20xbf16> loc(#loc234) + %380 = xten_nn.subgraph (%arg5 = %379: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_352", OutputName = "Sigmoid_352"} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc235) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc235) + } -> tensor<1x128x12x20xbf16> loc(#loc235) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc235) + } -> tensor<1x128x12x20xbf16> loc(#loc235) + %381 = xten_nn.subgraph (%arg5 = %380: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc236) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc236) + } -> tensor<1x64x12x20xbf16> loc(#loc236) + %382 = xten_nn.subgraph (%arg5 = %380: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc236) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc236) + } -> tensor<1x64x12x20xbf16> loc(#loc236) + %383 = xten_nn.subgraph (%arg5 = %30: tensor<1x64x12x20xbf16>, %arg6 = %382: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_359", OutputName = "Sub_359"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + %384 = xten_nn.subgraph (%arg5 = %383: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_360", + OutputName = "Mul_360", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + %385 = xten_nn.subgraph (%arg5 = %381: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "788", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "788", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_354", + OutputName = "Mul_354", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + %386 = xten_nn.subgraph (%arg5 = %377: tensor<1x64x12x20xbf16>, %arg6 = %385: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_355", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "782", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "788", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_355", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "789", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_355", + OutputName = "Concat_355", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc239) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc239) + } -> tensor<1x128x12x20xbf16> loc(#loc239) + %387 = xten_nn.subgraph (%arg5 = %386: tensor<1x128x12x20xbf16>, %arg6 = %29: tensor<64x128x3x3xbf16>, %arg7 = %28: tensor<64xbf16>) attributes { + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "789", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "788", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<64x128x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "789", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "788", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<64x128x3x3xbf16>, tensor<4xi32>) -> tensor<64x3x3x128xbf16> loc(#loc240) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc240) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_356", + PartOfOutputName = "Conv_356", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<64x3x3x128xbf16>, tensor<64xbf16>) -> tensor<1x12x20x64xbf16> loc(#loc240) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x64xbf16>, tensor<4xi32>) -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %467 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + %388 = xten_nn.subgraph (%arg5 = %387: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_357", OutputName = "Tanh_357"} : (tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + %389 = xten_nn.subgraph (%arg5 = %382: tensor<1x64x12x20xbf16>, %arg6 = %388: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "795", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "795", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_361", + OutputName = "Mul_361", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc242) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc242) + } -> tensor<1x64x12x20xbf16> loc(#loc242) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc242) + } -> tensor<1x64x12x20xbf16> loc(#loc242) + %390 = xten_nn.subgraph (%arg5 = %384: tensor<1x64x12x20xbf16>, %arg6 = %389: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "794", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "796", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "794", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "796", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_362", OutputName = "Add_362"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc243) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc243) + } -> tensor<1x64x12x20xbf16> loc(#loc243) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc243) + } -> tensor<1x64x12x20xbf16> loc(#loc243) + %391 = xten_nn.subgraph (%arg5 = %376: tensor<1x64x12x20xbf16>, %arg6 = %390: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_363", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "781", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_363", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "797", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_363", + OutputName = "Concat_363", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc244) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc244) + } -> tensor<1x128x12x20xbf16> loc(#loc244) + %392 = xten_nn.subgraph (%arg5 = %391: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Resize_365", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "797", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Resize_365", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "802", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 128 : ui32, + config.dim_2 = 12 : ui32, + config.dim_3 = 20 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 24 : ui32, + config.output_W = 40 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_365", + OutputName = "Resize_365", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x24x40xbf16> loc(#loc245) + xten_nn.output %461 : tensor<1x128x24x40xbf16> loc(#loc245) + } -> tensor<1x128x24x40xbf16> loc(#loc245) + %393 = xten_nn.subgraph (%arg5 = %392: tensor<1x128x24x40xbf16>) attributes { + LayerName = "Slice_371", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "802", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + OutputName = "Slice_371", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "812", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 128 : ui32, + config.dim_h = 24 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 23 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_371", + OutputName = "Slice_371", + size = array, + start = array} : (tensor<1x128x24x40xbf16>) -> tensor<1x128x23x40xbf16> loc(#loc246) + xten_nn.output %461 : tensor<1x128x23x40xbf16> loc(#loc246) + } -> tensor<1x128x23x40xbf16> loc(#loc246) + %394 = xten_nn.subgraph (%arg5 = %393: tensor<1x128x23x40xbf16>, %arg6 = %220: tensor<1x40x23x40xbf16>, %arg7 = %169: tensor<1x3x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_372", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "812", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "802", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_372", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "813", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_372", + OutputName = "Concat_372", + axis = 1 : i32} : (tensor<1x128x23x40xbf16>, tensor<1x40x23x40xbf16>, tensor<1x3x23x40xbf16>) -> tensor<1x171x23x40xbf16> loc(#loc247) + xten_nn.output %461 : tensor<1x171x23x40xbf16> loc(#loc247) + } -> tensor<1x171x23x40xbf16> loc(#loc247) + %395 = xten_nn.subgraph (%arg5 = %394: tensor<1x171x23x40xbf16>, %arg6 = %27: tensor<80x171x3x3xbf16>, %arg7 = %26: tensor<80xbf16>) attributes { + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "813", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Name = "812", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Name = "813", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x171x23x40xbf16>, %arg9 = %arg6: tensor<80x171x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "813", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Name = "812", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Name = "813", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x171x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x171xbf16> loc(#loc354) + %465 = tosa.transpose %arg8, %463 : (tensor<1x171x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x171xbf16> loc(#loc354) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_373", + PartOfOutputName = "Conv_373", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x171xbf16>, tensor<80x3x3x171xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc248) + %467 = tosa.clamp %466 { + LayerName = "Relu_374", + OutputName = "Relu_374", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc249) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc354) + xten_nn.output %468 : tensor<1x80x23x40xbf16> loc(#loc249) + } -> tensor<1x80x23x40xbf16> loc(#loc354) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc354) + } -> tensor<1x80x23x40xbf16> loc(#loc354) + %396 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "817", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %397 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "817", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %398 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_376", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_376", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "819", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_376", + OutputName = "Concat_376", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc251) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc251) + } -> tensor<1x80x23x40xbf16> loc(#loc251) + %399 = xten_nn.subgraph (%arg5 = %398: tensor<1x80x23x40xbf16>, %arg6 = %25: tensor<80x80x3x3xbf16>, %arg7 = %24: tensor<80xbf16>) attributes { + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "819", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "396", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<80x80x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "819", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "396", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x80x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x80xbf16> loc(#loc252) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_377", + PartOfOutputName = "Conv_377", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<80x3x3x80xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %467 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + %400 = xten_nn.subgraph (%arg5 = %399: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_378", OutputName = "Sigmoid_378"} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %462 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + %401 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %402 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %403 = xten_nn.subgraph (%arg5 = %23: tensor<1x40x23x40xbf16>, %arg6 = %402: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_385", OutputName = "Sub_385"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + %404 = xten_nn.subgraph (%arg5 = %403: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_386", + OutputName = "Mul_386", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + %405 = xten_nn.subgraph (%arg5 = %401: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "824", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "824", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_380", + OutputName = "Mul_380", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc256) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc256) + } -> tensor<1x40x23x40xbf16> loc(#loc256) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc256) + } -> tensor<1x40x23x40xbf16> loc(#loc256) + %406 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %405: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_381", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "818", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "824", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_381", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "825", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_381", + OutputName = "Concat_381", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc257) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc257) + } -> tensor<1x80x23x40xbf16> loc(#loc257) + %407 = xten_nn.subgraph (%arg5 = %406: tensor<1x80x23x40xbf16>, %arg6 = %22: tensor<40x80x3x3xbf16>, %arg7 = %21: tensor<40xbf16>) attributes { + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "825", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "824", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<40x80x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "825", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "824", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x80x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x80xbf16> loc(#loc258) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc258) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_382", + PartOfOutputName = "Conv_382", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<40x3x3x80xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc258) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + %408 = xten_nn.subgraph (%arg5 = %407: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_383", OutputName = "Tanh_383"} : (tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + %409 = xten_nn.subgraph (%arg5 = %402: tensor<1x40x23x40xbf16>, %arg6 = %408: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "831", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "831", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_387", + OutputName = "Mul_387", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + %410 = xten_nn.subgraph (%arg5 = %404: tensor<1x40x23x40xbf16>, %arg6 = %409: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "830", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "832", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "830", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "832", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_388", OutputName = "Add_388"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + %411 = xten_nn.subgraph (%arg5 = %396: tensor<1x40x23x40xbf16>, %arg6 = %410: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_389", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "817", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_389", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "833", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_389", + OutputName = "Concat_389", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc262) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc262) + } -> tensor<1x80x23x40xbf16> loc(#loc262) + %412 = xten_nn.subgraph (%arg5 = %411: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Resize_391", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "833", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Resize_391", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "838", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 80 : ui32, + config.dim_2 = 23 : ui32, + config.dim_3 = 40 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 46 : ui32, + config.output_W = 80 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_391", + OutputName = "Resize_391", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x46x80xbf16> loc(#loc263) + xten_nn.output %461 : tensor<1x80x46x80xbf16> loc(#loc263) + } -> tensor<1x80x46x80xbf16> loc(#loc263) + %413 = xten_nn.subgraph (%arg5 = %412: tensor<1x80x46x80xbf16>) attributes { + LayerName = "Slice_397", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "838", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + OutputName = "Slice_397", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "848", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 80 : ui32, + config.dim_h = 46 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 45 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_397", + OutputName = "Slice_397", + size = array, + start = array} : (tensor<1x80x46x80xbf16>) -> tensor<1x80x45x80xbf16> loc(#loc264) + xten_nn.output %461 : tensor<1x80x45x80xbf16> loc(#loc264) + } -> tensor<1x80x45x80xbf16> loc(#loc264) + %414 = xten_nn.subgraph (%arg5 = %413: tensor<1x80x45x80xbf16>, %arg6 = %184: tensor<1x24x45x80xbf16>, %arg7 = %168: tensor<1x3x45x80xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_398", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "848", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "838", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "849", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_398", + OutputName = "Concat_398", + axis = 1 : i32} : (tensor<1x80x45x80xbf16>, tensor<1x24x45x80xbf16>, tensor<1x3x45x80xbf16>) -> tensor<1x107x45x80xbf16> loc(#loc265) + xten_nn.output %461 : tensor<1x107x45x80xbf16> loc(#loc265) + } -> tensor<1x107x45x80xbf16> loc(#loc265) + %415 = xten_nn.subgraph (%arg5 = %414: tensor<1x107x45x80xbf16>, %arg6 = %20: tensor<40x107x3x3xbf16>, %arg7 = %19: tensor<40xbf16>) attributes { + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "849", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Name = "848", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Name = "849", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x107x45x80xbf16>, %arg9 = %arg6: tensor<40x107x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "849", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Name = "848", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Name = "849", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x107x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x107xbf16> loc(#loc355) + %465 = tosa.transpose %arg8, %463 : (tensor<1x107x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x107xbf16> loc(#loc355) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_399", + PartOfOutputName = "Conv_399", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x107xbf16>, tensor<40x3x3x107xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc266) + %467 = tosa.clamp %466 { + LayerName = "Relu_400", + OutputName = "Relu_400", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc267) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc355) + xten_nn.output %468 : tensor<1x40x45x80xbf16> loc(#loc267) + } -> tensor<1x40x45x80xbf16> loc(#loc355) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc355) + } -> tensor<1x40x45x80xbf16> loc(#loc355) + %416 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "853", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %417 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "853", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %418 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_402", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_402", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "855", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_402", + OutputName = "Concat_402", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc269) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc269) + } -> tensor<1x40x45x80xbf16> loc(#loc269) + %419 = xten_nn.subgraph (%arg5 = %418: tensor<1x40x45x80xbf16>, %arg6 = %18: tensor<40x40x3x3xbf16>, %arg7 = %17: tensor<40xbf16>) attributes { + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "855", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "395", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<40x40x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "855", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "395", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x40x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x40xbf16> loc(#loc270) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_403", + PartOfOutputName = "Conv_403", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<40x3x3x40xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %467 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + %420 = xten_nn.subgraph (%arg5 = %419: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_404", OutputName = "Sigmoid_404"} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %462 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + %421 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %422 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %423 = xten_nn.subgraph (%arg5 = %16: tensor<1x20x45x80xbf16>, %arg6 = %422: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_411", OutputName = "Sub_411"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + %424 = xten_nn.subgraph (%arg5 = %423: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_412", + OutputName = "Mul_412", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + %425 = xten_nn.subgraph (%arg5 = %421: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "860", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "860", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_406", + OutputName = "Mul_406", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc274) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc274) + } -> tensor<1x20x45x80xbf16> loc(#loc274) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc274) + } -> tensor<1x20x45x80xbf16> loc(#loc274) + %426 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %425: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_407", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "854", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "860", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_407", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "861", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_407", + OutputName = "Concat_407", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc275) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc275) + } -> tensor<1x40x45x80xbf16> loc(#loc275) + %427 = xten_nn.subgraph (%arg5 = %426: tensor<1x40x45x80xbf16>, %arg6 = %15: tensor<20x40x3x3xbf16>, %arg7 = %14: tensor<20xbf16>) attributes { + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "861", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "860", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<20x40x3x3xbf16>, %arg10 = %arg7: tensor<20xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "861", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "860", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<20x40x3x3xbf16>, tensor<4xi32>) -> tensor<20x3x3x40xbf16> loc(#loc276) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc276) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_408", + PartOfOutputName = "Conv_408", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<20x3x3x40xbf16>, tensor<20xbf16>) -> tensor<1x45x80x20xbf16> loc(#loc276) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x20xbf16>, tensor<4xi32>) -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %467 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + %428 = xten_nn.subgraph (%arg5 = %427: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_409", OutputName = "Tanh_409"} : (tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + %429 = xten_nn.subgraph (%arg5 = %422: tensor<1x20x45x80xbf16>, %arg6 = %428: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "867", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "867", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_413", + OutputName = "Mul_413", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + %430 = xten_nn.subgraph (%arg5 = %424: tensor<1x20x45x80xbf16>, %arg6 = %429: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "866", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "868", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "866", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "868", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_414", OutputName = "Add_414"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + %431 = xten_nn.subgraph (%arg5 = %416: tensor<1x20x45x80xbf16>, %arg6 = %430: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_415", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "853", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_415", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "869", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_415", + OutputName = "Concat_415", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc280) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc280) + } -> tensor<1x40x45x80xbf16> loc(#loc280) + %432 = xten_nn.subgraph (%arg5 = %431: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Resize_417", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "869", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Resize_417", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "874", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 40 : ui32, + config.dim_2 = 45 : ui32, + config.dim_3 = 80 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 90 : ui32, + config.output_W = 160 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_417", + OutputName = "Resize_417", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x90x160xbf16> loc(#loc281) + xten_nn.output %461 : tensor<1x40x90x160xbf16> loc(#loc281) + } -> tensor<1x40x90x160xbf16> loc(#loc281) + %433 = xten_nn.subgraph (%arg5 = %432: tensor<1x40x90x160xbf16>, %arg6 = %178: tensor<1x16x90x160xbf16>, %arg7 = %167: tensor<1x3x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_418", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "874", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "869", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_418", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "875", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_418", + OutputName = "Concat_418", + axis = 1 : i32} : (tensor<1x40x90x160xbf16>, tensor<1x16x90x160xbf16>, tensor<1x3x90x160xbf16>) -> tensor<1x59x90x160xbf16> loc(#loc282) + xten_nn.output %461 : tensor<1x59x90x160xbf16> loc(#loc282) + } -> tensor<1x59x90x160xbf16> loc(#loc282) + %434 = xten_nn.subgraph (%arg5 = %433: tensor<1x59x90x160xbf16>, %arg6 = %13: tensor<32x59x3x3xbf16>, %arg7 = %12: tensor<32xbf16>) attributes { + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "875", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Name = "874", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Name = "875", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x59x90x160xbf16>, %arg9 = %arg6: tensor<32x59x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "875", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Name = "874", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Name = "875", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x59x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x59xbf16> loc(#loc356) + %465 = tosa.transpose %arg8, %463 : (tensor<1x59x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x59xbf16> loc(#loc356) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_419", + PartOfOutputName = "Conv_419", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x59xbf16>, tensor<32x3x3x59xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc283) + %467 = tosa.clamp %466 { + LayerName = "Relu_420", + OutputName = "Relu_420", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc284) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc356) + xten_nn.output %468 : tensor<1x32x90x160xbf16> loc(#loc284) + } -> tensor<1x32x90x160xbf16> loc(#loc356) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc356) + } -> tensor<1x32x90x160xbf16> loc(#loc356) + %435 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "879", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %436 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "879", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %437 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_422", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_422", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "881", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_422", + OutputName = "Concat_422", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc286) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc286) + } -> tensor<1x32x90x160xbf16> loc(#loc286) + %438 = xten_nn.subgraph (%arg5 = %437: tensor<1x32x90x160xbf16>, %arg6 = %11: tensor<32x32x3x3xbf16>, %arg7 = %10: tensor<32xbf16>) attributes { + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "881", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "394", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<32x32x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "881", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "394", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x32x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x32xbf16> loc(#loc287) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_423", + PartOfOutputName = "Conv_423", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<32x3x3x32xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %467 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + %439 = xten_nn.subgraph (%arg5 = %438: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_424", OutputName = "Sigmoid_424"} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %462 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + %440 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %441 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %442 = xten_nn.subgraph (%arg5 = %9: tensor<1x16x90x160xbf16>, %arg6 = %441: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_431", OutputName = "Sub_431"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + %443 = xten_nn.subgraph (%arg5 = %442: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_432", + OutputName = "Mul_432", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + %444 = xten_nn.subgraph (%arg5 = %440: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "886", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "886", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_426", + OutputName = "Mul_426", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc291) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc291) + } -> tensor<1x16x90x160xbf16> loc(#loc291) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc291) + } -> tensor<1x16x90x160xbf16> loc(#loc291) + %445 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %444: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_427", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "880", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "886", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_427", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "887", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_427", + OutputName = "Concat_427", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc292) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc292) + } -> tensor<1x32x90x160xbf16> loc(#loc292) + %446 = xten_nn.subgraph (%arg5 = %445: tensor<1x32x90x160xbf16>, %arg6 = %8: tensor<16x32x3x3xbf16>, %arg7 = %7: tensor<16xbf16>) attributes { + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "887", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "886", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<16x32x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "887", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "886", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x32x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x32xbf16> loc(#loc293) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc293) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_428", + PartOfOutputName = "Conv_428", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<16x3x3x32xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc293) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + %447 = xten_nn.subgraph (%arg5 = %446: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_429", OutputName = "Tanh_429"} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + %448 = xten_nn.subgraph (%arg5 = %441: tensor<1x16x90x160xbf16>, %arg6 = %447: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "893", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "893", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_433", + OutputName = "Mul_433", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + %449 = xten_nn.subgraph (%arg5 = %443: tensor<1x16x90x160xbf16>, %arg6 = %448: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "892", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "894", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "892", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "894", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_434", OutputName = "Add_434"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + %450 = xten_nn.subgraph (%arg5 = %435: tensor<1x16x90x160xbf16>, %arg6 = %449: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_435", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "879", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_435", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "895", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_435", + OutputName = "Concat_435", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc297) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc297) + } -> tensor<1x32x90x160xbf16> loc(#loc297) + %451 = xten_nn.subgraph (%arg5 = %450: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Resize_437", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "895", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Resize_437", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "900", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 32 : ui32, + config.dim_2 = 90 : ui32, + config.dim_3 = 160 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 180 : ui32, + config.output_W = 320 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_437", + OutputName = "Resize_437", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x180x320xbf16> loc(#loc298) + xten_nn.output %461 : tensor<1x32x180x320xbf16> loc(#loc298) + } -> tensor<1x32x180x320xbf16> loc(#loc298) + %452 = xten_nn.subgraph (%arg5 = %451: tensor<1x32x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_438", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "900", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "895", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Concat_438", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "901", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_438", + OutputName = "Concat_438", + axis = 1 : i32} : (tensor<1x32x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x35x180x320xbf16> loc(#loc299) + xten_nn.output %461 : tensor<1x35x180x320xbf16> loc(#loc299) + } -> tensor<1x35x180x320xbf16> loc(#loc299) + %453 = xten_nn.subgraph (%arg5 = %452: tensor<1x35x180x320xbf16>, %arg6 = %6: tensor<16x35x3x3xbf16>, %arg7 = %5: tensor<16xbf16>) attributes { + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "901", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Name = "900", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Name = "901", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x35x180x320xbf16>, %arg9 = %arg6: tensor<16x35x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "901", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Name = "900", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Name = "901", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x35x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x35xbf16> loc(#loc357) + %465 = tosa.transpose %arg8, %463 : (tensor<1x35x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x35xbf16> loc(#loc357) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_439", + PartOfOutputName = "Conv_439", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x35xbf16>, tensor<16x3x3x35xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc300) + %467 = tosa.clamp %466 { + LayerName = "Relu_440", + OutputName = "Relu_440", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc301) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc357) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc301) + } -> tensor<1x16x180x320xbf16> loc(#loc357) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc357) + } -> tensor<1x16x180x320xbf16> loc(#loc357) + %454 = xten_nn.subgraph (%arg5 = %453: tensor<1x16x180x320xbf16>, %arg6 = %4: tensor<16x16x3x3xbf16>, %arg7 = %3: tensor<16xbf16>) attributes { + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1078", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Name = "1082", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<16x16x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1078", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Name = "1082", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x16x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x16xbf16> loc(#loc358) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc358) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_441", + PartOfOutputName = "Conv_441", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<16x3x3x16xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc302) + %467 = tosa.clamp %466 { + LayerName = "Relu_442", + OutputName = "Relu_442", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc303) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc358) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc303) + } -> tensor<1x16x180x320xbf16> loc(#loc358) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc358) + } -> tensor<1x16x180x320xbf16> loc(#loc358) + %455 = xten_nn.subgraph (%arg5 = %454: tensor<1x16x180x320xbf16>, %arg6 = %2: tensor<4x16x1x1xbf16>, %arg7 = %1: tensor<4xbf16>) attributes { + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1081", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "project_mat.conv.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<4x16x1x1xbf16>, %arg10 = %arg7: tensor<4xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1081", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "project_mat.conv.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc304) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<4x16x1x1xbf16>) -> tensor<4x1x1x16xbf16> loc(#loc304) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc304) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_443", + PartOfOutputName = "Conv_443", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<4x1x1x16xbf16>, tensor<4xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc304) + %467 = tosa.transpose %466, %462 : (tensor<1x180x320x4xbf16>, tensor<4xi32>) -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %467 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %461 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + %456 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc305) + } -> tensor<1x3x180x320xbf16> loc(#loc305) + %457 = xten_nn.subgraph (%arg5 = %456: tensor<1x3x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_445", OutputName = "Add_445"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + %458 = xten_nn.subgraph (%arg5 = %457: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "916", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "916", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_446", + OutputName = "Clip_446", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + %459 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 4 : ui32, + config.start = 3 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc305) + } -> tensor<1x1x180x320xbf16> loc(#loc305) + %460 = xten_nn.subgraph (%arg5 = %459: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "910", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "921", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "910", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "921", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_447", + OutputName = "Clip_447", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %462 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + return %449, %430, %410, %390, %458, %460 : tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16> loc(#loc313) + } loc(#loc313) +} loc(#loc) +#loc = loc(unknown) +#loc1 = loc("Div_2") +#loc2 = loc("Sub_431") +#loc3 = loc("Sub_411") +#loc4 = loc("Sub_385") +#loc5 = loc("Sub_359") +#loc6 = loc("Div_16") +#loc7 = loc("Sub_14") +#loc8 = loc("Initializer_398") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("AveragePool_347") +#loc14 = loc("AveragePool_348") +#loc15 = loc("Conv_17") +#loc16 = loc("Add_19") +#loc17 = loc("Clip_22") +#loc18 = loc("Div_24") +#loc19 = loc("Mul_25") +#loc20 = loc("Conv_26") +#loc21 = loc("Relu_27") +#loc22 = loc("Conv_28") +#loc23 = loc("Add_29") +#loc24 = loc("Conv_30") +#loc25 = loc("Relu_31") +#loc26 = loc("Conv_32") +#loc27 = loc("Relu_33") +#loc28 = loc("Conv_34") +#loc29 = loc("Conv_35") +#loc30 = loc("Relu_36") +#loc31 = loc("Conv_37") +#loc32 = loc("Relu_38") +#loc33 = loc("Conv_39") +#loc34 = loc("Add_40") +#loc35 = loc("Conv_41") +#loc36 = loc("Relu_42") +#loc37 = loc("Conv_43") +#loc38 = loc("Relu_44") +#loc39 = loc("GlobalAveragePool_45") +#loc40 = loc("Conv_46") +#loc41 = loc("Relu_47") +#loc42 = loc("Conv_48") +#loc43 = loc("Add_50") +#loc44 = loc("Clip_53") +#loc45 = loc("Div_55") +#loc46 = loc("Mul_56") +#loc47 = loc("Conv_57") +#loc48 = loc("Conv_58") +#loc49 = loc("Relu_59") +#loc50 = loc("Conv_60") +#loc51 = loc("Relu_61") +#loc52 = loc("GlobalAveragePool_62") +#loc53 = loc("Conv_63") +#loc54 = loc("Relu_64") +#loc55 = loc("Conv_65") +#loc56 = loc("Add_67") +#loc57 = loc("Clip_70") +#loc58 = loc("Div_72") +#loc59 = loc("Mul_73") +#loc60 = loc("Conv_74") +#loc61 = loc("Add_75") +#loc62 = loc("Conv_76") +#loc63 = loc("Relu_77") +#loc64 = loc("Conv_78") +#loc65 = loc("Relu_79") +#loc66 = loc("GlobalAveragePool_80") +#loc67 = loc("Conv_81") +#loc68 = loc("Relu_82") +#loc69 = loc("Conv_83") +#loc70 = loc("Add_85") +#loc71 = loc("Clip_88") +#loc72 = loc("Div_90") +#loc73 = loc("Mul_91") +#loc74 = loc("Conv_92") +#loc75 = loc("Add_93") +#loc76 = loc("Conv_94") +#loc77 = loc("Add_96") +#loc78 = loc("Clip_99") +#loc79 = loc("Div_101") +#loc80 = loc("Mul_102") +#loc81 = loc("Conv_103") +#loc82 = loc("Add_105") +#loc83 = loc("Clip_108") +#loc84 = loc("Div_110") +#loc85 = loc("Mul_111") +#loc86 = loc("Conv_112") +#loc87 = loc("Conv_113") +#loc88 = loc("Add_115") +#loc89 = loc("Clip_118") +#loc90 = loc("Div_120") +#loc91 = loc("Mul_121") +#loc92 = loc("Conv_122") +#loc93 = loc("Add_124") +#loc94 = loc("Clip_127") +#loc95 = loc("Div_129") +#loc96 = loc("Mul_130") +#loc97 = loc("Conv_131") +#loc98 = loc("Add_132") +#loc99 = loc("Conv_133") +#loc100 = loc("Add_135") +#loc101 = loc("Clip_138") +#loc102 = loc("Div_140") +#loc103 = loc("Mul_141") +#loc104 = loc("Conv_142") +#loc105 = loc("Add_144") +#loc106 = loc("Clip_147") +#loc107 = loc("Div_149") +#loc108 = loc("Mul_150") +#loc109 = loc("Conv_151") +#loc110 = loc("Add_152") +#loc111 = loc("Conv_153") +#loc112 = loc("Add_155") +#loc113 = loc("Clip_158") +#loc114 = loc("Div_160") +#loc115 = loc("Mul_161") +#loc116 = loc("Conv_162") +#loc117 = loc("Add_164") +#loc118 = loc("Clip_167") +#loc119 = loc("Div_169") +#loc120 = loc("Mul_170") +#loc121 = loc("Conv_171") +#loc122 = loc("Add_172") +#loc123 = loc("Conv_173") +#loc124 = loc("Add_175") +#loc125 = loc("Clip_178") +#loc126 = loc("Div_180") +#loc127 = loc("Mul_181") +#loc128 = loc("Conv_182") +#loc129 = loc("Add_184") +#loc130 = loc("Clip_187") +#loc131 = loc("Div_189") +#loc132 = loc("Mul_190") +#loc133 = loc("GlobalAveragePool_191") +#loc134 = loc("Conv_192") +#loc135 = loc("Relu_193") +#loc136 = loc("Conv_194") +#loc137 = loc("Add_196") +#loc138 = loc("Clip_199") +#loc139 = loc("Div_201") +#loc140 = loc("Mul_202") +#loc141 = loc("Conv_203") +#loc142 = loc("Conv_204") +#loc143 = loc("Add_206") +#loc144 = loc("Clip_209") +#loc145 = loc("Div_211") +#loc146 = loc("Mul_212") +#loc147 = loc("Conv_213") +#loc148 = loc("Add_215") +#loc149 = loc("Clip_218") +#loc150 = loc("Div_220") +#loc151 = loc("Mul_221") +#loc152 = loc("GlobalAveragePool_222") +#loc153 = loc("Conv_223") +#loc154 = loc("Relu_224") +#loc155 = loc("Conv_225") +#loc156 = loc("Add_227") +#loc157 = loc("Clip_230") +#loc158 = loc("Div_232") +#loc159 = loc("Mul_233") +#loc160 = loc("Conv_234") +#loc161 = loc("Add_235") +#loc162 = loc("Conv_236") +#loc163 = loc("Add_238") +#loc164 = loc("Clip_241") +#loc165 = loc("Div_243") +#loc166 = loc("Mul_244") +#loc167 = loc("Conv_245") +#loc168 = loc("Add_247") +#loc169 = loc("Clip_250") +#loc170 = loc("Div_252") +#loc171 = loc("Mul_253") +#loc172 = loc("GlobalAveragePool_254") +#loc173 = loc("Conv_255") +#loc174 = loc("Relu_256") +#loc175 = loc("Conv_257") +#loc176 = loc("Add_259") +#loc177 = loc("Clip_262") +#loc178 = loc("Div_264") +#loc179 = loc("Mul_265") +#loc180 = loc("Conv_266") +#loc181 = loc("Conv_267") +#loc182 = loc("Add_269") +#loc183 = loc("Clip_272") +#loc184 = loc("Div_274") +#loc185 = loc("Mul_275") +#loc186 = loc("Conv_276") +#loc187 = loc("Add_278") +#loc188 = loc("Clip_281") +#loc189 = loc("Div_283") +#loc190 = loc("Mul_284") +#loc191 = loc("GlobalAveragePool_285") +#loc192 = loc("Conv_286") +#loc193 = loc("Relu_287") +#loc194 = loc("Conv_288") +#loc195 = loc("Add_290") +#loc196 = loc("Clip_293") +#loc197 = loc("Div_295") +#loc198 = loc("Mul_296") +#loc199 = loc("Conv_297") +#loc200 = loc("Add_298") +#loc201 = loc("Conv_299") +#loc202 = loc("Add_301") +#loc203 = loc("Clip_304") +#loc204 = loc("Div_306") +#loc205 = loc("Mul_307") +#loc206 = loc("Conv_308") +#loc207 = loc("Add_310") +#loc208 = loc("Clip_313") +#loc209 = loc("Div_315") +#loc210 = loc("Mul_316") +#loc211 = loc("GlobalAveragePool_317") +#loc212 = loc("Conv_318") +#loc213 = loc("Relu_319") +#loc214 = loc("Conv_320") +#loc215 = loc("Add_322") +#loc216 = loc("Clip_325") +#loc217 = loc("Div_327") +#loc218 = loc("Mul_328") +#loc219 = loc("Conv_329") +#loc220 = loc("Add_330") +#loc221 = loc("Conv_331") +#loc222 = loc("Add_333") +#loc223 = loc("Clip_336") +#loc224 = loc("Div_338") +#loc225 = loc("Mul_339") +#loc226 = loc("GlobalAveragePool_342") +#loc227 = loc("Conv_343") +#loc228 = loc("Sigmoid_344") +#loc229 = loc("Mul_345") +#loc230 = loc("Conv_340") +#loc231 = loc("Relu_341") +#loc232 = loc("Split_349") +#loc233 = loc("Concat_350") +#loc234 = loc("Conv_351") +#loc235 = loc("Sigmoid_352") +#loc236 = loc("Split_353") +#loc237 = loc("Mul_360") +#loc238 = loc("Mul_354") +#loc239 = loc("Concat_355") +#loc240 = loc("Conv_356") +#loc241 = loc("Tanh_357") +#loc242 = loc("Mul_361") +#loc243 = loc("Add_362") +#loc244 = loc("Concat_363") +#loc245 = loc("Resize_365") +#loc246 = loc("Slice_371") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_386") +#loc256 = loc("Mul_380") +#loc257 = loc("Concat_381") +#loc258 = loc("Conv_382") +#loc259 = loc("Tanh_383") +#loc260 = loc("Mul_387") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_412") +#loc274 = loc("Mul_406") +#loc275 = loc("Concat_407") +#loc276 = loc("Conv_408") +#loc277 = loc("Tanh_409") +#loc278 = loc("Mul_413") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_432") +#loc291 = loc("Mul_426") +#loc292 = loc("Concat_427") +#loc293 = loc("Conv_428") +#loc294 = loc("Tanh_429") +#loc295 = loc("Mul_433") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_446") +#loc307 = loc("Clip_447") +#loc313 = loc(fused[#loc1, #loc2, #loc3, #loc4, #loc5, #loc6, #loc7, #loc8, #loc9, #loc10, #loc11, #loc12, #loc13, #loc14, #loc15, #loc16, #loc17, #loc18, #loc19, #loc20, #loc21, #loc22, #loc23, #loc24, #loc25, #loc26, #loc27, #loc28, #loc29, #loc30, #loc31, #loc32, #loc33, #loc34, #loc35, #loc36, #loc37, #loc38, #loc39, #loc40, #loc41, #loc42, #loc43, #loc44, #loc45, #loc46, #loc47, #loc48, #loc49, #loc50, #loc51, #loc52, #loc53, #loc54, #loc55, #loc56, #loc57, #loc58, #loc59, #loc60, #loc61, #loc62, #loc63, #loc64, #loc65, #loc66, #loc67, #loc68, #loc69, #loc70, #loc71, #loc72, #loc73, #loc74, #loc75, #loc76, #loc77, #loc78, #loc79, #loc80, #loc81, #loc82, #loc83, #loc84, #loc85, #loc86, #loc87, #loc88, #loc89, #loc90, #loc91, #loc92, #loc93, #loc94, #loc95, #loc96, #loc97, #loc98, #loc99, #loc100, #loc101, #loc102, #loc103, #loc104, #loc105, #loc106, #loc107, #loc108, #loc109, #loc110, #loc111, #loc112, #loc113, #loc114, #loc115, #loc116, #loc117, #loc118, #loc119, #loc120, #loc121, #loc122, #loc123, #loc124, #loc125, #loc126, #loc127, #loc128, #loc129, #loc130, #loc131, #loc132, #loc133, #loc134, #loc135, #loc136, #loc137, #loc138, #loc139, #loc140, #loc141, #loc142, #loc143, #loc144, #loc145, #loc146, #loc147, #loc148, #loc149, #loc150, #loc151, #loc152, #loc153, #loc154, #loc155, #loc156, #loc157, #loc158, #loc159, #loc160, #loc161, #loc162, #loc163, #loc164, #loc165, #loc166, #loc167, #loc168, #loc169, #loc170, #loc171, #loc172, #loc173, #loc174, #loc175, #loc176, #loc177, #loc178, #loc179, #loc180, #loc181, #loc182, #loc183, #loc184, #loc185, #loc186, #loc187, #loc188, #loc189, #loc190, #loc191, #loc192, #loc193, #loc194, #loc195, #loc196, #loc197, #loc198, #loc199, #loc200, #loc201, #loc202, #loc203, #loc204, #loc205, #loc206, #loc207, #loc208, #loc209, #loc210, #loc211, #loc212, #loc213, #loc214, #loc215, #loc216, #loc217, #loc218, #loc219, #loc220, #loc221, #loc222, #loc223, #loc224, #loc225, #loc226, #loc227, #loc228, #loc229, #loc230, #loc231, #loc232, #loc233, #loc234, #loc235, #loc236, #loc237, #loc238, #loc239, #loc240, #loc241, #loc242, #loc243, #loc244, #loc245, #loc246, #loc247, #loc248, #loc249, #loc250, #loc251, #loc252, #loc253, #loc254, #loc255, #loc256, #loc257, #loc258, #loc259, #loc260, #loc261, #loc262, #loc263, #loc264, #loc265, #loc266, #loc267, #loc268, #loc269, #loc270, #loc271, #loc272, #loc273, #loc274, #loc275, #loc276, #loc277, #loc278, #loc279, #loc280, #loc281, #loc282, #loc283, #loc284, #loc285, #loc286, #loc287, #loc288, #loc289, #loc290, #loc291, #loc292, #loc293, #loc294, #loc295, #loc296, #loc297, #loc298, #loc299, #loc300, #loc301, #loc302, #loc303, #loc304, #loc305, #loc306, #loc307]) +#loc314 = loc(fused[#loc7, #loc8]) +#loc315 = loc(fused[#loc11, #loc9, #loc12]) +#loc316 = loc(fused[#loc9, #loc12, #loc11]) +#loc317 = loc(fused[#loc20, #loc21]) +#loc318 = loc(fused[#loc22, #loc23]) +#loc319 = loc(fused[#loc24, #loc25]) +#loc320 = loc(fused[#loc26, #loc27]) +#loc321 = loc(fused[#loc29, #loc30]) +#loc322 = loc(fused[#loc31, #loc32]) +#loc323 = loc(fused[#loc33, #loc34]) +#loc324 = loc(fused[#loc35, #loc36]) +#loc325 = loc(fused[#loc37, #loc38]) +#loc326 = loc(fused[#loc40, #loc41]) +#loc327 = loc(fused[#loc48, #loc49]) +#loc328 = loc(fused[#loc50, #loc51]) +#loc329 = loc(fused[#loc53, #loc54]) +#loc330 = loc(fused[#loc60, #loc61]) +#loc331 = loc(fused[#loc62, #loc63]) +#loc332 = loc(fused[#loc64, #loc65]) +#loc333 = loc(fused[#loc67, #loc68]) +#loc334 = loc(fused[#loc74, #loc75]) +#loc335 = loc(fused[#loc97, #loc98]) +#loc336 = loc(fused[#loc109, #loc110]) +#loc337 = loc(fused[#loc121, #loc122]) +#loc338 = loc(fused[#loc132, #loc133]) +#loc339 = loc(fused[#loc134, #loc135]) +#loc340 = loc(fused[#loc151, #loc152]) +#loc341 = loc(fused[#loc153, #loc154]) +#loc342 = loc(fused[#loc160, #loc161]) +#loc343 = loc(fused[#loc171, #loc172]) +#loc344 = loc(fused[#loc173, #loc174]) +#loc345 = loc(fused[#loc190, #loc191]) +#loc346 = loc(fused[#loc192, #loc193]) +#loc347 = loc(fused[#loc199, #loc200]) +#loc348 = loc(fused[#loc210, #loc211]) +#loc349 = loc(fused[#loc212, #loc213]) +#loc350 = loc(fused[#loc219, #loc220]) +#loc351 = loc(fused[#loc225, #loc226]) +#loc352 = loc(fused[#loc230, #loc231, #loc229]) +#loc353 = loc(fused[#loc230, #loc231]) +#loc354 = loc(fused[#loc248, #loc249]) +#loc355 = loc(fused[#loc266, #loc267]) +#loc356 = loc(fused[#loc283, #loc284]) +#loc357 = loc(fused[#loc300, #loc301]) +#loc358 = loc(fused[#loc302, #loc303]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.chained_kernel.tosa.mlir b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.chained_kernel.tosa.mlir new file mode 100644 index 0000000000000000000000000000000000000000..cfdfc2e577455de53419ea6f78bb07721ce1c237 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.chained_kernel.tosa.mlir @@ -0,0 +1,26473 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func @forward(%arg0: tensor<1x180x320x4xui8> {onnx.name = "src"} loc(unknown), %arg1: tensor<1x90x160x16xbf16> {onnx.name = "r1i"} loc(unknown), %arg2: tensor<1x45x80x20xbf16> {onnx.name = "r2i"} loc(unknown), %arg3: tensor<1x23x40x40xbf16> {onnx.name = "r3i"} loc(unknown), %arg4: tensor<1x12x20x64xbf16> {onnx.name = "r4i"} loc(unknown)) -> (tensor<1x180x320x3xbf16> {onnx.name = "fgr"}, tensor<1x180x320x1xbf16> {onnx.name = "pha"}, tensor<1x90x160x16xbf16> {onnx.name = "r1o"}, tensor<1x45x80x20xbf16> {onnx.name = "r2o"}, tensor<1x23x40x40xbf16> {onnx.name = "r3o"}, tensor<1x12x20x64xbf16> {onnx.name = "r4o"}) { + %0 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/biases"} -> tensor<4xbf16> loc(#loc) + %1 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/weights"} -> tensor<4x16x1x1xbf16> loc(#loc) + %2 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/biases"} -> tensor<16xbf16> loc(#loc) + %3 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/weights"} -> tensor<16x16x3x3xbf16> loc(#loc) + %4 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/biases"} -> tensor<16xbf16> loc(#loc) + %5 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/weights"} -> tensor<16x35x3x3xbf16> loc(#loc) + %6 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/biases"} -> tensor<16xbf16> loc(#loc) + %7 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/weights"} -> tensor<16x32x3x3xbf16> loc(#loc) + %8 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_431/Constant_0_0"} -> tensor<1x16x90x160xbf16> loc(#loc1) + %9 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/biases"} -> tensor<32xbf16> loc(#loc) + %10 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/weights"} -> tensor<32x32x3x3xbf16> loc(#loc) + %11 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/biases"} -> tensor<32xbf16> loc(#loc) + %12 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/weights"} -> tensor<32x59x3x3xbf16> loc(#loc) + %13 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/biases"} -> tensor<20xbf16> loc(#loc) + %14 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/weights"} -> tensor<20x40x3x3xbf16> loc(#loc) + %15 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_411/Constant_0_0"} -> tensor<1x20x45x80xbf16> loc(#loc2) + %16 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/biases"} -> tensor<40xbf16> loc(#loc) + %17 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/weights"} -> tensor<40x40x3x3xbf16> loc(#loc) + %18 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/biases"} -> tensor<40xbf16> loc(#loc) + %19 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/weights"} -> tensor<40x107x3x3xbf16> loc(#loc) + %20 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/biases"} -> tensor<40xbf16> loc(#loc) + %21 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/weights"} -> tensor<40x80x3x3xbf16> loc(#loc) + %22 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_385/Constant_0_0"} -> tensor<1x40x23x40xbf16> loc(#loc3) + %23 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/biases"} -> tensor<80xbf16> loc(#loc) + %24 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/weights"} -> tensor<80x80x3x3xbf16> loc(#loc) + %25 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/biases"} -> tensor<80xbf16> loc(#loc) + %26 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/weights"} -> tensor<80x171x3x3xbf16> loc(#loc) + %27 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/biases"} -> tensor<64xbf16> loc(#loc) + %28 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/weights"} -> tensor<64x128x3x3xbf16> loc(#loc) + %29 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_359/Constant_0_0"} -> tensor<1x64x12x20xbf16> loc(#loc4) + %30 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/biases"} -> tensor<128xbf16> loc(#loc) + %31 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/weights"} -> tensor<128x128x3x3xbf16> loc(#loc) + %32 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/biases"} -> tensor<128xbf16> loc(#loc) + %33 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %34 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/biases"} -> tensor<128xbf16> loc(#loc) + %35 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %36 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/biases"} -> tensor<960xbf16> loc(#loc) + %37 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %38 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/biases"} -> tensor<160xbf16> loc(#loc) + %39 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %40 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/biases"} -> tensor<960xbf16> loc(#loc) + %41 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %42 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/biases"} -> tensor<240xbf16> loc(#loc) + %43 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %44 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/biases"} -> tensor<960xbf16> loc(#loc) + %45 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %46 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/biases"} -> tensor<960xbf16> loc(#loc) + %47 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %48 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/biases"} -> tensor<160xbf16> loc(#loc) + %49 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %50 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/biases"} -> tensor<960xbf16> loc(#loc) + %51 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %52 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/biases"} -> tensor<240xbf16> loc(#loc) + %53 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %54 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/biases"} -> tensor<960xbf16> loc(#loc) + %55 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %56 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/biases"} -> tensor<960xbf16> loc(#loc) + %57 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %58 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/biases"} -> tensor<160xbf16> loc(#loc) + %59 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/weights"} -> tensor<160x672x1x1xbf16> loc(#loc) + %60 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/biases"} -> tensor<672xbf16> loc(#loc) + %61 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %62 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/biases"} -> tensor<168xbf16> loc(#loc) + %63 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %64 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/biases"} -> tensor<672xbf16> loc(#loc) + %65 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/weights"} -> tensor<672x1x9x9xbf16> loc(#loc) + %66 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/biases"} -> tensor<672xbf16> loc(#loc) + %67 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %68 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/biases"} -> tensor<112xbf16> loc(#loc) + %69 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/weights"} -> tensor<112x672x1x1xbf16> loc(#loc) + %70 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/biases"} -> tensor<672xbf16> loc(#loc) + %71 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %72 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/biases"} -> tensor<168xbf16> loc(#loc) + %73 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %74 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/biases"} -> tensor<672xbf16> loc(#loc) + %75 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/weights"} -> tensor<672x1x3x3xbf16> loc(#loc) + %76 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/biases"} -> tensor<672xbf16> loc(#loc) + %77 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %78 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/biases"} -> tensor<112xbf16> loc(#loc) + %79 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/weights"} -> tensor<112x480x1x1xbf16> loc(#loc) + %80 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/biases"} -> tensor<480xbf16> loc(#loc) + %81 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/weights"} -> tensor<480x120x1x1xbf16> loc(#loc) + %82 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/biases"} -> tensor<120xbf16> loc(#loc) + %83 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/weights"} -> tensor<120x480x1x1xbf16> loc(#loc) + %84 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/biases"} -> tensor<480xbf16> loc(#loc) + %85 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/weights"} -> tensor<480x1x3x3xbf16> loc(#loc) + %86 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/biases"} -> tensor<480xbf16> loc(#loc) + %87 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/weights"} -> tensor<480x80x1x1xbf16> loc(#loc) + %88 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/biases"} -> tensor<80xbf16> loc(#loc) + %89 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %90 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/biases"} -> tensor<184xbf16> loc(#loc) + %91 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %92 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/biases"} -> tensor<184xbf16> loc(#loc) + %93 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %94 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/biases"} -> tensor<80xbf16> loc(#loc) + %95 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %96 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/biases"} -> tensor<184xbf16> loc(#loc) + %97 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %98 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/biases"} -> tensor<184xbf16> loc(#loc) + %99 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %100 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/biases"} -> tensor<80xbf16> loc(#loc) + %101 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/weights"} -> tensor<80x200x1x1xbf16> loc(#loc) + %102 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/biases"} -> tensor<200xbf16> loc(#loc) + %103 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/weights"} -> tensor<200x1x3x3xbf16> loc(#loc) + %104 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/biases"} -> tensor<200xbf16> loc(#loc) + %105 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/weights"} -> tensor<200x80x1x1xbf16> loc(#loc) + %106 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/biases"} -> tensor<80xbf16> loc(#loc) + %107 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/weights"} -> tensor<80x240x1x1xbf16> loc(#loc) + %108 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/biases"} -> tensor<240xbf16> loc(#loc) + %109 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/weights"} -> tensor<240x1x3x3xbf16> loc(#loc) + %110 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/biases"} -> tensor<240xbf16> loc(#loc) + %111 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/weights"} -> tensor<240x40x1x1xbf16> loc(#loc) + %112 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/biases"} -> tensor<40xbf16> loc(#loc) + %113 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %114 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/biases"} -> tensor<120xbf16> loc(#loc) + %115 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %116 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/biases"} -> tensor<32xbf16> loc(#loc) + %117 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %118 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/biases"} -> tensor<120xbf16> loc(#loc) + %119 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %120 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/biases"} -> tensor<120xbf16> loc(#loc) + %121 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %122 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/biases"} -> tensor<40xbf16> loc(#loc) + %123 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %124 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/biases"} -> tensor<120xbf16> loc(#loc) + %125 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %126 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/biases"} -> tensor<32xbf16> loc(#loc) + %127 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %128 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/biases"} -> tensor<120xbf16> loc(#loc) + %129 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %130 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/biases"} -> tensor<120xbf16> loc(#loc) + %131 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %132 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/biases"} -> tensor<40xbf16> loc(#loc) + %133 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/weights"} -> tensor<40x72x1x1xbf16> loc(#loc) + %134 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/biases"} -> tensor<72xbf16> loc(#loc) + %135 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %136 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/biases"} -> tensor<24xbf16> loc(#loc) + %137 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %138 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/biases"} -> tensor<72xbf16> loc(#loc) + %139 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/weights"} -> tensor<72x1x5x5xbf16> loc(#loc) + %140 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/biases"} -> tensor<72xbf16> loc(#loc) + %141 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %142 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/biases"} -> tensor<24xbf16> loc(#loc) + %143 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %144 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/biases"} -> tensor<72xbf16> loc(#loc) + %145 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/weights"} -> tensor<72x1x3x3xbf16> loc(#loc) + %146 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/biases"} -> tensor<72xbf16> loc(#loc) + %147 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %148 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/biases"} -> tensor<24xbf16> loc(#loc) + %149 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/weights"} -> tensor<24x64x1x1xbf16> loc(#loc) + %150 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/biases"} -> tensor<64xbf16> loc(#loc) + %151 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/weights"} -> tensor<64x1x3x3xbf16> loc(#loc) + %152 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/biases"} -> tensor<64xbf16> loc(#loc) + %153 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/weights"} -> tensor<64x16x1x1xbf16> loc(#loc) + %154 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/biases"} -> tensor<16xbf16> loc(#loc) + %155 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/weights"} -> tensor<16x16x1x1xbf16> loc(#loc) + %156 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/biases"} -> tensor<16xbf16> loc(#loc) + %157 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/weights"} -> tensor<16x1x3x3xbf16> loc(#loc) + %158 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/biases"} -> tensor<16xbf16> loc(#loc) + %159 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/weights"} -> tensor<16x3x3x3xbf16> loc(#loc) + %160 = xten_nn.load_external_const {file = "constants.h5", key = "Div_16/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc5) + %161 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_14/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc319) + %162 = xten_nn.subgraph (%arg5 = %arg1: tensor<1x90x160x16xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %472 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc8) + %473 = tosa.transpose %arg5, %472 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc8) + xten_nn.output %473 : tensor<1x16x90x160xbf16> loc(#loc8) + } -> tensor<1x16x90x160xbf16> loc(#loc8) + %163 = xten_nn.subgraph (%arg5 = %arg2: tensor<1x45x80x20xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %472 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc9) + %473 = tosa.transpose %arg5, %472 : (tensor<1x45x80x20xbf16>, tensor<4xi32>) -> tensor<1x20x45x80xbf16> loc(#loc9) + xten_nn.output %473 : tensor<1x20x45x80xbf16> loc(#loc9) + } -> tensor<1x20x45x80xbf16> loc(#loc9) + %164 = xten_nn.subgraph (%arg5 = %arg3: tensor<1x23x40x40xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %472 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc10) + %473 = tosa.transpose %arg5, %472 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc10) + xten_nn.output %473 : tensor<1x40x23x40xbf16> loc(#loc10) + } -> tensor<1x40x23x40xbf16> loc(#loc10) + %165 = tosa.cast %arg0 {LayerName = "Cast_0", OutputName = "Cast_0"} : (tensor<1x180x320x4xui8>) -> tensor<1x180x320x4xbf16> loc(#loc11) + %166 = xten_nn.subgraph (%arg5 = %165: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "386", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "386", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.906250e-03 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.906250e-03> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_2", + OutputName = "Div_2", + shift = 0 : i8} : (tensor<1x180x320x4xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc12) + xten_nn.output %474 : tensor<1x180x320x4xbf16> loc(#loc12) + } -> tensor<1x180x320x4xbf16> loc(#loc12) + xten_nn.output %472 : tensor<1x180x320x4xbf16> loc(#loc12) + } -> tensor<1x180x320x4xbf16> loc(#loc12) + %167 = xten_nn.subgraph (%arg5 = %166: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Slice_7_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Slice_7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "392", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "W", + config.dim_c = 184 : ui32, + config.dim_h = 320 : ui32, + config.dim_w = 4 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + LayerName = "Slice_7", + OutputName = "Slice_7", + size = array, + start = array} : (tensor<1x180x320x4xbf16>) -> tensor<1x180x320x3xbf16> loc(#loc13) + xten_nn.output %472 : tensor<1x180x320x3xbf16> loc(#loc13) + } -> tensor<1x180x320x3xbf16> loc(#loc13) + %168 = xten_nn.subgraph (%arg5 = %167: tensor<1x180x320x3xbf16>) attributes { + LayerName = "CompilerGenerated_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "CompilerGenerated_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "BufferPadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_0_padded = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_1_padded = 23 : ui32, + config.dim_2 = 3 : ui32, + config.dim_2_padded = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_padded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x180x320x3xbf16> loc(#loc14) + } -> tensor<1x180x320x3xbf16> loc(#loc14) + %169 = xten_nn.subgraph (%arg5 = %168: tensor<1x180x320x3xbf16>) attributes { + LayerName = "Slice_7_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Add_445_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_2 = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 10 : ui32 + }} { + %472 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc15) + %473 = tosa.transpose %arg5, %472 : (tensor<1x180x320x3xbf16>, tensor<4xi32>) -> tensor<1x3x180x320xbf16> loc(#loc321) + xten_nn.output %473 : tensor<1x3x180x320xbf16> loc(#loc321) + } -> tensor<1x3x180x320xbf16> loc(#loc320) + %170 = xten_nn.subgraph (%arg5 = %169: tensor<1x3x180x320xbf16>) attributes { + LayerName = "CompilerGenerated_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "CompilerGenerated_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 184 : ui32, + config.dim_0_unpadded = 180 : ui32, + config.dim_1 = 1 : ui32, + config.dim_1_unpadded = 1 : ui32, + config.dim_2 = 320 : ui32, + config.dim_2_unpadded = 320 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x3x180x320xbf16> loc(#loc14) + } -> tensor<1x3x180x320xbf16> loc(#loc14) + %171 = xten_nn.subgraph (%arg5 = %170: tensor<1x3x180x320xbf16>) attributes { + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc16) + %475 = tosa.transpose %arg6, %474 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc16) + %476 = tosa.avg_pool2d %475 { + PartOfLayerName = "AveragePool_346", + PartOfOutputName = "AveragePool_346", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>) -> tensor<1x90x160x3xbf16> loc(#loc16) + %477 = tosa.transpose %476, %473 : (tensor<1x90x160x3xbf16>, tensor<4xi32>) -> tensor<1x3x90x160xbf16> loc(#loc16) + xten_nn.output %477 : tensor<1x3x90x160xbf16> loc(#loc16) + } -> tensor<1x3x90x160xbf16> loc(#loc16) + xten_nn.output %472 : tensor<1x3x90x160xbf16> loc(#loc16) + } -> tensor<1x3x90x160xbf16> loc(#loc16) + %172 = xten_nn.subgraph (%arg5 = %171: tensor<1x3x90x160xbf16>) attributes { + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x90x160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc17) + %475 = tosa.transpose %arg6, %474 : (tensor<1x3x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x3xbf16> loc(#loc17) + %476 = tosa.avg_pool2d %475 { + PartOfLayerName = "AveragePool_347", + PartOfOutputName = "AveragePool_347", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x90x160x3xbf16>) -> tensor<1x45x80x3xbf16> loc(#loc17) + %477 = tosa.transpose %476, %473 : (tensor<1x45x80x3xbf16>, tensor<4xi32>) -> tensor<1x3x45x80xbf16> loc(#loc17) + xten_nn.output %477 : tensor<1x3x45x80xbf16> loc(#loc17) + } -> tensor<1x3x45x80xbf16> loc(#loc17) + xten_nn.output %472 : tensor<1x3x45x80xbf16> loc(#loc17) + } -> tensor<1x3x45x80xbf16> loc(#loc17) + %173 = xten_nn.subgraph (%arg5 = %172: tensor<1x3x45x80xbf16>) attributes { + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "780", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x45x80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 1], [0, 0]], + HWPaddingNotCounted = [[0, 1], [0, 0]], + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "780", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc18) + %475 = tosa.transpose %arg6, %474 : (tensor<1x3x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x3xbf16> loc(#loc18) + %476 = tosa.avg_pool2d %475 { + PartOfLayerName = "AveragePool_348", + PartOfOutputName = "AveragePool_348", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x45x80x3xbf16>) -> tensor<1x23x40x3xbf16> loc(#loc18) + %477 = tosa.transpose %476, %473 : (tensor<1x23x40x3xbf16>, tensor<4xi32>) -> tensor<1x3x23x40xbf16> loc(#loc18) + xten_nn.output %477 : tensor<1x3x23x40xbf16> loc(#loc18) + } -> tensor<1x3x23x40xbf16> loc(#loc18) + xten_nn.output %472 : tensor<1x3x23x40xbf16> loc(#loc18) + } -> tensor<1x3x23x40xbf16> loc(#loc18) + %174 = xten_nn.subgraph (%arg5 = %arg4: tensor<1x12x20x64xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %472 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc19) + %473 = tosa.transpose %arg5, %472 : (tensor<1x12x20x64xbf16>, tensor<4xi32>) -> tensor<1x64x12x20xbf16> loc(#loc19) + xten_nn.output %473 : tensor<1x64x12x20xbf16> loc(#loc19) + } -> tensor<1x64x12x20xbf16> loc(#loc19) + %175 = xten_nn.subgraph (%arg5 = %170: tensor<1x3x180x320xbf16>, %arg6 = %161: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "392", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Sub_14", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "392", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Sub_14", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.add %arg7, %arg8 {LayerName = "Sub_14", OutputName = "Initializer_398"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc319) + xten_nn.output %473 : tensor<1x3x180x320xbf16> loc(#loc319) + } -> tensor<1x3x180x320xbf16> loc(#loc319) + xten_nn.output %472 : tensor<1x3x180x320xbf16> loc(#loc319) + } -> tensor<1x3x180x320xbf16> loc(#loc319) + %176 = xten_nn.subgraph (%arg5 = %175: tensor<1x3x180x320xbf16>, %arg6 = %160: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + OutputName = "Div_16", + PartOfLayerName = "Div_16", + shift = 0 : i8} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc5) + xten_nn.output %473 : tensor<1x3x180x320xbf16> loc(#loc5) + } -> tensor<1x3x180x320xbf16> loc(#loc5) + xten_nn.output %472 : tensor<1x3x180x320xbf16> loc(#loc5) + } -> tensor<1x3x180x320xbf16> loc(#loc5) + %177 = xten_nn.subgraph (%arg5 = %176: tensor<1x3x180x320xbf16>, %arg6 = %159: tensor<16x3x3x3xbf16>, %arg7 = %158: tensor<16xbf16>) attributes { + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Name = "399", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Name = "929", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x3x180x320xbf16>, %arg9 = %arg6: tensor<16x3x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Name = "399", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Name = "929", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 2 : ui8, + config.stride_w = 2 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<16x3x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x3xbf16> loc(#loc20) + %476 = tosa.transpose %arg8, %474 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc20) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_17", + PartOfOutputName = "Conv_17", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>, tensor<16x3x3x3xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %478 = tosa.transpose %477, %473 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc20) + xten_nn.output %478 : tensor<1x16x90x160xbf16> loc(#loc20) + } -> tensor<1x16x90x160xbf16> loc(#loc20) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc20) + } -> tensor<1x16x90x160xbf16> loc(#loc20) + %178 = xten_nn.subgraph (%arg5 = %177: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_19", OutputName = "Add_19"} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %474 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc21) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc21) + %179 = xten_nn.subgraph (%arg5 = %178: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_22", + OutputName = "Clip_22", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc22) + xten_nn.output %473 : tensor<1x16x90x160xbf16> loc(#loc22) + } -> tensor<1x16x90x160xbf16> loc(#loc22) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc22) + } -> tensor<1x16x90x160xbf16> loc(#loc22) + %180 = xten_nn.subgraph (%arg5 = %179: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "410", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "410", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_24", + OutputName = "Div_24", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc23) + xten_nn.output %474 : tensor<1x16x90x160xbf16> loc(#loc23) + } -> tensor<1x16x90x160xbf16> loc(#loc23) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc23) + } -> tensor<1x16x90x160xbf16> loc(#loc23) + %181 = xten_nn.subgraph (%arg5 = %177: tensor<1x16x90x160xbf16>, %arg6 = %180: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_25", + OutputName = "Mul_25", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc24) + xten_nn.output %473 : tensor<1x16x90x160xbf16> loc(#loc24) + } -> tensor<1x16x90x160xbf16> loc(#loc24) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc24) + } -> tensor<1x16x90x160xbf16> loc(#loc24) + %182 = xten_nn.subgraph (%arg5 = %181: tensor<1x16x90x160xbf16>, %arg6 = %157: tensor<16x1x3x3xbf16>, %arg7 = %156: tensor<16xbf16>) attributes { + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "411", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<16x1x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "411", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc322) + %476 = tosa.transpose %arg9, %475 : (tensor<16x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x16x1xbf16> loc(#loc322) + %477 = tosa.transpose %arg8, %474 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc322) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_26", + PartOfOutputName = "Conv_26", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<3x3x16x1xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc25) + %479 = tosa.clamp %478 { + LayerName = "Relu_27", + OutputName = "Relu_27", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc26) + %480 = tosa.transpose %479, %473 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc322) + xten_nn.output %480 : tensor<1x16x90x160xbf16> loc(#loc26) + } -> tensor<1x16x90x160xbf16> loc(#loc322) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc322) + } -> tensor<1x16x90x160xbf16> loc(#loc322) + %183 = xten_nn.subgraph (%arg5 = %182: tensor<1x16x90x160xbf16>, %arg6 = %155: tensor<16x16x1x1xbf16>, %arg7 = %154: tensor<16xbf16>, %arg8 = %181: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Conv_28", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "931", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "935", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x16x90x160xbf16>, %arg10 = %arg6: tensor<16x16x1x1xbf16>, %arg11 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_28", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "931", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "935", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_28", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "934", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc27) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<16x16x1x1xbf16>) -> tensor<16x1x1x16xbf16> loc(#loc27) + %477 = tosa.transpose %arg9, %475 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc27) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_28", + PartOfOutputName = "Conv_28", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<16x1x1x16xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc27) + %479 = tosa.transpose %478, %474 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc27) + xten_nn.output %479 : tensor<1x16x90x160xbf16> loc(#loc27) + } -> tensor<1x16x90x160xbf16> loc(#loc27) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x16x90x160xbf16>, %arg10 = %arg8: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_29", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "934", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.add %arg9, %arg10 {LayerName = "Add_29", OutputName = "Add_29"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc28) + xten_nn.output %474 : tensor<1x16x90x160xbf16> loc(#loc28) + } -> tensor<1x16x90x160xbf16> loc(#loc28) + xten_nn.output %473 : tensor<1x16x90x160xbf16> loc(#loc28) + } -> tensor<1x16x90x160xbf16> loc(#loc323) + %184 = xten_nn.subgraph (%arg5 = %183: tensor<1x16x90x160xbf16>, %arg6 = %153: tensor<64x16x1x1xbf16>, %arg7 = %152: tensor<64xbf16>) attributes { + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "934", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "417", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<64x16x1x1xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "934", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "417", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc324) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<64x16x1x1xbf16>) -> tensor<64x1x1x16xbf16> loc(#loc324) + %476 = tosa.transpose %arg8, %474 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc324) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_30", + PartOfOutputName = "Conv_30", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<64x1x1x16xbf16>, tensor<64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc29) + %478 = tosa.clamp %477 { + LayerName = "Relu_31", + OutputName = "Relu_31", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc30) + %479 = tosa.transpose %478, %473 : (tensor<1x90x160x64xbf16>, tensor<4xi32>) -> tensor<1x64x90x160xbf16> loc(#loc324) + xten_nn.output %479 : tensor<1x64x90x160xbf16> loc(#loc30) + } -> tensor<1x64x90x160xbf16> loc(#loc324) + xten_nn.output %472 : tensor<1x64x90x160xbf16> loc(#loc324) + } -> tensor<1x64x90x160xbf16> loc(#loc324) + %185 = xten_nn.subgraph (%arg5 = %184: tensor<1x64x90x160xbf16>, %arg6 = %151: tensor<64x1x3x3xbf16>, %arg7 = %150: tensor<64xbf16>) attributes { + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "937", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "941", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x90x160xbf16>, %arg9 = %arg6: tensor<64x1x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "937", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "941", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc325) + %476 = tosa.transpose %arg9, %475 : (tensor<64x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x64x1xbf16> loc(#loc325) + %477 = tosa.transpose %arg8, %474 : (tensor<1x64x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x64xbf16> loc(#loc325) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_32", + PartOfOutputName = "Conv_32", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x64xbf16>, tensor<3x3x64x1xbf16>, tensor<64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc31) + %479 = tosa.clamp %478 { + LayerName = "Relu_33", + OutputName = "Relu_33", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc32) + %480 = tosa.transpose %479, %473 : (tensor<1x45x80x64xbf16>, tensor<4xi32>) -> tensor<1x64x45x80xbf16> loc(#loc325) + xten_nn.output %480 : tensor<1x64x45x80xbf16> loc(#loc32) + } -> tensor<1x64x45x80xbf16> loc(#loc325) + xten_nn.output %472 : tensor<1x64x45x80xbf16> loc(#loc325) + } -> tensor<1x64x45x80xbf16> loc(#loc325) + %186 = xten_nn.subgraph (%arg5 = %185: tensor<1x64x45x80xbf16>, %arg6 = %149: tensor<24x64x1x1xbf16>, %arg7 = %148: tensor<24xbf16>) attributes { + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Name = "940", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Name = "944", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x45x80xbf16>, %arg9 = %arg6: tensor<24x64x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Name = "940", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Name = "944", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc33) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x64x1x1xbf16>) -> tensor<24x1x1x64xbf16> loc(#loc33) + %476 = tosa.transpose %arg8, %474 : (tensor<1x64x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x64xbf16> loc(#loc33) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_34", + PartOfOutputName = "Conv_34", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x64xbf16>, tensor<24x1x1x64xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc33) + %478 = tosa.transpose %477, %473 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc33) + xten_nn.output %478 : tensor<1x24x45x80xbf16> loc(#loc33) + } -> tensor<1x24x45x80xbf16> loc(#loc33) + xten_nn.output %472 : tensor<1x24x45x80xbf16> loc(#loc33) + } -> tensor<1x24x45x80xbf16> loc(#loc33) + %187 = xten_nn.subgraph (%arg5 = %186: tensor<1x24x45x80xbf16>, %arg6 = %147: tensor<72x24x1x1xbf16>, %arg7 = %146: tensor<72xbf16>) attributes { + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "423", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "947", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "423", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "947", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc326) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc326) + %476 = tosa.transpose %arg8, %474 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc326) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_35", + PartOfOutputName = "Conv_35", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc34) + %478 = tosa.clamp %477 { + LayerName = "Relu_36", + OutputName = "Relu_36", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc35) + %479 = tosa.transpose %478, %473 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc326) + xten_nn.output %479 : tensor<1x72x45x80xbf16> loc(#loc35) + } -> tensor<1x72x45x80xbf16> loc(#loc326) + xten_nn.output %472 : tensor<1x72x45x80xbf16> loc(#loc326) + } -> tensor<1x72x45x80xbf16> loc(#loc326) + %188 = xten_nn.subgraph (%arg5 = %187: tensor<1x72x45x80xbf16>, %arg6 = %145: tensor<72x1x3x3xbf16>, %arg7 = %144: tensor<72xbf16>) attributes { + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "946", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "950", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x3x3xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "946", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "950", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc327) + %476 = tosa.transpose %arg9, %475 : (tensor<72x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x72x1xbf16> loc(#loc327) + %477 = tosa.transpose %arg8, %474 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc327) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_37", + PartOfOutputName = "Conv_37", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<3x3x72x1xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc36) + %479 = tosa.clamp %478 { + LayerName = "Relu_38", + OutputName = "Relu_38", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc37) + %480 = tosa.transpose %479, %473 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc327) + xten_nn.output %480 : tensor<1x72x45x80xbf16> loc(#loc37) + } -> tensor<1x72x45x80xbf16> loc(#loc327) + xten_nn.output %472 : tensor<1x72x45x80xbf16> loc(#loc327) + } -> tensor<1x72x45x80xbf16> loc(#loc327) + %189 = xten_nn.subgraph (%arg5 = %188: tensor<1x72x45x80xbf16>, %arg6 = %143: tensor<24x72x1x1xbf16>, %arg7 = %142: tensor<24xbf16>, %arg8 = %186: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Conv_39", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Name = "949", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "953", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x72x45x80xbf16>, %arg10 = %arg6: tensor<24x72x1x1xbf16>, %arg11 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_39", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Name = "949", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "953", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_39", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "952", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc38) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc38) + %477 = tosa.transpose %arg9, %475 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc38) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_39", + PartOfOutputName = "Conv_39", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc38) + %479 = tosa.transpose %478, %474 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc38) + xten_nn.output %479 : tensor<1x24x45x80xbf16> loc(#loc38) + } -> tensor<1x24x45x80xbf16> loc(#loc38) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x24x45x80xbf16>, %arg10 = %arg8: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Add_40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "952", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.add %arg9, %arg10 {LayerName = "Add_40", OutputName = "Add_40"} : (tensor<1x24x45x80xbf16>, tensor<1x24x45x80xbf16>) -> tensor<1x24x45x80xbf16> loc(#loc39) + xten_nn.output %474 : tensor<1x24x45x80xbf16> loc(#loc39) + } -> tensor<1x24x45x80xbf16> loc(#loc39) + xten_nn.output %473 : tensor<1x24x45x80xbf16> loc(#loc39) + } -> tensor<1x24x45x80xbf16> loc(#loc328) + %190 = xten_nn.subgraph (%arg5 = %189: tensor<1x24x45x80xbf16>, %arg6 = %141: tensor<72x24x1x1xbf16>, %arg7 = %140: tensor<72xbf16>) attributes { + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "952", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "434", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "952", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "434", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc329) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc329) + %476 = tosa.transpose %arg8, %474 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc329) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_41", + PartOfOutputName = "Conv_41", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc40) + %478 = tosa.clamp %477 { + LayerName = "Relu_42", + OutputName = "Relu_42", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc41) + %479 = tosa.transpose %478, %473 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc329) + xten_nn.output %479 : tensor<1x72x45x80xbf16> loc(#loc41) + } -> tensor<1x72x45x80xbf16> loc(#loc329) + xten_nn.output %472 : tensor<1x72x45x80xbf16> loc(#loc329) + } -> tensor<1x72x45x80xbf16> loc(#loc329) + %191 = xten_nn.subgraph (%arg5 = %190: tensor<1x72x45x80xbf16>, %arg6 = %139: tensor<72x1x5x5xbf16>, %arg7 = %138: tensor<72xbf16>) attributes { + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "955", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "959", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x5x5xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 1]], + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "955", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "959", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 2 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc330) + %476 = tosa.transpose %arg9, %475 : (tensor<72x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x72x1xbf16> loc(#loc330) + %477 = tosa.transpose %arg8, %474 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc330) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_43", + PartOfOutputName = "Conv_43", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<5x5x72x1xbf16>, tensor<72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc42) + %479 = tosa.clamp %478 { + LayerName = "Relu_44", + OutputName = "Relu_44", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc43) + %480 = tosa.transpose %479, %473 : (tensor<1x23x40x72xbf16>, tensor<4xi32>) -> tensor<1x72x23x40xbf16> loc(#loc330) + xten_nn.output %480 : tensor<1x72x23x40xbf16> loc(#loc43) + } -> tensor<1x72x23x40xbf16> loc(#loc330) + xten_nn.output %472 : tensor<1x72x23x40xbf16> loc(#loc330) + } -> tensor<1x72x23x40xbf16> loc(#loc330) + %192 = xten_nn.subgraph (%arg5 = %191: tensor<1x72x23x40xbf16>) attributes { + LayerName = "GlobalAveragePool_45_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_45_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 9 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %472 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x72x23x40xbf16>) -> tensor<1x72x1x920xbf16> loc(#loc44) + xten_nn.output %472 : tensor<1x72x1x920xbf16> loc(#loc44) + } -> tensor<1x72x1x920xbf16> loc(#loc44) + %193 = xten_nn.subgraph (%arg5 = %192: tensor<1x72x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_45_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_45_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_45_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_45_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 72 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %473 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x72x1x920xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc44) + xten_nn.output %473 : tensor<1x72x1x1xbf16> loc(#loc44) + } -> tensor<1x72x1x1xbf16> loc(#loc44) + xten_nn.output %472 : tensor<1x72x1x1xbf16> loc(#loc44) + } -> tensor<1x72x1x1xbf16> loc(#loc44) + %194 = xten_nn.subgraph (%arg5 = %193: tensor<1x72x1x1xbf16>, %arg6 = %137: tensor<24x72x1x1xbf16>, %arg7 = %136: tensor<24xbf16>) attributes { + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "440", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x1x1xbf16>, %arg9 = %arg6: tensor<24x72x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "440", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc331) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc331) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_46", + PartOfOutputName = "Conv_46", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc45) + %476 = tosa.clamp %475 { + LayerName = "Relu_47", + OutputName = "Relu_47", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc46) + %477 = tosa.reshape %476 {new_shape = array} : (tensor<1x1x1x24xbf16>) -> tensor<1x24x1x1xbf16> loc(#loc331) + xten_nn.output %477 : tensor<1x24x1x1xbf16> loc(#loc46) + } -> tensor<1x24x1x1xbf16> loc(#loc331) + xten_nn.output %472 : tensor<1x24x1x1xbf16> loc(#loc331) + } -> tensor<1x24x1x1xbf16> loc(#loc331) + %195 = xten_nn.subgraph (%arg5 = %194: tensor<1x24x1x1xbf16>, %arg6 = %135: tensor<72x24x1x1xbf16>, %arg7 = %134: tensor<72xbf16>) attributes { + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "442", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x1x1xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "442", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc47) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x24x1x1xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc47) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_48", + PartOfOutputName = "Conv_48", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc47) + %476 = tosa.reshape %475 {new_shape = array} : (tensor<1x1x1x72xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc47) + xten_nn.output %476 : tensor<1x72x1x1xbf16> loc(#loc47) + } -> tensor<1x72x1x1xbf16> loc(#loc47) + xten_nn.output %472 : tensor<1x72x1x1xbf16> loc(#loc47) + } -> tensor<1x72x1x1xbf16> loc(#loc47) + %196 = xten_nn.subgraph (%arg5 = %195: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_50", OutputName = "Add_50"} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc48) + xten_nn.output %474 : tensor<1x72x1x1xbf16> loc(#loc48) + } -> tensor<1x72x1x1xbf16> loc(#loc48) + xten_nn.output %472 : tensor<1x72x1x1xbf16> loc(#loc48) + } -> tensor<1x72x1x1xbf16> loc(#loc48) + %197 = xten_nn.subgraph (%arg5 = %196: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_53", + OutputName = "Clip_53", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc49) + xten_nn.output %473 : tensor<1x72x1x1xbf16> loc(#loc49) + } -> tensor<1x72x1x1xbf16> loc(#loc49) + xten_nn.output %472 : tensor<1x72x1x1xbf16> loc(#loc49) + } -> tensor<1x72x1x1xbf16> loc(#loc49) + %198 = xten_nn.subgraph (%arg5 = %197: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_55", + OutputName = "Div_55", + shift = 0 : i8} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc50) + xten_nn.output %474 : tensor<1x72x1x1xbf16> loc(#loc50) + } -> tensor<1x72x1x1xbf16> loc(#loc50) + xten_nn.output %472 : tensor<1x72x1x1xbf16> loc(#loc50) + } -> tensor<1x72x1x1xbf16> loc(#loc50) + %199 = xten_nn.subgraph (%arg5 = %198: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Mul_56_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_56_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 72 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %472 = tosa.tile %arg5 {multiples = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc51) + xten_nn.output %472 : tensor<1x72x23x40xbf16> loc(#loc51) + } -> tensor<1x72x23x40xbf16> loc(#loc51) + %200 = xten_nn.subgraph (%arg5 = %199: tensor<1x72x23x40xbf16>, %arg6 = %191: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x72x23x40xbf16>, %arg8 = %arg6: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_56", + OutputName = "Mul_56", + shift = 0 : i8} : (tensor<1x72x23x40xbf16>, tensor<1x72x23x40xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc51) + xten_nn.output %473 : tensor<1x72x23x40xbf16> loc(#loc51) + } -> tensor<1x72x23x40xbf16> loc(#loc51) + xten_nn.output %472 : tensor<1x72x23x40xbf16> loc(#loc51) + } -> tensor<1x72x23x40xbf16> loc(#loc51) + %201 = xten_nn.subgraph (%arg5 = %200: tensor<1x72x23x40xbf16>, %arg6 = %133: tensor<40x72x1x1xbf16>, %arg7 = %132: tensor<40xbf16>) attributes { + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Name = "451", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "452", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x23x40xbf16>, %arg9 = %arg6: tensor<40x72x1x1xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Name = "451", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "452", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc52) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<40x72x1x1xbf16>) -> tensor<40x1x1x72xbf16> loc(#loc52) + %476 = tosa.transpose %arg8, %474 : (tensor<1x72x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x72xbf16> loc(#loc52) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_57", + PartOfOutputName = "Conv_57", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x72xbf16>, tensor<40x1x1x72xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc52) + %478 = tosa.transpose %477, %473 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc52) + xten_nn.output %478 : tensor<1x40x23x40xbf16> loc(#loc52) + } -> tensor<1x40x23x40xbf16> loc(#loc52) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc52) + } -> tensor<1x40x23x40xbf16> loc(#loc52) + %202 = xten_nn.subgraph (%arg5 = %201: tensor<1x40x23x40xbf16>, %arg6 = %131: tensor<120x40x1x1xbf16>, %arg7 = %130: tensor<120xbf16>) attributes { + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "452", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "965", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "452", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "965", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc332) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc332) + %476 = tosa.transpose %arg8, %474 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc332) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_58", + PartOfOutputName = "Conv_58", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc53) + %478 = tosa.clamp %477 { + LayerName = "Relu_59", + OutputName = "Relu_59", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc54) + %479 = tosa.transpose %478, %473 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc332) + xten_nn.output %479 : tensor<1x120x23x40xbf16> loc(#loc54) + } -> tensor<1x120x23x40xbf16> loc(#loc332) + xten_nn.output %472 : tensor<1x120x23x40xbf16> loc(#loc332) + } -> tensor<1x120x23x40xbf16> loc(#loc332) + %203 = xten_nn.subgraph (%arg5 = %202: tensor<1x120x23x40xbf16>, %arg6 = %129: tensor<120x1x5x5xbf16>, %arg7 = %128: tensor<120xbf16>) attributes { + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "964", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "968", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "964", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "968", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc333) + %476 = tosa.transpose %arg9, %475 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc333) + %477 = tosa.transpose %arg8, %474 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc333) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_60", + PartOfOutputName = "Conv_60", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc55) + %479 = tosa.clamp %478 { + LayerName = "Relu_61", + OutputName = "Relu_61", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc56) + %480 = tosa.transpose %479, %473 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc333) + xten_nn.output %480 : tensor<1x120x23x40xbf16> loc(#loc56) + } -> tensor<1x120x23x40xbf16> loc(#loc333) + xten_nn.output %472 : tensor<1x120x23x40xbf16> loc(#loc333) + } -> tensor<1x120x23x40xbf16> loc(#loc333) + %204 = xten_nn.subgraph (%arg5 = %203: tensor<1x120x23x40xbf16>) attributes { + LayerName = "GlobalAveragePool_62_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_62_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %472 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc57) + xten_nn.output %472 : tensor<1x120x1x920xbf16> loc(#loc57) + } -> tensor<1x120x1x920xbf16> loc(#loc57) + %205 = xten_nn.subgraph (%arg5 = %204: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_62_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_62_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_62_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_62_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %473 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc57) + xten_nn.output %473 : tensor<1x120x1x1xbf16> loc(#loc57) + } -> tensor<1x120x1x1xbf16> loc(#loc57) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc57) + } -> tensor<1x120x1x1xbf16> loc(#loc57) + %206 = xten_nn.subgraph (%arg5 = %205: tensor<1x120x1x1xbf16>, %arg6 = %127: tensor<32x120x1x1xbf16>, %arg7 = %126: tensor<32xbf16>) attributes { + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "460", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "460", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc334) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc334) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_63", + PartOfOutputName = "Conv_63", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc58) + %476 = tosa.clamp %475 { + LayerName = "Relu_64", + OutputName = "Relu_64", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc59) + %477 = tosa.reshape %476 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc334) + xten_nn.output %477 : tensor<1x32x1x1xbf16> loc(#loc59) + } -> tensor<1x32x1x1xbf16> loc(#loc334) + xten_nn.output %472 : tensor<1x32x1x1xbf16> loc(#loc334) + } -> tensor<1x32x1x1xbf16> loc(#loc334) + %207 = xten_nn.subgraph (%arg5 = %206: tensor<1x32x1x1xbf16>, %arg6 = %125: tensor<120x32x1x1xbf16>, %arg7 = %124: tensor<120xbf16>) attributes { + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "462", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "462", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc60) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc60) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_65", + PartOfOutputName = "Conv_65", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc60) + %476 = tosa.reshape %475 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc60) + xten_nn.output %476 : tensor<1x120x1x1xbf16> loc(#loc60) + } -> tensor<1x120x1x1xbf16> loc(#loc60) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc60) + } -> tensor<1x120x1x1xbf16> loc(#loc60) + %208 = xten_nn.subgraph (%arg5 = %207: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_67", OutputName = "Add_67"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc61) + xten_nn.output %474 : tensor<1x120x1x1xbf16> loc(#loc61) + } -> tensor<1x120x1x1xbf16> loc(#loc61) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc61) + } -> tensor<1x120x1x1xbf16> loc(#loc61) + %209 = xten_nn.subgraph (%arg5 = %208: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_70", + OutputName = "Clip_70", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc62) + xten_nn.output %473 : tensor<1x120x1x1xbf16> loc(#loc62) + } -> tensor<1x120x1x1xbf16> loc(#loc62) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc62) + } -> tensor<1x120x1x1xbf16> loc(#loc62) + %210 = xten_nn.subgraph (%arg5 = %209: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_72", + OutputName = "Div_72", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc63) + xten_nn.output %474 : tensor<1x120x1x1xbf16> loc(#loc63) + } -> tensor<1x120x1x1xbf16> loc(#loc63) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc63) + } -> tensor<1x120x1x1xbf16> loc(#loc63) + %211 = xten_nn.subgraph (%arg5 = %210: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Mul_73_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_73_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %472 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc64) + xten_nn.output %472 : tensor<1x120x23x40xbf16> loc(#loc64) + } -> tensor<1x120x23x40xbf16> loc(#loc64) + %212 = xten_nn.subgraph (%arg5 = %211: tensor<1x120x23x40xbf16>, %arg6 = %203: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_73", + OutputName = "Mul_73", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc64) + xten_nn.output %473 : tensor<1x120x23x40xbf16> loc(#loc64) + } -> tensor<1x120x23x40xbf16> loc(#loc64) + xten_nn.output %472 : tensor<1x120x23x40xbf16> loc(#loc64) + } -> tensor<1x120x23x40xbf16> loc(#loc64) + %213 = xten_nn.subgraph (%arg5 = %212: tensor<1x120x23x40xbf16>, %arg6 = %123: tensor<40x120x1x1xbf16>, %arg7 = %122: tensor<40xbf16>, %arg8 = %201: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_74", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "471", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "472", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_74", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "471", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "472", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_74", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "970", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc65) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc65) + %477 = tosa.transpose %arg9, %475 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc65) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_74", + PartOfOutputName = "Conv_74", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc65) + %479 = tosa.transpose %478, %474 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc65) + xten_nn.output %479 : tensor<1x40x23x40xbf16> loc(#loc65) + } -> tensor<1x40x23x40xbf16> loc(#loc65) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_75", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "970", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.add %arg9, %arg10 {LayerName = "Add_75", OutputName = "Add_75"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc66) + xten_nn.output %474 : tensor<1x40x23x40xbf16> loc(#loc66) + } -> tensor<1x40x23x40xbf16> loc(#loc66) + xten_nn.output %473 : tensor<1x40x23x40xbf16> loc(#loc66) + } -> tensor<1x40x23x40xbf16> loc(#loc335) + %214 = xten_nn.subgraph (%arg5 = %213: tensor<1x40x23x40xbf16>, %arg6 = %121: tensor<120x40x1x1xbf16>, %arg7 = %120: tensor<120xbf16>) attributes { + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "970", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "475", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "970", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "475", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc336) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc336) + %476 = tosa.transpose %arg8, %474 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc336) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_76", + PartOfOutputName = "Conv_76", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc67) + %478 = tosa.clamp %477 { + LayerName = "Relu_77", + OutputName = "Relu_77", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc68) + %479 = tosa.transpose %478, %473 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc336) + xten_nn.output %479 : tensor<1x120x23x40xbf16> loc(#loc68) + } -> tensor<1x120x23x40xbf16> loc(#loc336) + xten_nn.output %472 : tensor<1x120x23x40xbf16> loc(#loc336) + } -> tensor<1x120x23x40xbf16> loc(#loc336) + %215 = xten_nn.subgraph (%arg5 = %214: tensor<1x120x23x40xbf16>, %arg6 = %119: tensor<120x1x5x5xbf16>, %arg7 = %118: tensor<120xbf16>) attributes { + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "973", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "977", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "973", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "977", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc337) + %476 = tosa.transpose %arg9, %475 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc337) + %477 = tosa.transpose %arg8, %474 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc337) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_78", + PartOfOutputName = "Conv_78", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc69) + %479 = tosa.clamp %478 { + LayerName = "Relu_79", + OutputName = "Relu_79", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc70) + %480 = tosa.transpose %479, %473 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc337) + xten_nn.output %480 : tensor<1x120x23x40xbf16> loc(#loc70) + } -> tensor<1x120x23x40xbf16> loc(#loc337) + xten_nn.output %472 : tensor<1x120x23x40xbf16> loc(#loc337) + } -> tensor<1x120x23x40xbf16> loc(#loc337) + %216 = xten_nn.subgraph (%arg5 = %215: tensor<1x120x23x40xbf16>) attributes { + LayerName = "GlobalAveragePool_80_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_80_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %472 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc71) + xten_nn.output %472 : tensor<1x120x1x920xbf16> loc(#loc71) + } -> tensor<1x120x1x920xbf16> loc(#loc71) + %217 = xten_nn.subgraph (%arg5 = %216: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_80_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_80_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_80_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_80_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %473 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc71) + xten_nn.output %473 : tensor<1x120x1x1xbf16> loc(#loc71) + } -> tensor<1x120x1x1xbf16> loc(#loc71) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc71) + } -> tensor<1x120x1x1xbf16> loc(#loc71) + %218 = xten_nn.subgraph (%arg5 = %217: tensor<1x120x1x1xbf16>, %arg6 = %117: tensor<32x120x1x1xbf16>, %arg7 = %116: tensor<32xbf16>) attributes { + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "481", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "481", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc338) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc338) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_81", + PartOfOutputName = "Conv_81", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc72) + %476 = tosa.clamp %475 { + LayerName = "Relu_82", + OutputName = "Relu_82", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc73) + %477 = tosa.reshape %476 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc338) + xten_nn.output %477 : tensor<1x32x1x1xbf16> loc(#loc73) + } -> tensor<1x32x1x1xbf16> loc(#loc338) + xten_nn.output %472 : tensor<1x32x1x1xbf16> loc(#loc338) + } -> tensor<1x32x1x1xbf16> loc(#loc338) + %219 = xten_nn.subgraph (%arg5 = %218: tensor<1x32x1x1xbf16>, %arg6 = %115: tensor<120x32x1x1xbf16>, %arg7 = %114: tensor<120xbf16>) attributes { + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "483", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "483", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc74) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc74) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_83", + PartOfOutputName = "Conv_83", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc74) + %476 = tosa.reshape %475 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc74) + xten_nn.output %476 : tensor<1x120x1x1xbf16> loc(#loc74) + } -> tensor<1x120x1x1xbf16> loc(#loc74) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc74) + } -> tensor<1x120x1x1xbf16> loc(#loc74) + %220 = xten_nn.subgraph (%arg5 = %219: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_85", OutputName = "Add_85"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc75) + xten_nn.output %474 : tensor<1x120x1x1xbf16> loc(#loc75) + } -> tensor<1x120x1x1xbf16> loc(#loc75) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc75) + } -> tensor<1x120x1x1xbf16> loc(#loc75) + %221 = xten_nn.subgraph (%arg5 = %220: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_88", + OutputName = "Clip_88", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc76) + xten_nn.output %473 : tensor<1x120x1x1xbf16> loc(#loc76) + } -> tensor<1x120x1x1xbf16> loc(#loc76) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc76) + } -> tensor<1x120x1x1xbf16> loc(#loc76) + %222 = xten_nn.subgraph (%arg5 = %221: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_90", + OutputName = "Div_90", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc77) + xten_nn.output %474 : tensor<1x120x1x1xbf16> loc(#loc77) + } -> tensor<1x120x1x1xbf16> loc(#loc77) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc77) + } -> tensor<1x120x1x1xbf16> loc(#loc77) + %223 = xten_nn.subgraph (%arg5 = %222: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Mul_91_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_91_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %472 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc78) + xten_nn.output %472 : tensor<1x120x23x40xbf16> loc(#loc78) + } -> tensor<1x120x23x40xbf16> loc(#loc78) + %224 = xten_nn.subgraph (%arg5 = %223: tensor<1x120x23x40xbf16>, %arg6 = %215: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_91", + OutputName = "Mul_91", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc78) + xten_nn.output %473 : tensor<1x120x23x40xbf16> loc(#loc78) + } -> tensor<1x120x23x40xbf16> loc(#loc78) + xten_nn.output %472 : tensor<1x120x23x40xbf16> loc(#loc78) + } -> tensor<1x120x23x40xbf16> loc(#loc78) + %225 = xten_nn.subgraph (%arg5 = %224: tensor<1x120x23x40xbf16>, %arg6 = %113: tensor<40x120x1x1xbf16>, %arg7 = %112: tensor<40xbf16>, %arg8 = %213: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_92", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "492", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "493", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_92", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "492", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "493", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_92", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "979", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc79) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc79) + %477 = tosa.transpose %arg9, %475 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc79) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_92", + PartOfOutputName = "Conv_92", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc79) + %479 = tosa.transpose %478, %474 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc79) + xten_nn.output %479 : tensor<1x40x23x40xbf16> loc(#loc79) + } -> tensor<1x40x23x40xbf16> loc(#loc79) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_93", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "979", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.add %arg9, %arg10 {LayerName = "Add_93", OutputName = "Add_93"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc80) + xten_nn.output %474 : tensor<1x40x23x40xbf16> loc(#loc80) + } -> tensor<1x40x23x40xbf16> loc(#loc80) + xten_nn.output %473 : tensor<1x40x23x40xbf16> loc(#loc80) + } -> tensor<1x40x23x40xbf16> loc(#loc339) + %226 = xten_nn.subgraph (%arg5 = %225: tensor<1x40x23x40xbf16>, %arg6 = %111: tensor<240x40x1x1xbf16>, %arg7 = %110: tensor<240xbf16>) attributes { + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "979", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "496", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<240x40x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "979", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "496", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc81) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x40x1x1xbf16>) -> tensor<240x1x1x40xbf16> loc(#loc81) + %476 = tosa.transpose %arg8, %474 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc81) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_94", + PartOfOutputName = "Conv_94", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<240x1x1x40xbf16>, tensor<240xbf16>) -> tensor<1x23x40x240xbf16> loc(#loc81) + %478 = tosa.transpose %477, %473 : (tensor<1x23x40x240xbf16>, tensor<4xi32>) -> tensor<1x240x23x40xbf16> loc(#loc81) + xten_nn.output %478 : tensor<1x240x23x40xbf16> loc(#loc81) + } -> tensor<1x240x23x40xbf16> loc(#loc81) + xten_nn.output %472 : tensor<1x240x23x40xbf16> loc(#loc81) + } -> tensor<1x240x23x40xbf16> loc(#loc81) + %227 = xten_nn.subgraph (%arg5 = %226: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_96", OutputName = "Add_96"} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc82) + xten_nn.output %474 : tensor<1x240x23x40xbf16> loc(#loc82) + } -> tensor<1x240x23x40xbf16> loc(#loc82) + xten_nn.output %472 : tensor<1x240x23x40xbf16> loc(#loc82) + } -> tensor<1x240x23x40xbf16> loc(#loc82) + %228 = xten_nn.subgraph (%arg5 = %227: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_99", + OutputName = "Clip_99", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc83) + xten_nn.output %473 : tensor<1x240x23x40xbf16> loc(#loc83) + } -> tensor<1x240x23x40xbf16> loc(#loc83) + xten_nn.output %472 : tensor<1x240x23x40xbf16> loc(#loc83) + } -> tensor<1x240x23x40xbf16> loc(#loc83) + %229 = xten_nn.subgraph (%arg5 = %228: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "505", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "505", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_101", + OutputName = "Div_101", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc84) + xten_nn.output %474 : tensor<1x240x23x40xbf16> loc(#loc84) + } -> tensor<1x240x23x40xbf16> loc(#loc84) + xten_nn.output %472 : tensor<1x240x23x40xbf16> loc(#loc84) + } -> tensor<1x240x23x40xbf16> loc(#loc84) + %230 = xten_nn.subgraph (%arg5 = %226: tensor<1x240x23x40xbf16>, %arg6 = %229: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x23x40xbf16>, %arg8 = %arg6: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_102", + OutputName = "Mul_102", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc85) + xten_nn.output %473 : tensor<1x240x23x40xbf16> loc(#loc85) + } -> tensor<1x240x23x40xbf16> loc(#loc85) + xten_nn.output %472 : tensor<1x240x23x40xbf16> loc(#loc85) + } -> tensor<1x240x23x40xbf16> loc(#loc85) + %231 = xten_nn.subgraph (%arg5 = %230: tensor<1x240x23x40xbf16>, %arg6 = %109: tensor<240x1x3x3xbf16>, %arg7 = %108: tensor<240xbf16>) attributes { + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "506", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x23x40xbf16>, %arg9 = %arg6: tensor<240x1x3x3xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 0]], + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "506", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc86) + %476 = tosa.transpose %arg9, %475 : (tensor<240x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x240x1xbf16> loc(#loc86) + %477 = tosa.transpose %arg8, %474 : (tensor<1x240x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x240xbf16> loc(#loc86) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_103", + PartOfOutputName = "Conv_103", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x240xbf16>, tensor<3x3x240x1xbf16>, tensor<240xbf16>) -> tensor<1x12x20x240xbf16> loc(#loc86) + %479 = tosa.transpose %478, %473 : (tensor<1x12x20x240xbf16>, tensor<4xi32>) -> tensor<1x240x12x20xbf16> loc(#loc86) + xten_nn.output %479 : tensor<1x240x12x20xbf16> loc(#loc86) + } -> tensor<1x240x12x20xbf16> loc(#loc86) + xten_nn.output %472 : tensor<1x240x12x20xbf16> loc(#loc86) + } -> tensor<1x240x12x20xbf16> loc(#loc86) + %232 = xten_nn.subgraph (%arg5 = %231: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_105", OutputName = "Add_105"} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc87) + xten_nn.output %474 : tensor<1x240x12x20xbf16> loc(#loc87) + } -> tensor<1x240x12x20xbf16> loc(#loc87) + xten_nn.output %472 : tensor<1x240x12x20xbf16> loc(#loc87) + } -> tensor<1x240x12x20xbf16> loc(#loc87) + %233 = xten_nn.subgraph (%arg5 = %232: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_108", + OutputName = "Clip_108", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc88) + xten_nn.output %473 : tensor<1x240x12x20xbf16> loc(#loc88) + } -> tensor<1x240x12x20xbf16> loc(#loc88) + xten_nn.output %472 : tensor<1x240x12x20xbf16> loc(#loc88) + } -> tensor<1x240x12x20xbf16> loc(#loc88) + %234 = xten_nn.subgraph (%arg5 = %233: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "515", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "515", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_110", + OutputName = "Div_110", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc89) + xten_nn.output %474 : tensor<1x240x12x20xbf16> loc(#loc89) + } -> tensor<1x240x12x20xbf16> loc(#loc89) + xten_nn.output %472 : tensor<1x240x12x20xbf16> loc(#loc89) + } -> tensor<1x240x12x20xbf16> loc(#loc89) + %235 = xten_nn.subgraph (%arg5 = %231: tensor<1x240x12x20xbf16>, %arg6 = %234: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x12x20xbf16>, %arg8 = %arg6: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_111", + OutputName = "Mul_111", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc90) + xten_nn.output %473 : tensor<1x240x12x20xbf16> loc(#loc90) + } -> tensor<1x240x12x20xbf16> loc(#loc90) + xten_nn.output %472 : tensor<1x240x12x20xbf16> loc(#loc90) + } -> tensor<1x240x12x20xbf16> loc(#loc90) + %236 = xten_nn.subgraph (%arg5 = %235: tensor<1x240x12x20xbf16>, %arg6 = %107: tensor<80x240x1x1xbf16>, %arg7 = %106: tensor<80xbf16>) attributes { + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Name = "985", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "516", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x12x20xbf16>, %arg9 = %arg6: tensor<80x240x1x1xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Name = "985", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "516", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc91) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<80x240x1x1xbf16>) -> tensor<80x1x1x240xbf16> loc(#loc91) + %476 = tosa.transpose %arg8, %474 : (tensor<1x240x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x240xbf16> loc(#loc91) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_112", + PartOfOutputName = "Conv_112", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x240xbf16>, tensor<80x1x1x240xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc91) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc91) + xten_nn.output %478 : tensor<1x80x12x20xbf16> loc(#loc91) + } -> tensor<1x80x12x20xbf16> loc(#loc91) + xten_nn.output %472 : tensor<1x80x12x20xbf16> loc(#loc91) + } -> tensor<1x80x12x20xbf16> loc(#loc91) + %237 = xten_nn.subgraph (%arg5 = %236: tensor<1x80x12x20xbf16>, %arg6 = %105: tensor<200x80x1x1xbf16>, %arg7 = %104: tensor<200xbf16>) attributes { + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "516", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "992", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<200x80x1x1xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "516", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "992", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc92) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<200x80x1x1xbf16>) -> tensor<200x1x1x80xbf16> loc(#loc92) + %476 = tosa.transpose %arg8, %474 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc92) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_113", + PartOfOutputName = "Conv_113", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<200x1x1x80xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc92) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %478 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %472 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + %238 = xten_nn.subgraph (%arg5 = %237: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_115", OutputName = "Add_115"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %474 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %472 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + %239 = xten_nn.subgraph (%arg5 = %238: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_118", + OutputName = "Clip_118", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %473 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %472 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + %240 = xten_nn.subgraph (%arg5 = %239: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "527", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "527", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_120", + OutputName = "Div_120", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc95) + xten_nn.output %474 : tensor<1x200x12x20xbf16> loc(#loc95) + } -> tensor<1x200x12x20xbf16> loc(#loc95) + xten_nn.output %472 : tensor<1x200x12x20xbf16> loc(#loc95) + } -> tensor<1x200x12x20xbf16> loc(#loc95) + %241 = xten_nn.subgraph (%arg5 = %237: tensor<1x200x12x20xbf16>, %arg6 = %240: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_121", + OutputName = "Mul_121", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc96) + xten_nn.output %473 : tensor<1x200x12x20xbf16> loc(#loc96) + } -> tensor<1x200x12x20xbf16> loc(#loc96) + xten_nn.output %472 : tensor<1x200x12x20xbf16> loc(#loc96) + } -> tensor<1x200x12x20xbf16> loc(#loc96) + %242 = xten_nn.subgraph (%arg5 = %241: tensor<1x200x12x20xbf16>, %arg6 = %103: tensor<200x1x3x3xbf16>, %arg7 = %102: tensor<200xbf16>) attributes { + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "528", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x200x12x20xbf16>, %arg9 = %arg6: tensor<200x1x3x3xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "528", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc97) + %476 = tosa.transpose %arg9, %475 : (tensor<200x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x200x1xbf16> loc(#loc97) + %477 = tosa.transpose %arg8, %474 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc97) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_122", + PartOfOutputName = "Conv_122", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<3x3x200x1xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc97) + %479 = tosa.transpose %478, %473 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc97) + xten_nn.output %479 : tensor<1x200x12x20xbf16> loc(#loc97) + } -> tensor<1x200x12x20xbf16> loc(#loc97) + xten_nn.output %472 : tensor<1x200x12x20xbf16> loc(#loc97) + } -> tensor<1x200x12x20xbf16> loc(#loc97) + %243 = xten_nn.subgraph (%arg5 = %242: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_124", OutputName = "Add_124"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc98) + xten_nn.output %474 : tensor<1x200x12x20xbf16> loc(#loc98) + } -> tensor<1x200x12x20xbf16> loc(#loc98) + xten_nn.output %472 : tensor<1x200x12x20xbf16> loc(#loc98) + } -> tensor<1x200x12x20xbf16> loc(#loc98) + %244 = xten_nn.subgraph (%arg5 = %243: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_127", + OutputName = "Clip_127", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc99) + xten_nn.output %473 : tensor<1x200x12x20xbf16> loc(#loc99) + } -> tensor<1x200x12x20xbf16> loc(#loc99) + xten_nn.output %472 : tensor<1x200x12x20xbf16> loc(#loc99) + } -> tensor<1x200x12x20xbf16> loc(#loc99) + %245 = xten_nn.subgraph (%arg5 = %244: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "537", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "537", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_129", + OutputName = "Div_129", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc100) + xten_nn.output %474 : tensor<1x200x12x20xbf16> loc(#loc100) + } -> tensor<1x200x12x20xbf16> loc(#loc100) + xten_nn.output %472 : tensor<1x200x12x20xbf16> loc(#loc100) + } -> tensor<1x200x12x20xbf16> loc(#loc100) + %246 = xten_nn.subgraph (%arg5 = %242: tensor<1x200x12x20xbf16>, %arg6 = %245: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_130", + OutputName = "Mul_130", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc101) + xten_nn.output %473 : tensor<1x200x12x20xbf16> loc(#loc101) + } -> tensor<1x200x12x20xbf16> loc(#loc101) + xten_nn.output %472 : tensor<1x200x12x20xbf16> loc(#loc101) + } -> tensor<1x200x12x20xbf16> loc(#loc101) + %247 = xten_nn.subgraph (%arg5 = %246: tensor<1x200x12x20xbf16>, %arg6 = %101: tensor<80x200x1x1xbf16>, %arg7 = %100: tensor<80xbf16>, %arg8 = %236: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_131", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Name = "994", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Name = "538", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x200x12x20xbf16>, %arg10 = %arg6: tensor<80x200x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_131", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Name = "994", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Name = "538", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_131", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "997", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc102) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x200x1x1xbf16>) -> tensor<80x1x1x200xbf16> loc(#loc102) + %477 = tosa.transpose %arg9, %475 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc102) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_131", + PartOfOutputName = "Conv_131", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<80x1x1x200xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc102) + %479 = tosa.transpose %478, %474 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc102) + xten_nn.output %479 : tensor<1x80x12x20xbf16> loc(#loc102) + } -> tensor<1x80x12x20xbf16> loc(#loc102) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_132", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "997", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.add %arg9, %arg10 {LayerName = "Add_132", OutputName = "Add_132"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc103) + xten_nn.output %474 : tensor<1x80x12x20xbf16> loc(#loc103) + } -> tensor<1x80x12x20xbf16> loc(#loc103) + xten_nn.output %473 : tensor<1x80x12x20xbf16> loc(#loc103) + } -> tensor<1x80x12x20xbf16> loc(#loc340) + %248 = xten_nn.subgraph (%arg5 = %247: tensor<1x80x12x20xbf16>, %arg6 = %99: tensor<184x80x1x1xbf16>, %arg7 = %98: tensor<184xbf16>) attributes { + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "997", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "541", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "997", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "541", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc104) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc104) + %476 = tosa.transpose %arg8, %474 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc104) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_133", + PartOfOutputName = "Conv_133", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc104) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %478 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + %249 = xten_nn.subgraph (%arg5 = %248: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_135", OutputName = "Add_135"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %474 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + %250 = xten_nn.subgraph (%arg5 = %249: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_138", + OutputName = "Clip_138", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %473 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + %251 = xten_nn.subgraph (%arg5 = %250: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "550", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "550", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_140", + OutputName = "Div_140", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc107) + xten_nn.output %474 : tensor<1x184x12x20xbf16> loc(#loc107) + } -> tensor<1x184x12x20xbf16> loc(#loc107) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc107) + } -> tensor<1x184x12x20xbf16> loc(#loc107) + %252 = xten_nn.subgraph (%arg5 = %248: tensor<1x184x12x20xbf16>, %arg6 = %251: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_141", + OutputName = "Mul_141", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc108) + xten_nn.output %473 : tensor<1x184x12x20xbf16> loc(#loc108) + } -> tensor<1x184x12x20xbf16> loc(#loc108) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc108) + } -> tensor<1x184x12x20xbf16> loc(#loc108) + %253 = xten_nn.subgraph (%arg5 = %252: tensor<1x184x12x20xbf16>, %arg6 = %97: tensor<184x1x3x3xbf16>, %arg7 = %96: tensor<184xbf16>) attributes { + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "551", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "551", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc109) + %476 = tosa.transpose %arg9, %475 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc109) + %477 = tosa.transpose %arg8, %474 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc109) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_142", + PartOfOutputName = "Conv_142", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc109) + %479 = tosa.transpose %478, %473 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %479 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc109) + } -> tensor<1x184x12x20xbf16> loc(#loc109) + %254 = xten_nn.subgraph (%arg5 = %253: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_144", OutputName = "Add_144"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %474 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc110) + } -> tensor<1x184x12x20xbf16> loc(#loc110) + %255 = xten_nn.subgraph (%arg5 = %254: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_147", + OutputName = "Clip_147", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %473 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + %256 = xten_nn.subgraph (%arg5 = %255: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "560", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "560", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_149", + OutputName = "Div_149", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %474 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + %257 = xten_nn.subgraph (%arg5 = %253: tensor<1x184x12x20xbf16>, %arg6 = %256: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_150", + OutputName = "Mul_150", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %473 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + %258 = xten_nn.subgraph (%arg5 = %257: tensor<1x184x12x20xbf16>, %arg6 = %95: tensor<80x184x1x1xbf16>, %arg7 = %94: tensor<80xbf16>, %arg8 = %247: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_151", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1003", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "561", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_151", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1003", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "561", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_151", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1006", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc114) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc114) + %477 = tosa.transpose %arg9, %475 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc114) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_151", + PartOfOutputName = "Conv_151", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc114) + %479 = tosa.transpose %478, %474 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc114) + xten_nn.output %479 : tensor<1x80x12x20xbf16> loc(#loc114) + } -> tensor<1x80x12x20xbf16> loc(#loc114) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_152", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1006", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.add %arg9, %arg10 {LayerName = "Add_152", OutputName = "Add_152"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc115) + xten_nn.output %474 : tensor<1x80x12x20xbf16> loc(#loc115) + } -> tensor<1x80x12x20xbf16> loc(#loc115) + xten_nn.output %473 : tensor<1x80x12x20xbf16> loc(#loc115) + } -> tensor<1x80x12x20xbf16> loc(#loc341) + %259 = xten_nn.subgraph (%arg5 = %258: tensor<1x80x12x20xbf16>, %arg6 = %93: tensor<184x80x1x1xbf16>, %arg7 = %92: tensor<184xbf16>) attributes { + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1006", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "564", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1006", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "564", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc116) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc116) + %476 = tosa.transpose %arg8, %474 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc116) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_153", + PartOfOutputName = "Conv_153", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc116) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %478 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + %260 = xten_nn.subgraph (%arg5 = %259: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_155", OutputName = "Add_155"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %474 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + %261 = xten_nn.subgraph (%arg5 = %260: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_158", + OutputName = "Clip_158", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %473 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + %262 = xten_nn.subgraph (%arg5 = %261: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "573", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "573", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_160", + OutputName = "Div_160", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc119) + xten_nn.output %474 : tensor<1x184x12x20xbf16> loc(#loc119) + } -> tensor<1x184x12x20xbf16> loc(#loc119) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc119) + } -> tensor<1x184x12x20xbf16> loc(#loc119) + %263 = xten_nn.subgraph (%arg5 = %259: tensor<1x184x12x20xbf16>, %arg6 = %262: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_161", + OutputName = "Mul_161", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc120) + xten_nn.output %473 : tensor<1x184x12x20xbf16> loc(#loc120) + } -> tensor<1x184x12x20xbf16> loc(#loc120) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc120) + } -> tensor<1x184x12x20xbf16> loc(#loc120) + %264 = xten_nn.subgraph (%arg5 = %263: tensor<1x184x12x20xbf16>, %arg6 = %91: tensor<184x1x3x3xbf16>, %arg7 = %90: tensor<184xbf16>) attributes { + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "574", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "574", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc121) + %476 = tosa.transpose %arg9, %475 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc121) + %477 = tosa.transpose %arg8, %474 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc121) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_162", + PartOfOutputName = "Conv_162", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc121) + %479 = tosa.transpose %478, %473 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc121) + xten_nn.output %479 : tensor<1x184x12x20xbf16> loc(#loc121) + } -> tensor<1x184x12x20xbf16> loc(#loc121) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc121) + } -> tensor<1x184x12x20xbf16> loc(#loc121) + %265 = xten_nn.subgraph (%arg5 = %264: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_164", OutputName = "Add_164"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc122) + xten_nn.output %474 : tensor<1x184x12x20xbf16> loc(#loc122) + } -> tensor<1x184x12x20xbf16> loc(#loc122) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc122) + } -> tensor<1x184x12x20xbf16> loc(#loc122) + %266 = xten_nn.subgraph (%arg5 = %265: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_167", + OutputName = "Clip_167", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc123) + xten_nn.output %473 : tensor<1x184x12x20xbf16> loc(#loc123) + } -> tensor<1x184x12x20xbf16> loc(#loc123) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc123) + } -> tensor<1x184x12x20xbf16> loc(#loc123) + %267 = xten_nn.subgraph (%arg5 = %266: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "583", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "583", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_169", + OutputName = "Div_169", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc124) + xten_nn.output %474 : tensor<1x184x12x20xbf16> loc(#loc124) + } -> tensor<1x184x12x20xbf16> loc(#loc124) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc124) + } -> tensor<1x184x12x20xbf16> loc(#loc124) + %268 = xten_nn.subgraph (%arg5 = %264: tensor<1x184x12x20xbf16>, %arg6 = %267: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_170", + OutputName = "Mul_170", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc125) + xten_nn.output %473 : tensor<1x184x12x20xbf16> loc(#loc125) + } -> tensor<1x184x12x20xbf16> loc(#loc125) + xten_nn.output %472 : tensor<1x184x12x20xbf16> loc(#loc125) + } -> tensor<1x184x12x20xbf16> loc(#loc125) + %269 = xten_nn.subgraph (%arg5 = %268: tensor<1x184x12x20xbf16>, %arg6 = %89: tensor<80x184x1x1xbf16>, %arg7 = %88: tensor<80xbf16>, %arg8 = %258: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_171", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1012", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "584", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_171", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1012", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "584", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_171", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1015", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc126) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc126) + %477 = tosa.transpose %arg9, %475 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc126) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_171", + PartOfOutputName = "Conv_171", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc126) + %479 = tosa.transpose %478, %474 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc126) + xten_nn.output %479 : tensor<1x80x12x20xbf16> loc(#loc126) + } -> tensor<1x80x12x20xbf16> loc(#loc126) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_172", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1015", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.add %arg9, %arg10 {LayerName = "Add_172", OutputName = "Add_172"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc127) + xten_nn.output %474 : tensor<1x80x12x20xbf16> loc(#loc127) + } -> tensor<1x80x12x20xbf16> loc(#loc127) + xten_nn.output %473 : tensor<1x80x12x20xbf16> loc(#loc127) + } -> tensor<1x80x12x20xbf16> loc(#loc342) + %270 = xten_nn.subgraph (%arg5 = %269: tensor<1x80x12x20xbf16>, %arg6 = %87: tensor<480x80x1x1xbf16>, %arg7 = %86: tensor<480xbf16>) attributes { + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1015", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "587", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<480x80x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1015", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "587", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc128) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x80x1x1xbf16>) -> tensor<480x1x1x80xbf16> loc(#loc128) + %476 = tosa.transpose %arg8, %474 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc128) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_173", + PartOfOutputName = "Conv_173", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<480x1x1x80xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc128) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %478 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + %271 = xten_nn.subgraph (%arg5 = %270: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_175", OutputName = "Add_175"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %474 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + %272 = xten_nn.subgraph (%arg5 = %271: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_178", + OutputName = "Clip_178", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %473 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + %273 = xten_nn.subgraph (%arg5 = %272: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "596", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "596", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_180", + OutputName = "Div_180", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc131) + xten_nn.output %474 : tensor<1x480x12x20xbf16> loc(#loc131) + } -> tensor<1x480x12x20xbf16> loc(#loc131) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc131) + } -> tensor<1x480x12x20xbf16> loc(#loc131) + %274 = xten_nn.subgraph (%arg5 = %270: tensor<1x480x12x20xbf16>, %arg6 = %273: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_181", + OutputName = "Mul_181", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc132) + xten_nn.output %473 : tensor<1x480x12x20xbf16> loc(#loc132) + } -> tensor<1x480x12x20xbf16> loc(#loc132) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc132) + } -> tensor<1x480x12x20xbf16> loc(#loc132) + %275 = xten_nn.subgraph (%arg5 = %274: tensor<1x480x12x20xbf16>, %arg6 = %85: tensor<480x1x3x3xbf16>, %arg7 = %84: tensor<480xbf16>) attributes { + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "597", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<480x1x3x3xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "597", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc133) + %476 = tosa.transpose %arg9, %475 : (tensor<480x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x480x1xbf16> loc(#loc133) + %477 = tosa.transpose %arg8, %474 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc133) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_182", + PartOfOutputName = "Conv_182", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<3x3x480x1xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc133) + %479 = tosa.transpose %478, %473 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc133) + xten_nn.output %479 : tensor<1x480x12x20xbf16> loc(#loc133) + } -> tensor<1x480x12x20xbf16> loc(#loc133) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc133) + } -> tensor<1x480x12x20xbf16> loc(#loc133) + %276 = xten_nn.subgraph (%arg5 = %275: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_184", OutputName = "Add_184"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc134) + xten_nn.output %474 : tensor<1x480x12x20xbf16> loc(#loc134) + } -> tensor<1x480x12x20xbf16> loc(#loc134) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc134) + } -> tensor<1x480x12x20xbf16> loc(#loc134) + %277 = xten_nn.subgraph (%arg5 = %276: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_187", + OutputName = "Clip_187", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc135) + xten_nn.output %473 : tensor<1x480x12x20xbf16> loc(#loc135) + } -> tensor<1x480x12x20xbf16> loc(#loc135) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc135) + } -> tensor<1x480x12x20xbf16> loc(#loc135) + %278 = xten_nn.subgraph (%arg5 = %277: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "606", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "606", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_189", + OutputName = "Div_189", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc136) + xten_nn.output %474 : tensor<1x480x12x20xbf16> loc(#loc136) + } -> tensor<1x480x12x20xbf16> loc(#loc136) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc136) + } -> tensor<1x480x12x20xbf16> loc(#loc136) + %279 = xten_nn.subgraph (%arg5 = %275: tensor<1x480x12x20xbf16>, %arg6 = %278: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_190", + OutputName = "Mul_190", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc137) + xten_nn.output %473 : tensor<1x480x12x20xbf16> loc(#loc137) + } -> tensor<1x480x12x20xbf16> loc(#loc137) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc137) + } -> tensor<1x480x12x20xbf16> loc(#loc137) + %280 = xten_nn.subgraph (%arg5 = %279: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_191_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 60 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %472 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x1x240xbf16> loc(#loc343) + xten_nn.output %472 : tensor<1x480x1x240xbf16> loc(#loc343) + } -> tensor<1x480x1x240xbf16> loc(#loc343) + %281 = xten_nn.subgraph (%arg5 = %280: tensor<1x480x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_191", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_191_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_191", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_191_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 480 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %473 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x480x1x240xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc138) + xten_nn.output %473 : tensor<1x480x1x1xbf16> loc(#loc138) + } -> tensor<1x480x1x1xbf16> loc(#loc138) + xten_nn.output %472 : tensor<1x480x1x1xbf16> loc(#loc138) + } -> tensor<1x480x1x1xbf16> loc(#loc138) + %282 = xten_nn.subgraph (%arg5 = %281: tensor<1x480x1x1xbf16>, %arg6 = %83: tensor<120x480x1x1xbf16>, %arg7 = %82: tensor<120xbf16>) attributes { + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "607", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x1x1xbf16>, %arg9 = %arg6: tensor<120x480x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "607", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x480x1x1xbf16>) -> tensor<120x1x1x480xbf16> loc(#loc344) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc344) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_192", + PartOfOutputName = "Conv_192", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x480xbf16>, tensor<120x1x1x480xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc139) + %476 = tosa.clamp %475 { + LayerName = "Relu_193", + OutputName = "Relu_193", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc140) + %477 = tosa.reshape %476 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc344) + xten_nn.output %477 : tensor<1x120x1x1xbf16> loc(#loc140) + } -> tensor<1x120x1x1xbf16> loc(#loc344) + xten_nn.output %472 : tensor<1x120x1x1xbf16> loc(#loc344) + } -> tensor<1x120x1x1xbf16> loc(#loc344) + %283 = xten_nn.subgraph (%arg5 = %282: tensor<1x120x1x1xbf16>, %arg6 = %81: tensor<480x120x1x1xbf16>, %arg7 = %80: tensor<480xbf16>) attributes { + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "609", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<480x120x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "609", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x120x1x1xbf16>) -> tensor<480x1x1x120xbf16> loc(#loc141) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc141) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_194", + PartOfOutputName = "Conv_194", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<480x1x1x120xbf16>, tensor<480xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc141) + %476 = tosa.reshape %475 {new_shape = array} : (tensor<1x1x1x480xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc141) + xten_nn.output %476 : tensor<1x480x1x1xbf16> loc(#loc141) + } -> tensor<1x480x1x1xbf16> loc(#loc141) + xten_nn.output %472 : tensor<1x480x1x1xbf16> loc(#loc141) + } -> tensor<1x480x1x1xbf16> loc(#loc141) + %284 = xten_nn.subgraph (%arg5 = %283: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_196", OutputName = "Add_196"} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc142) + xten_nn.output %474 : tensor<1x480x1x1xbf16> loc(#loc142) + } -> tensor<1x480x1x1xbf16> loc(#loc142) + xten_nn.output %472 : tensor<1x480x1x1xbf16> loc(#loc142) + } -> tensor<1x480x1x1xbf16> loc(#loc142) + %285 = xten_nn.subgraph (%arg5 = %284: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_199", + OutputName = "Clip_199", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc143) + xten_nn.output %473 : tensor<1x480x1x1xbf16> loc(#loc143) + } -> tensor<1x480x1x1xbf16> loc(#loc143) + xten_nn.output %472 : tensor<1x480x1x1xbf16> loc(#loc143) + } -> tensor<1x480x1x1xbf16> loc(#loc143) + %286 = xten_nn.subgraph (%arg5 = %285: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_201", + OutputName = "Div_201", + shift = 0 : i8} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc144) + xten_nn.output %474 : tensor<1x480x1x1xbf16> loc(#loc144) + } -> tensor<1x480x1x1xbf16> loc(#loc144) + xten_nn.output %472 : tensor<1x480x1x1xbf16> loc(#loc144) + } -> tensor<1x480x1x1xbf16> loc(#loc144) + %287 = xten_nn.subgraph (%arg5 = %286: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Mul_202_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_202_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 480 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %472 = tosa.tile %arg5 {multiples = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc145) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc145) + } -> tensor<1x480x12x20xbf16> loc(#loc145) + %288 = xten_nn.subgraph (%arg5 = %287: tensor<1x480x12x20xbf16>, %arg6 = %279: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_202", + OutputName = "Mul_202", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc145) + xten_nn.output %473 : tensor<1x480x12x20xbf16> loc(#loc145) + } -> tensor<1x480x12x20xbf16> loc(#loc145) + xten_nn.output %472 : tensor<1x480x12x20xbf16> loc(#loc145) + } -> tensor<1x480x12x20xbf16> loc(#loc145) + %289 = xten_nn.subgraph (%arg5 = %288: tensor<1x480x12x20xbf16>, %arg6 = %79: tensor<112x480x1x1xbf16>, %arg7 = %78: tensor<112xbf16>) attributes { + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Name = "618", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "619", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<112x480x1x1xbf16>, %arg10 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Name = "618", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "619", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc146) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<112x480x1x1xbf16>) -> tensor<112x1x1x480xbf16> loc(#loc146) + %476 = tosa.transpose %arg8, %474 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc146) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_203", + PartOfOutputName = "Conv_203", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<112x1x1x480xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc146) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc146) + xten_nn.output %478 : tensor<1x112x12x20xbf16> loc(#loc146) + } -> tensor<1x112x12x20xbf16> loc(#loc146) + xten_nn.output %472 : tensor<1x112x12x20xbf16> loc(#loc146) + } -> tensor<1x112x12x20xbf16> loc(#loc146) + %290 = xten_nn.subgraph (%arg5 = %289: tensor<1x112x12x20xbf16>, %arg6 = %77: tensor<672x112x1x1xbf16>, %arg7 = %76: tensor<672xbf16>) attributes { + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "619", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "1028", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "619", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "1028", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc147) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc147) + %476 = tosa.transpose %arg8, %474 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc147) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_204", + PartOfOutputName = "Conv_204", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc147) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %478 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + %291 = xten_nn.subgraph (%arg5 = %290: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_206", OutputName = "Add_206"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %474 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + %292 = xten_nn.subgraph (%arg5 = %291: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_209", + OutputName = "Clip_209", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %473 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + %293 = xten_nn.subgraph (%arg5 = %292: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "630", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "630", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_211", + OutputName = "Div_211", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc150) + xten_nn.output %474 : tensor<1x672x12x20xbf16> loc(#loc150) + } -> tensor<1x672x12x20xbf16> loc(#loc150) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc150) + } -> tensor<1x672x12x20xbf16> loc(#loc150) + %294 = xten_nn.subgraph (%arg5 = %290: tensor<1x672x12x20xbf16>, %arg6 = %293: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_212", + OutputName = "Mul_212", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc151) + xten_nn.output %473 : tensor<1x672x12x20xbf16> loc(#loc151) + } -> tensor<1x672x12x20xbf16> loc(#loc151) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc151) + } -> tensor<1x672x12x20xbf16> loc(#loc151) + %295 = xten_nn.subgraph (%arg5 = %294: tensor<1x672x12x20xbf16>, %arg6 = %75: tensor<672x1x3x3xbf16>, %arg7 = %74: tensor<672xbf16>) attributes { + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "631", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x3x3xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "631", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc152) + %476 = tosa.transpose %arg9, %475 : (tensor<672x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x672x1xbf16> loc(#loc152) + %477 = tosa.transpose %arg8, %474 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc152) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_213", + PartOfOutputName = "Conv_213", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<3x3x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc152) + %479 = tosa.transpose %478, %473 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc152) + xten_nn.output %479 : tensor<1x672x12x20xbf16> loc(#loc152) + } -> tensor<1x672x12x20xbf16> loc(#loc152) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc152) + } -> tensor<1x672x12x20xbf16> loc(#loc152) + %296 = xten_nn.subgraph (%arg5 = %295: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_215", OutputName = "Add_215"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc153) + xten_nn.output %474 : tensor<1x672x12x20xbf16> loc(#loc153) + } -> tensor<1x672x12x20xbf16> loc(#loc153) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc153) + } -> tensor<1x672x12x20xbf16> loc(#loc153) + %297 = xten_nn.subgraph (%arg5 = %296: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_218", + OutputName = "Clip_218", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc154) + xten_nn.output %473 : tensor<1x672x12x20xbf16> loc(#loc154) + } -> tensor<1x672x12x20xbf16> loc(#loc154) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc154) + } -> tensor<1x672x12x20xbf16> loc(#loc154) + %298 = xten_nn.subgraph (%arg5 = %297: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "640", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "640", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_220", + OutputName = "Div_220", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc155) + xten_nn.output %474 : tensor<1x672x12x20xbf16> loc(#loc155) + } -> tensor<1x672x12x20xbf16> loc(#loc155) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc155) + } -> tensor<1x672x12x20xbf16> loc(#loc155) + %299 = xten_nn.subgraph (%arg5 = %295: tensor<1x672x12x20xbf16>, %arg6 = %298: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_221", + OutputName = "Mul_221", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc156) + xten_nn.output %473 : tensor<1x672x12x20xbf16> loc(#loc156) + } -> tensor<1x672x12x20xbf16> loc(#loc156) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc156) + } -> tensor<1x672x12x20xbf16> loc(#loc156) + %300 = xten_nn.subgraph (%arg5 = %299: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_222_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %472 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc345) + xten_nn.output %472 : tensor<1x672x1x240xbf16> loc(#loc345) + } -> tensor<1x672x1x240xbf16> loc(#loc345) + %301 = xten_nn.subgraph (%arg5 = %300: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_222", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_222_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_222", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_222_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %473 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc157) + xten_nn.output %473 : tensor<1x672x1x1xbf16> loc(#loc157) + } -> tensor<1x672x1x1xbf16> loc(#loc157) + xten_nn.output %472 : tensor<1x672x1x1xbf16> loc(#loc157) + } -> tensor<1x672x1x1xbf16> loc(#loc157) + %302 = xten_nn.subgraph (%arg5 = %301: tensor<1x672x1x1xbf16>, %arg6 = %73: tensor<168x672x1x1xbf16>, %arg7 = %72: tensor<168xbf16>) attributes { + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "641", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "641", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc346) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc346) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_223", + PartOfOutputName = "Conv_223", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc158) + %476 = tosa.clamp %475 { + LayerName = "Relu_224", + OutputName = "Relu_224", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc159) + %477 = tosa.reshape %476 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc346) + xten_nn.output %477 : tensor<1x168x1x1xbf16> loc(#loc159) + } -> tensor<1x168x1x1xbf16> loc(#loc346) + xten_nn.output %472 : tensor<1x168x1x1xbf16> loc(#loc346) + } -> tensor<1x168x1x1xbf16> loc(#loc346) + %303 = xten_nn.subgraph (%arg5 = %302: tensor<1x168x1x1xbf16>, %arg6 = %71: tensor<672x168x1x1xbf16>, %arg7 = %70: tensor<672xbf16>) attributes { + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "643", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "643", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc160) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc160) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_225", + PartOfOutputName = "Conv_225", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc160) + %476 = tosa.reshape %475 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc160) + xten_nn.output %476 : tensor<1x672x1x1xbf16> loc(#loc160) + } -> tensor<1x672x1x1xbf16> loc(#loc160) + xten_nn.output %472 : tensor<1x672x1x1xbf16> loc(#loc160) + } -> tensor<1x672x1x1xbf16> loc(#loc160) + %304 = xten_nn.subgraph (%arg5 = %303: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_227", OutputName = "Add_227"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc161) + xten_nn.output %474 : tensor<1x672x1x1xbf16> loc(#loc161) + } -> tensor<1x672x1x1xbf16> loc(#loc161) + xten_nn.output %472 : tensor<1x672x1x1xbf16> loc(#loc161) + } -> tensor<1x672x1x1xbf16> loc(#loc161) + %305 = xten_nn.subgraph (%arg5 = %304: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_230", + OutputName = "Clip_230", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc162) + xten_nn.output %473 : tensor<1x672x1x1xbf16> loc(#loc162) + } -> tensor<1x672x1x1xbf16> loc(#loc162) + xten_nn.output %472 : tensor<1x672x1x1xbf16> loc(#loc162) + } -> tensor<1x672x1x1xbf16> loc(#loc162) + %306 = xten_nn.subgraph (%arg5 = %305: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_232", + OutputName = "Div_232", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc163) + xten_nn.output %474 : tensor<1x672x1x1xbf16> loc(#loc163) + } -> tensor<1x672x1x1xbf16> loc(#loc163) + xten_nn.output %472 : tensor<1x672x1x1xbf16> loc(#loc163) + } -> tensor<1x672x1x1xbf16> loc(#loc163) + %307 = xten_nn.subgraph (%arg5 = %306: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Mul_233_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_233_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %472 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + %308 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>, %arg6 = %299: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_233", + OutputName = "Mul_233", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %473 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + %309 = xten_nn.subgraph (%arg5 = %308: tensor<1x672x12x20xbf16>, %arg6 = %69: tensor<112x672x1x1xbf16>, %arg7 = %68: tensor<112xbf16>, %arg8 = %289: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Conv_234", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "652", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "653", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x672x12x20xbf16>, %arg10 = %arg6: tensor<112x672x1x1xbf16>, %arg11 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_234", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "652", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "653", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_234", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1033", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc165) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<112x672x1x1xbf16>) -> tensor<112x1x1x672xbf16> loc(#loc165) + %477 = tosa.transpose %arg9, %475 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc165) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_234", + PartOfOutputName = "Conv_234", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<112x1x1x672xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc165) + %479 = tosa.transpose %478, %474 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc165) + xten_nn.output %479 : tensor<1x112x12x20xbf16> loc(#loc165) + } -> tensor<1x112x12x20xbf16> loc(#loc165) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x112x12x20xbf16>, %arg10 = %arg8: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Add_235", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1033", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.add %arg9, %arg10 {LayerName = "Add_235", OutputName = "Add_235"} : (tensor<1x112x12x20xbf16>, tensor<1x112x12x20xbf16>) -> tensor<1x112x12x20xbf16> loc(#loc166) + xten_nn.output %474 : tensor<1x112x12x20xbf16> loc(#loc166) + } -> tensor<1x112x12x20xbf16> loc(#loc166) + xten_nn.output %473 : tensor<1x112x12x20xbf16> loc(#loc166) + } -> tensor<1x112x12x20xbf16> loc(#loc347) + %310 = xten_nn.subgraph (%arg5 = %309: tensor<1x112x12x20xbf16>, %arg6 = %67: tensor<672x112x1x1xbf16>, %arg7 = %66: tensor<672xbf16>) attributes { + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "1033", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "656", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "1033", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "656", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc167) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc167) + %476 = tosa.transpose %arg8, %474 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc167) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_236", + PartOfOutputName = "Conv_236", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc167) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %478 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + %311 = xten_nn.subgraph (%arg5 = %310: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_238", OutputName = "Add_238"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %474 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + %312 = xten_nn.subgraph (%arg5 = %311: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_241", + OutputName = "Clip_241", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %473 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + %313 = xten_nn.subgraph (%arg5 = %312: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "665", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "665", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_243", + OutputName = "Div_243", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc170) + xten_nn.output %474 : tensor<1x672x12x20xbf16> loc(#loc170) + } -> tensor<1x672x12x20xbf16> loc(#loc170) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc170) + } -> tensor<1x672x12x20xbf16> loc(#loc170) + %314 = xten_nn.subgraph (%arg5 = %310: tensor<1x672x12x20xbf16>, %arg6 = %313: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_244", + OutputName = "Mul_244", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc171) + xten_nn.output %473 : tensor<1x672x12x20xbf16> loc(#loc171) + } -> tensor<1x672x12x20xbf16> loc(#loc171) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc171) + } -> tensor<1x672x12x20xbf16> loc(#loc171) + %315 = xten_nn.subgraph (%arg5 = %314: tensor<1x672x12x20xbf16>, %arg6 = %65: tensor<672x1x9x9xbf16>, %arg7 = %64: tensor<672xbf16>) attributes { + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "666", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x9x9xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "666", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc172) + %476 = tosa.transpose %arg9, %475 : (tensor<672x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x672x1xbf16> loc(#loc172) + %477 = tosa.transpose %arg8, %474 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc172) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_245", + PartOfOutputName = "Conv_245", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<9x9x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc172) + %479 = tosa.transpose %478, %473 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc172) + xten_nn.output %479 : tensor<1x672x12x20xbf16> loc(#loc172) + } -> tensor<1x672x12x20xbf16> loc(#loc172) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc172) + } -> tensor<1x672x12x20xbf16> loc(#loc172) + %316 = xten_nn.subgraph (%arg5 = %315: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_247", OutputName = "Add_247"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc173) + xten_nn.output %474 : tensor<1x672x12x20xbf16> loc(#loc173) + } -> tensor<1x672x12x20xbf16> loc(#loc173) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc173) + } -> tensor<1x672x12x20xbf16> loc(#loc173) + %317 = xten_nn.subgraph (%arg5 = %316: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_250", + OutputName = "Clip_250", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc174) + xten_nn.output %473 : tensor<1x672x12x20xbf16> loc(#loc174) + } -> tensor<1x672x12x20xbf16> loc(#loc174) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc174) + } -> tensor<1x672x12x20xbf16> loc(#loc174) + %318 = xten_nn.subgraph (%arg5 = %317: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "675", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "675", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_252", + OutputName = "Div_252", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc175) + xten_nn.output %474 : tensor<1x672x12x20xbf16> loc(#loc175) + } -> tensor<1x672x12x20xbf16> loc(#loc175) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc175) + } -> tensor<1x672x12x20xbf16> loc(#loc175) + %319 = xten_nn.subgraph (%arg5 = %315: tensor<1x672x12x20xbf16>, %arg6 = %318: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_253", + OutputName = "Mul_253", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc176) + xten_nn.output %473 : tensor<1x672x12x20xbf16> loc(#loc176) + } -> tensor<1x672x12x20xbf16> loc(#loc176) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc176) + } -> tensor<1x672x12x20xbf16> loc(#loc176) + %320 = xten_nn.subgraph (%arg5 = %319: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_254_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %472 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc348) + xten_nn.output %472 : tensor<1x672x1x240xbf16> loc(#loc348) + } -> tensor<1x672x1x240xbf16> loc(#loc348) + %321 = xten_nn.subgraph (%arg5 = %320: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_254", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_254_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_254", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_254_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %473 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc177) + xten_nn.output %473 : tensor<1x672x1x1xbf16> loc(#loc177) + } -> tensor<1x672x1x1xbf16> loc(#loc177) + xten_nn.output %472 : tensor<1x672x1x1xbf16> loc(#loc177) + } -> tensor<1x672x1x1xbf16> loc(#loc177) + %322 = xten_nn.subgraph (%arg5 = %321: tensor<1x672x1x1xbf16>, %arg6 = %63: tensor<168x672x1x1xbf16>, %arg7 = %62: tensor<168xbf16>) attributes { + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "676", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "676", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc349) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc349) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_255", + PartOfOutputName = "Conv_255", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc178) + %476 = tosa.clamp %475 { + LayerName = "Relu_256", + OutputName = "Relu_256", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc179) + %477 = tosa.reshape %476 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc349) + xten_nn.output %477 : tensor<1x168x1x1xbf16> loc(#loc179) + } -> tensor<1x168x1x1xbf16> loc(#loc349) + xten_nn.output %472 : tensor<1x168x1x1xbf16> loc(#loc349) + } -> tensor<1x168x1x1xbf16> loc(#loc349) + %323 = xten_nn.subgraph (%arg5 = %322: tensor<1x168x1x1xbf16>, %arg6 = %61: tensor<672x168x1x1xbf16>, %arg7 = %60: tensor<672xbf16>) attributes { + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "678", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "678", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc180) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc180) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_257", + PartOfOutputName = "Conv_257", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc180) + %476 = tosa.reshape %475 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc180) + xten_nn.output %476 : tensor<1x672x1x1xbf16> loc(#loc180) + } -> tensor<1x672x1x1xbf16> loc(#loc180) + xten_nn.output %472 : tensor<1x672x1x1xbf16> loc(#loc180) + } -> tensor<1x672x1x1xbf16> loc(#loc180) + %324 = xten_nn.subgraph (%arg5 = %323: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_259", OutputName = "Add_259"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc181) + xten_nn.output %474 : tensor<1x672x1x1xbf16> loc(#loc181) + } -> tensor<1x672x1x1xbf16> loc(#loc181) + xten_nn.output %472 : tensor<1x672x1x1xbf16> loc(#loc181) + } -> tensor<1x672x1x1xbf16> loc(#loc181) + %325 = xten_nn.subgraph (%arg5 = %324: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_262", + OutputName = "Clip_262", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc182) + xten_nn.output %473 : tensor<1x672x1x1xbf16> loc(#loc182) + } -> tensor<1x672x1x1xbf16> loc(#loc182) + xten_nn.output %472 : tensor<1x672x1x1xbf16> loc(#loc182) + } -> tensor<1x672x1x1xbf16> loc(#loc182) + %326 = xten_nn.subgraph (%arg5 = %325: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_264", + OutputName = "Div_264", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc183) + xten_nn.output %474 : tensor<1x672x1x1xbf16> loc(#loc183) + } -> tensor<1x672x1x1xbf16> loc(#loc183) + xten_nn.output %472 : tensor<1x672x1x1xbf16> loc(#loc183) + } -> tensor<1x672x1x1xbf16> loc(#loc183) + %327 = xten_nn.subgraph (%arg5 = %326: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Mul_265_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_265_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %472 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc184) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc184) + } -> tensor<1x672x12x20xbf16> loc(#loc184) + %328 = xten_nn.subgraph (%arg5 = %327: tensor<1x672x12x20xbf16>, %arg6 = %319: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_265", + OutputName = "Mul_265", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc184) + xten_nn.output %473 : tensor<1x672x12x20xbf16> loc(#loc184) + } -> tensor<1x672x12x20xbf16> loc(#loc184) + xten_nn.output %472 : tensor<1x672x12x20xbf16> loc(#loc184) + } -> tensor<1x672x12x20xbf16> loc(#loc184) + %329 = xten_nn.subgraph (%arg5 = %328: tensor<1x672x12x20xbf16>, %arg6 = %59: tensor<160x672x1x1xbf16>, %arg7 = %58: tensor<160xbf16>) attributes { + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "687", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "688", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<160x672x1x1xbf16>, %arg10 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "687", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "688", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc185) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<160x672x1x1xbf16>) -> tensor<160x1x1x672xbf16> loc(#loc185) + %476 = tosa.transpose %arg8, %474 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc185) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_266", + PartOfOutputName = "Conv_266", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<160x1x1x672xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc185) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc185) + xten_nn.output %478 : tensor<1x160x12x20xbf16> loc(#loc185) + } -> tensor<1x160x12x20xbf16> loc(#loc185) + xten_nn.output %472 : tensor<1x160x12x20xbf16> loc(#loc185) + } -> tensor<1x160x12x20xbf16> loc(#loc185) + %330 = xten_nn.subgraph (%arg5 = %329: tensor<1x160x12x20xbf16>, %arg6 = %57: tensor<960x160x1x1xbf16>, %arg7 = %56: tensor<960xbf16>) attributes { + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "688", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "1046", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "688", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "1046", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc186) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc186) + %476 = tosa.transpose %arg8, %474 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc186) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_267", + PartOfOutputName = "Conv_267", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc186) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %478 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + %331 = xten_nn.subgraph (%arg5 = %330: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_269", OutputName = "Add_269"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %474 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + %332 = xten_nn.subgraph (%arg5 = %331: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_272", + OutputName = "Clip_272", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + %333 = xten_nn.subgraph (%arg5 = %332: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "699", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "699", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_274", + OutputName = "Div_274", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc189) + xten_nn.output %474 : tensor<1x960x12x20xbf16> loc(#loc189) + } -> tensor<1x960x12x20xbf16> loc(#loc189) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc189) + } -> tensor<1x960x12x20xbf16> loc(#loc189) + %334 = xten_nn.subgraph (%arg5 = %330: tensor<1x960x12x20xbf16>, %arg6 = %333: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_275", + OutputName = "Mul_275", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc190) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc190) + } -> tensor<1x960x12x20xbf16> loc(#loc190) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc190) + } -> tensor<1x960x12x20xbf16> loc(#loc190) + %335 = xten_nn.subgraph (%arg5 = %334: tensor<1x960x12x20xbf16>, %arg6 = %55: tensor<960x1x9x9xbf16>, %arg7 = %54: tensor<960xbf16>) attributes { + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "700", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "700", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc191) + %476 = tosa.transpose %arg9, %475 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc191) + %477 = tosa.transpose %arg8, %474 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc191) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_276", + PartOfOutputName = "Conv_276", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc191) + %479 = tosa.transpose %478, %473 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc191) + xten_nn.output %479 : tensor<1x960x12x20xbf16> loc(#loc191) + } -> tensor<1x960x12x20xbf16> loc(#loc191) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc191) + } -> tensor<1x960x12x20xbf16> loc(#loc191) + %336 = xten_nn.subgraph (%arg5 = %335: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_278", OutputName = "Add_278"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc192) + xten_nn.output %474 : tensor<1x960x12x20xbf16> loc(#loc192) + } -> tensor<1x960x12x20xbf16> loc(#loc192) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc192) + } -> tensor<1x960x12x20xbf16> loc(#loc192) + %337 = xten_nn.subgraph (%arg5 = %336: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_281", + OutputName = "Clip_281", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc193) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc193) + } -> tensor<1x960x12x20xbf16> loc(#loc193) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc193) + } -> tensor<1x960x12x20xbf16> loc(#loc193) + %338 = xten_nn.subgraph (%arg5 = %337: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "709", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "709", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_283", + OutputName = "Div_283", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc194) + xten_nn.output %474 : tensor<1x960x12x20xbf16> loc(#loc194) + } -> tensor<1x960x12x20xbf16> loc(#loc194) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc194) + } -> tensor<1x960x12x20xbf16> loc(#loc194) + %339 = xten_nn.subgraph (%arg5 = %335: tensor<1x960x12x20xbf16>, %arg6 = %338: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_284", + OutputName = "Mul_284", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc195) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc195) + } -> tensor<1x960x12x20xbf16> loc(#loc195) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc195) + } -> tensor<1x960x12x20xbf16> loc(#loc195) + %340 = xten_nn.subgraph (%arg5 = %339: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_285_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %472 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc350) + xten_nn.output %472 : tensor<1x960x1x240xbf16> loc(#loc350) + } -> tensor<1x960x1x240xbf16> loc(#loc350) + %341 = xten_nn.subgraph (%arg5 = %340: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_285", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_285_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_285", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_285_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %473 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc196) + xten_nn.output %473 : tensor<1x960x1x1xbf16> loc(#loc196) + } -> tensor<1x960x1x1xbf16> loc(#loc196) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc196) + } -> tensor<1x960x1x1xbf16> loc(#loc196) + %342 = xten_nn.subgraph (%arg5 = %341: tensor<1x960x1x1xbf16>, %arg6 = %53: tensor<240x960x1x1xbf16>, %arg7 = %52: tensor<240xbf16>) attributes { + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "710", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "710", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc351) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc351) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_286", + PartOfOutputName = "Conv_286", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc197) + %476 = tosa.clamp %475 { + LayerName = "Relu_287", + OutputName = "Relu_287", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc198) + %477 = tosa.reshape %476 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc351) + xten_nn.output %477 : tensor<1x240x1x1xbf16> loc(#loc198) + } -> tensor<1x240x1x1xbf16> loc(#loc351) + xten_nn.output %472 : tensor<1x240x1x1xbf16> loc(#loc351) + } -> tensor<1x240x1x1xbf16> loc(#loc351) + %343 = xten_nn.subgraph (%arg5 = %342: tensor<1x240x1x1xbf16>, %arg6 = %51: tensor<960x240x1x1xbf16>, %arg7 = %50: tensor<960xbf16>) attributes { + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "712", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "712", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc199) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc199) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_288", + PartOfOutputName = "Conv_288", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc199) + %476 = tosa.reshape %475 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc199) + xten_nn.output %476 : tensor<1x960x1x1xbf16> loc(#loc199) + } -> tensor<1x960x1x1xbf16> loc(#loc199) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc199) + } -> tensor<1x960x1x1xbf16> loc(#loc199) + %344 = xten_nn.subgraph (%arg5 = %343: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_290", OutputName = "Add_290"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc200) + xten_nn.output %474 : tensor<1x960x1x1xbf16> loc(#loc200) + } -> tensor<1x960x1x1xbf16> loc(#loc200) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc200) + } -> tensor<1x960x1x1xbf16> loc(#loc200) + %345 = xten_nn.subgraph (%arg5 = %344: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_293", + OutputName = "Clip_293", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc201) + xten_nn.output %473 : tensor<1x960x1x1xbf16> loc(#loc201) + } -> tensor<1x960x1x1xbf16> loc(#loc201) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc201) + } -> tensor<1x960x1x1xbf16> loc(#loc201) + %346 = xten_nn.subgraph (%arg5 = %345: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_295", + OutputName = "Div_295", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc202) + xten_nn.output %474 : tensor<1x960x1x1xbf16> loc(#loc202) + } -> tensor<1x960x1x1xbf16> loc(#loc202) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc202) + } -> tensor<1x960x1x1xbf16> loc(#loc202) + %347 = xten_nn.subgraph (%arg5 = %346: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Mul_296_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_296_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %472 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + %348 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>, %arg6 = %339: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_296", + OutputName = "Mul_296", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + %349 = xten_nn.subgraph (%arg5 = %348: tensor<1x960x12x20xbf16>, %arg6 = %49: tensor<160x960x1x1xbf16>, %arg7 = %48: tensor<160xbf16>, %arg8 = %329: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_297", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "721", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "722", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_297", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "721", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "722", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_297", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1051", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc204) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc204) + %477 = tosa.transpose %arg9, %475 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc204) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_297", + PartOfOutputName = "Conv_297", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc204) + %479 = tosa.transpose %478, %474 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc204) + xten_nn.output %479 : tensor<1x160x12x20xbf16> loc(#loc204) + } -> tensor<1x160x12x20xbf16> loc(#loc204) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_298", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1051", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.add %arg9, %arg10 {LayerName = "Add_298", OutputName = "Add_298"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc205) + xten_nn.output %474 : tensor<1x160x12x20xbf16> loc(#loc205) + } -> tensor<1x160x12x20xbf16> loc(#loc205) + xten_nn.output %473 : tensor<1x160x12x20xbf16> loc(#loc205) + } -> tensor<1x160x12x20xbf16> loc(#loc352) + %350 = xten_nn.subgraph (%arg5 = %349: tensor<1x160x12x20xbf16>, %arg6 = %47: tensor<960x160x1x1xbf16>, %arg7 = %46: tensor<960xbf16>) attributes { + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1051", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "725", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1051", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "725", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc206) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc206) + %476 = tosa.transpose %arg8, %474 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc206) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_299", + PartOfOutputName = "Conv_299", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc206) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %478 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + %351 = xten_nn.subgraph (%arg5 = %350: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_301", OutputName = "Add_301"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %474 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + %352 = xten_nn.subgraph (%arg5 = %351: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_304", + OutputName = "Clip_304", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + %353 = xten_nn.subgraph (%arg5 = %352: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "734", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "734", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_306", + OutputName = "Div_306", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc209) + xten_nn.output %474 : tensor<1x960x12x20xbf16> loc(#loc209) + } -> tensor<1x960x12x20xbf16> loc(#loc209) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc209) + } -> tensor<1x960x12x20xbf16> loc(#loc209) + %354 = xten_nn.subgraph (%arg5 = %350: tensor<1x960x12x20xbf16>, %arg6 = %353: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_307", + OutputName = "Mul_307", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc210) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc210) + } -> tensor<1x960x12x20xbf16> loc(#loc210) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc210) + } -> tensor<1x960x12x20xbf16> loc(#loc210) + %355 = xten_nn.subgraph (%arg5 = %354: tensor<1x960x12x20xbf16>, %arg6 = %45: tensor<960x1x9x9xbf16>, %arg7 = %44: tensor<960xbf16>) attributes { + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "735", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "735", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc211) + %476 = tosa.transpose %arg9, %475 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc211) + %477 = tosa.transpose %arg8, %474 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc211) + %478 = tosa.depthwise_conv2d %477, %476, %arg10 { + PartOfLayerName = "Conv_308", + PartOfOutputName = "Conv_308", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc211) + %479 = tosa.transpose %478, %473 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc211) + xten_nn.output %479 : tensor<1x960x12x20xbf16> loc(#loc211) + } -> tensor<1x960x12x20xbf16> loc(#loc211) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc211) + } -> tensor<1x960x12x20xbf16> loc(#loc211) + %356 = xten_nn.subgraph (%arg5 = %355: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_310", OutputName = "Add_310"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc212) + xten_nn.output %474 : tensor<1x960x12x20xbf16> loc(#loc212) + } -> tensor<1x960x12x20xbf16> loc(#loc212) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc212) + } -> tensor<1x960x12x20xbf16> loc(#loc212) + %357 = xten_nn.subgraph (%arg5 = %356: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_313", + OutputName = "Clip_313", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc213) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc213) + } -> tensor<1x960x12x20xbf16> loc(#loc213) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc213) + } -> tensor<1x960x12x20xbf16> loc(#loc213) + %358 = xten_nn.subgraph (%arg5 = %357: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "744", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "744", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_315", + OutputName = "Div_315", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc214) + xten_nn.output %474 : tensor<1x960x12x20xbf16> loc(#loc214) + } -> tensor<1x960x12x20xbf16> loc(#loc214) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc214) + } -> tensor<1x960x12x20xbf16> loc(#loc214) + %359 = xten_nn.subgraph (%arg5 = %355: tensor<1x960x12x20xbf16>, %arg6 = %358: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_316", + OutputName = "Mul_316", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc215) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc215) + } -> tensor<1x960x12x20xbf16> loc(#loc215) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc215) + } -> tensor<1x960x12x20xbf16> loc(#loc215) + %360 = xten_nn.subgraph (%arg5 = %359: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_317_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %472 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc353) + xten_nn.output %472 : tensor<1x960x1x240xbf16> loc(#loc353) + } -> tensor<1x960x1x240xbf16> loc(#loc353) + %361 = xten_nn.subgraph (%arg5 = %360: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_317", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_317_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_317", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_317_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %473 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc216) + xten_nn.output %473 : tensor<1x960x1x1xbf16> loc(#loc216) + } -> tensor<1x960x1x1xbf16> loc(#loc216) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc216) + } -> tensor<1x960x1x1xbf16> loc(#loc216) + %362 = xten_nn.subgraph (%arg5 = %361: tensor<1x960x1x1xbf16>, %arg6 = %43: tensor<240x960x1x1xbf16>, %arg7 = %42: tensor<240xbf16>) attributes { + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "745", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "745", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc354) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc354) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_318", + PartOfOutputName = "Conv_318", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc217) + %476 = tosa.clamp %475 { + LayerName = "Relu_319", + OutputName = "Relu_319", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc218) + %477 = tosa.reshape %476 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc354) + xten_nn.output %477 : tensor<1x240x1x1xbf16> loc(#loc218) + } -> tensor<1x240x1x1xbf16> loc(#loc354) + xten_nn.output %472 : tensor<1x240x1x1xbf16> loc(#loc354) + } -> tensor<1x240x1x1xbf16> loc(#loc354) + %363 = xten_nn.subgraph (%arg5 = %362: tensor<1x240x1x1xbf16>, %arg6 = %41: tensor<960x240x1x1xbf16>, %arg7 = %40: tensor<960xbf16>) attributes { + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "747", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "747", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc219) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc219) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_320", + PartOfOutputName = "Conv_320", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc219) + %476 = tosa.reshape %475 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc219) + xten_nn.output %476 : tensor<1x960x1x1xbf16> loc(#loc219) + } -> tensor<1x960x1x1xbf16> loc(#loc219) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc219) + } -> tensor<1x960x1x1xbf16> loc(#loc219) + %364 = xten_nn.subgraph (%arg5 = %363: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_322", OutputName = "Add_322"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc220) + xten_nn.output %474 : tensor<1x960x1x1xbf16> loc(#loc220) + } -> tensor<1x960x1x1xbf16> loc(#loc220) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc220) + } -> tensor<1x960x1x1xbf16> loc(#loc220) + %365 = xten_nn.subgraph (%arg5 = %364: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_325", + OutputName = "Clip_325", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc221) + xten_nn.output %473 : tensor<1x960x1x1xbf16> loc(#loc221) + } -> tensor<1x960x1x1xbf16> loc(#loc221) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc221) + } -> tensor<1x960x1x1xbf16> loc(#loc221) + %366 = xten_nn.subgraph (%arg5 = %365: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_327", + OutputName = "Div_327", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc222) + xten_nn.output %474 : tensor<1x960x1x1xbf16> loc(#loc222) + } -> tensor<1x960x1x1xbf16> loc(#loc222) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc222) + } -> tensor<1x960x1x1xbf16> loc(#loc222) + %367 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Mul_328_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_328_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %472 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + %368 = xten_nn.subgraph (%arg5 = %367: tensor<1x960x12x20xbf16>, %arg6 = %359: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_328", + OutputName = "Mul_328", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + %369 = xten_nn.subgraph (%arg5 = %368: tensor<1x960x12x20xbf16>, %arg6 = %39: tensor<160x960x1x1xbf16>, %arg7 = %38: tensor<160xbf16>, %arg8 = %349: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_329", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "756", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "757", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_329", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "756", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "757", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_329", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1060", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc224) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc224) + %477 = tosa.transpose %arg9, %475 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc224) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_329", + PartOfOutputName = "Conv_329", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc224) + %479 = tosa.transpose %478, %474 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc224) + xten_nn.output %479 : tensor<1x160x12x20xbf16> loc(#loc224) + } -> tensor<1x160x12x20xbf16> loc(#loc224) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_330", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1060", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.add %arg9, %arg10 {LayerName = "Add_330", OutputName = "Add_330"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc225) + xten_nn.output %474 : tensor<1x160x12x20xbf16> loc(#loc225) + } -> tensor<1x160x12x20xbf16> loc(#loc225) + xten_nn.output %473 : tensor<1x160x12x20xbf16> loc(#loc225) + } -> tensor<1x160x12x20xbf16> loc(#loc355) + %370 = xten_nn.subgraph (%arg5 = %369: tensor<1x160x12x20xbf16>, %arg6 = %37: tensor<960x160x1x1xbf16>, %arg7 = %36: tensor<960xbf16>) attributes { + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1060", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "760", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1060", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "760", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc226) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc226) + %476 = tosa.transpose %arg8, %474 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc226) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_331", + PartOfOutputName = "Conv_331", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc226) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc226) + xten_nn.output %478 : tensor<1x960x12x20xbf16> loc(#loc226) + } -> tensor<1x960x12x20xbf16> loc(#loc226) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc226) + } -> tensor<1x960x12x20xbf16> loc(#loc226) + %371 = xten_nn.subgraph (%arg5 = %370: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.add %arg6, %473 {LayerName = "Add_333", OutputName = "Add_333"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc227) + xten_nn.output %474 : tensor<1x960x12x20xbf16> loc(#loc227) + } -> tensor<1x960x12x20xbf16> loc(#loc227) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc227) + } -> tensor<1x960x12x20xbf16> loc(#loc227) + %372 = xten_nn.subgraph (%arg5 = %371: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_336", + OutputName = "Clip_336", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc228) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc228) + } -> tensor<1x960x12x20xbf16> loc(#loc228) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc228) + } -> tensor<1x960x12x20xbf16> loc(#loc228) + %373 = xten_nn.subgraph (%arg5 = %372: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "769", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "769", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %474 = tosa.mul %arg6, %473 { + LayerName = "Div_338", + OutputName = "Div_338", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc229) + xten_nn.output %474 : tensor<1x960x12x20xbf16> loc(#loc229) + } -> tensor<1x960x12x20xbf16> loc(#loc229) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc229) + } -> tensor<1x960x12x20xbf16> loc(#loc229) + %374 = xten_nn.subgraph (%arg5 = %370: tensor<1x960x12x20xbf16>, %arg6 = %373: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_339", + OutputName = "Mul_339", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc230) + xten_nn.output %473 : tensor<1x960x12x20xbf16> loc(#loc230) + } -> tensor<1x960x12x20xbf16> loc(#loc230) + xten_nn.output %472 : tensor<1x960x12x20xbf16> loc(#loc230) + } -> tensor<1x960x12x20xbf16> loc(#loc230) + %375 = xten_nn.subgraph (%arg5 = %374: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_342_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %472 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc356) + xten_nn.output %472 : tensor<1x960x1x240xbf16> loc(#loc356) + } -> tensor<1x960x1x240xbf16> loc(#loc356) + %376 = xten_nn.subgraph (%arg5 = %375: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_342", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_342_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_342", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_342_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %473 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc231) + xten_nn.output %473 : tensor<1x960x1x1xbf16> loc(#loc231) + } -> tensor<1x960x1x1xbf16> loc(#loc231) + xten_nn.output %472 : tensor<1x960x1x1xbf16> loc(#loc231) + } -> tensor<1x960x1x1xbf16> loc(#loc231) + %377 = xten_nn.subgraph (%arg5 = %376: tensor<1x960x1x1xbf16>, %arg6 = %35: tensor<128x960x1x1xbf16>, %arg7 = %34: tensor<128xbf16>) attributes { + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "aspp.aspp2.1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<128x960x1x1xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "aspp.aspp2.1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = tosa.reshape %arg9 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc232) + %474 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc232) + %475 = tosa.conv2d %474, %473, %arg10 { + PartOfLayerName = "Conv_343", + PartOfOutputName = "Conv_343", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x1x1x128xbf16> loc(#loc232) + %476 = tosa.reshape %475 {new_shape = array} : (tensor<1x1x1x128xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc232) + xten_nn.output %476 : tensor<1x128x1x1xbf16> loc(#loc232) + } -> tensor<1x128x1x1xbf16> loc(#loc232) + xten_nn.output %472 : tensor<1x128x1x1xbf16> loc(#loc232) + } -> tensor<1x128x1x1xbf16> loc(#loc232) + %378 = xten_nn.subgraph (%arg5 = %377: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "776", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "776", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_344", OutputName = "Sigmoid_344"} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc233) + xten_nn.output %473 : tensor<1x128x1x1xbf16> loc(#loc233) + } -> tensor<1x128x1x1xbf16> loc(#loc233) + xten_nn.output %472 : tensor<1x128x1x1xbf16> loc(#loc233) + } -> tensor<1x128x1x1xbf16> loc(#loc233) + %379 = xten_nn.subgraph (%arg5 = %378: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Mul_345_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "773", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_345_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 128 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %472 = tosa.tile %arg5 {multiples = array} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc234) + xten_nn.output %472 : tensor<1x128x12x20xbf16> loc(#loc234) + } -> tensor<1x128x12x20xbf16> loc(#loc234) + %380 = xten_nn.subgraph (%arg5 = %374: tensor<1x960x12x20xbf16>, %arg6 = %33: tensor<128x960x1x1xbf16>, %arg7 = %32: tensor<128xbf16>, %arg8 = %379: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Conv_340", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "1063", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1066", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<128x960x1x1xbf16>, %arg11 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_340", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "1063", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_341", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "773", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %474 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc358) + %476 = tosa.reshape %arg10 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc358) + %477 = tosa.transpose %arg9, %475 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc358) + %478 = tosa.conv2d %477, %476, %arg11 { + PartOfLayerName = "Conv_340", + PartOfOutputName = "Conv_340", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc235) + %479 = tosa.clamp %478 { + LayerName = "Relu_341", + OutputName = "Relu_341", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x12x20x128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc236) + %480 = tosa.transpose %479, %474 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc358) + xten_nn.output %480 : tensor<1x128x12x20xbf16> loc(#loc236) + } -> tensor<1x128x12x20xbf16> loc(#loc358) + %473 = xten_nn.subgraph (%arg9 = %472: tensor<1x128x12x20xbf16>, %arg10 = %arg8: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Mul_345_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "773", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1066", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %474 = tosa.mul %arg9, %arg10 { + LayerName = "Mul_345", + OutputName = "Mul_345", + shift = 0 : i8} : (tensor<1x128x12x20xbf16>, tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc234) + xten_nn.output %474 : tensor<1x128x12x20xbf16> loc(#loc234) + } -> tensor<1x128x12x20xbf16> loc(#loc234) + xten_nn.output %473 : tensor<1x128x12x20xbf16> loc(#loc234) + } -> tensor<1x128x12x20xbf16> loc(#loc357) + %381 = xten_nn.subgraph (%arg5 = %380: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "781", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + %382 = xten_nn.subgraph (%arg5 = %380: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "781", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + %383 = xten_nn.subgraph (%arg5 = %382: tensor<1x64x12x20xbf16>, %arg6 = %174: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_350", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_350", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "783", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_350", + OutputName = "Concat_350", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc238) + xten_nn.output %472 : tensor<1x128x12x20xbf16> loc(#loc238) + } -> tensor<1x128x12x20xbf16> loc(#loc238) + %384 = xten_nn.subgraph (%arg5 = %383: tensor<1x128x12x20xbf16>, %arg6 = %31: tensor<128x128x3x3xbf16>, %arg7 = %30: tensor<128xbf16>) attributes { + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "783", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "397", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<128x128x3x3xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "783", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "397", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<128x128x3x3xbf16>, tensor<4xi32>) -> tensor<128x3x3x128xbf16> loc(#loc239) + %476 = tosa.transpose %arg8, %474 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc239) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_351", + PartOfOutputName = "Conv_351", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<128x3x3x128xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc239) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc239) + xten_nn.output %478 : tensor<1x128x12x20xbf16> loc(#loc239) + } -> tensor<1x128x12x20xbf16> loc(#loc239) + xten_nn.output %472 : tensor<1x128x12x20xbf16> loc(#loc239) + } -> tensor<1x128x12x20xbf16> loc(#loc239) + %385 = xten_nn.subgraph (%arg5 = %384: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_352", OutputName = "Sigmoid_352"} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc240) + xten_nn.output %473 : tensor<1x128x12x20xbf16> loc(#loc240) + } -> tensor<1x128x12x20xbf16> loc(#loc240) + xten_nn.output %472 : tensor<1x128x12x20xbf16> loc(#loc240) + } -> tensor<1x128x12x20xbf16> loc(#loc240) + %386 = xten_nn.subgraph (%arg5 = %385: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + %387 = xten_nn.subgraph (%arg5 = %385: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + %388 = xten_nn.subgraph (%arg5 = %29: tensor<1x64x12x20xbf16>, %arg6 = %387: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.sub %arg7, %arg8 {LayerName = "Sub_359", OutputName = "Sub_359"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc4) + xten_nn.output %473 : tensor<1x64x12x20xbf16> loc(#loc4) + } -> tensor<1x64x12x20xbf16> loc(#loc4) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc4) + } -> tensor<1x64x12x20xbf16> loc(#loc4) + %389 = xten_nn.subgraph (%arg5 = %388: tensor<1x64x12x20xbf16>, %arg6 = %174: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_360", + OutputName = "Mul_360", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc242) + xten_nn.output %473 : tensor<1x64x12x20xbf16> loc(#loc242) + } -> tensor<1x64x12x20xbf16> loc(#loc242) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc242) + } -> tensor<1x64x12x20xbf16> loc(#loc242) + %390 = xten_nn.subgraph (%arg5 = %386: tensor<1x64x12x20xbf16>, %arg6 = %174: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "788", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "788", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_354", + OutputName = "Mul_354", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc243) + xten_nn.output %473 : tensor<1x64x12x20xbf16> loc(#loc243) + } -> tensor<1x64x12x20xbf16> loc(#loc243) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc243) + } -> tensor<1x64x12x20xbf16> loc(#loc243) + %391 = xten_nn.subgraph (%arg5 = %382: tensor<1x64x12x20xbf16>, %arg6 = %390: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_355", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "782", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "788", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_355", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "789", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_355", + OutputName = "Concat_355", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc244) + xten_nn.output %472 : tensor<1x128x12x20xbf16> loc(#loc244) + } -> tensor<1x128x12x20xbf16> loc(#loc244) + %392 = xten_nn.subgraph (%arg5 = %391: tensor<1x128x12x20xbf16>, %arg6 = %28: tensor<64x128x3x3xbf16>, %arg7 = %27: tensor<64xbf16>) attributes { + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "789", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "788", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<64x128x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "789", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "788", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<64x128x3x3xbf16>, tensor<4xi32>) -> tensor<64x3x3x128xbf16> loc(#loc245) + %476 = tosa.transpose %arg8, %474 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc245) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_356", + PartOfOutputName = "Conv_356", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<64x3x3x128xbf16>, tensor<64xbf16>) -> tensor<1x12x20x64xbf16> loc(#loc245) + %478 = tosa.transpose %477, %473 : (tensor<1x12x20x64xbf16>, tensor<4xi32>) -> tensor<1x64x12x20xbf16> loc(#loc245) + xten_nn.output %478 : tensor<1x64x12x20xbf16> loc(#loc245) + } -> tensor<1x64x12x20xbf16> loc(#loc245) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc245) + } -> tensor<1x64x12x20xbf16> loc(#loc245) + %393 = xten_nn.subgraph (%arg5 = %392: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.tanh %arg6 {LayerName = "Tanh_357", OutputName = "Tanh_357"} : (tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc246) + xten_nn.output %473 : tensor<1x64x12x20xbf16> loc(#loc246) + } -> tensor<1x64x12x20xbf16> loc(#loc246) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc246) + } -> tensor<1x64x12x20xbf16> loc(#loc246) + %394 = xten_nn.subgraph (%arg5 = %387: tensor<1x64x12x20xbf16>, %arg6 = %393: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "795", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "795", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_361", + OutputName = "Mul_361", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc247) + xten_nn.output %473 : tensor<1x64x12x20xbf16> loc(#loc247) + } -> tensor<1x64x12x20xbf16> loc(#loc247) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc247) + } -> tensor<1x64x12x20xbf16> loc(#loc247) + %395 = xten_nn.subgraph (%arg5 = %389: tensor<1x64x12x20xbf16>, %arg6 = %394: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "794", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "796", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "794", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "796", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.add %arg7, %arg8 {LayerName = "Add_362", OutputName = "Add_362"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc248) + xten_nn.output %473 : tensor<1x64x12x20xbf16> loc(#loc248) + } -> tensor<1x64x12x20xbf16> loc(#loc248) + xten_nn.output %472 : tensor<1x64x12x20xbf16> loc(#loc248) + } -> tensor<1x64x12x20xbf16> loc(#loc248) + %396 = xten_nn.subgraph (%arg5 = %381: tensor<1x64x12x20xbf16>, %arg6 = %395: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_363", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "781", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_363", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "797", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_363", + OutputName = "Concat_363", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc249) + xten_nn.output %472 : tensor<1x128x12x20xbf16> loc(#loc249) + } -> tensor<1x128x12x20xbf16> loc(#loc249) + %397 = xten_nn.subgraph (%arg5 = %396: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Resize_365", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "797", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Resize_365", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "802", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 128 : ui32, + config.dim_2 = 12 : ui32, + config.dim_3 = 20 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 24 : ui32, + config.output_W = 40 : ui32 + }} { + %472 = xten_nn.resize %arg5 { + LayerName = "Resize_365", + OutputName = "Resize_365", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x24x40xbf16> loc(#loc250) + xten_nn.output %472 : tensor<1x128x24x40xbf16> loc(#loc250) + } -> tensor<1x128x24x40xbf16> loc(#loc250) + %398 = xten_nn.subgraph (%arg5 = %397: tensor<1x128x24x40xbf16>) attributes { + LayerName = "Slice_371", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "802", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + OutputName = "Slice_371", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "812", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 128 : ui32, + config.dim_h = 24 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 23 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + LayerName = "Slice_371", + OutputName = "Slice_371", + size = array, + start = array} : (tensor<1x128x24x40xbf16>) -> tensor<1x128x23x40xbf16> loc(#loc251) + xten_nn.output %472 : tensor<1x128x23x40xbf16> loc(#loc251) + } -> tensor<1x128x23x40xbf16> loc(#loc251) + %399 = xten_nn.subgraph (%arg5 = %398: tensor<1x128x23x40xbf16>, %arg6 = %225: tensor<1x40x23x40xbf16>, %arg7 = %173: tensor<1x3x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_372", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "812", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "802", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_372", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "813", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_372", + OutputName = "Concat_372", + axis = 1 : i32} : (tensor<1x128x23x40xbf16>, tensor<1x40x23x40xbf16>, tensor<1x3x23x40xbf16>) -> tensor<1x171x23x40xbf16> loc(#loc252) + xten_nn.output %472 : tensor<1x171x23x40xbf16> loc(#loc252) + } -> tensor<1x171x23x40xbf16> loc(#loc252) + %400 = xten_nn.subgraph (%arg5 = %399: tensor<1x171x23x40xbf16>, %arg6 = %26: tensor<80x171x3x3xbf16>, %arg7 = %25: tensor<80xbf16>) attributes { + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "813", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Name = "812", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Name = "813", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x171x23x40xbf16>, %arg9 = %arg6: tensor<80x171x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "813", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Name = "812", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Name = "813", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<80x171x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x171xbf16> loc(#loc359) + %476 = tosa.transpose %arg8, %474 : (tensor<1x171x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x171xbf16> loc(#loc359) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_373", + PartOfOutputName = "Conv_373", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x171xbf16>, tensor<80x3x3x171xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc253) + %478 = tosa.clamp %477 { + LayerName = "Relu_374", + OutputName = "Relu_374", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc254) + %479 = tosa.transpose %478, %473 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc359) + xten_nn.output %479 : tensor<1x80x23x40xbf16> loc(#loc254) + } -> tensor<1x80x23x40xbf16> loc(#loc359) + xten_nn.output %472 : tensor<1x80x23x40xbf16> loc(#loc359) + } -> tensor<1x80x23x40xbf16> loc(#loc359) + %401 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "817", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + %402 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "817", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + %403 = xten_nn.subgraph (%arg5 = %402: tensor<1x40x23x40xbf16>, %arg6 = %164: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_376", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_376", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "819", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_376", + OutputName = "Concat_376", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc256) + xten_nn.output %472 : tensor<1x80x23x40xbf16> loc(#loc256) + } -> tensor<1x80x23x40xbf16> loc(#loc256) + %404 = xten_nn.subgraph (%arg5 = %403: tensor<1x80x23x40xbf16>, %arg6 = %24: tensor<80x80x3x3xbf16>, %arg7 = %23: tensor<80xbf16>) attributes { + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "819", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "396", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<80x80x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "819", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "396", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<80x80x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x80xbf16> loc(#loc257) + %476 = tosa.transpose %arg8, %474 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc257) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_377", + PartOfOutputName = "Conv_377", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<80x3x3x80xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc257) + %478 = tosa.transpose %477, %473 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc257) + xten_nn.output %478 : tensor<1x80x23x40xbf16> loc(#loc257) + } -> tensor<1x80x23x40xbf16> loc(#loc257) + xten_nn.output %472 : tensor<1x80x23x40xbf16> loc(#loc257) + } -> tensor<1x80x23x40xbf16> loc(#loc257) + %405 = xten_nn.subgraph (%arg5 = %404: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_378", OutputName = "Sigmoid_378"} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc258) + xten_nn.output %473 : tensor<1x80x23x40xbf16> loc(#loc258) + } -> tensor<1x80x23x40xbf16> loc(#loc258) + xten_nn.output %472 : tensor<1x80x23x40xbf16> loc(#loc258) + } -> tensor<1x80x23x40xbf16> loc(#loc258) + %406 = xten_nn.subgraph (%arg5 = %405: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + %407 = xten_nn.subgraph (%arg5 = %405: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + %408 = xten_nn.subgraph (%arg5 = %22: tensor<1x40x23x40xbf16>, %arg6 = %407: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.sub %arg7, %arg8 {LayerName = "Sub_385", OutputName = "Sub_385"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc3) + xten_nn.output %473 : tensor<1x40x23x40xbf16> loc(#loc3) + } -> tensor<1x40x23x40xbf16> loc(#loc3) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc3) + } -> tensor<1x40x23x40xbf16> loc(#loc3) + %409 = xten_nn.subgraph (%arg5 = %408: tensor<1x40x23x40xbf16>, %arg6 = %164: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_386", + OutputName = "Mul_386", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %473 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + %410 = xten_nn.subgraph (%arg5 = %406: tensor<1x40x23x40xbf16>, %arg6 = %164: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "824", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "824", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_380", + OutputName = "Mul_380", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %473 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + %411 = xten_nn.subgraph (%arg5 = %402: tensor<1x40x23x40xbf16>, %arg6 = %410: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_381", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "818", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "824", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_381", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "825", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_381", + OutputName = "Concat_381", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc262) + xten_nn.output %472 : tensor<1x80x23x40xbf16> loc(#loc262) + } -> tensor<1x80x23x40xbf16> loc(#loc262) + %412 = xten_nn.subgraph (%arg5 = %411: tensor<1x80x23x40xbf16>, %arg6 = %21: tensor<40x80x3x3xbf16>, %arg7 = %20: tensor<40xbf16>) attributes { + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "825", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "824", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<40x80x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "825", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "824", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<40x80x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x80xbf16> loc(#loc263) + %476 = tosa.transpose %arg8, %474 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc263) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_382", + PartOfOutputName = "Conv_382", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<40x3x3x80xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc263) + %478 = tosa.transpose %477, %473 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc263) + xten_nn.output %478 : tensor<1x40x23x40xbf16> loc(#loc263) + } -> tensor<1x40x23x40xbf16> loc(#loc263) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc263) + } -> tensor<1x40x23x40xbf16> loc(#loc263) + %413 = xten_nn.subgraph (%arg5 = %412: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.tanh %arg6 {LayerName = "Tanh_383", OutputName = "Tanh_383"} : (tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc264) + xten_nn.output %473 : tensor<1x40x23x40xbf16> loc(#loc264) + } -> tensor<1x40x23x40xbf16> loc(#loc264) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc264) + } -> tensor<1x40x23x40xbf16> loc(#loc264) + %414 = xten_nn.subgraph (%arg5 = %407: tensor<1x40x23x40xbf16>, %arg6 = %413: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "831", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "831", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_387", + OutputName = "Mul_387", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc265) + xten_nn.output %473 : tensor<1x40x23x40xbf16> loc(#loc265) + } -> tensor<1x40x23x40xbf16> loc(#loc265) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc265) + } -> tensor<1x40x23x40xbf16> loc(#loc265) + %415 = xten_nn.subgraph (%arg5 = %409: tensor<1x40x23x40xbf16>, %arg6 = %414: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "830", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "832", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "830", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "832", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.add %arg7, %arg8 {LayerName = "Add_388", OutputName = "Add_388"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc266) + xten_nn.output %473 : tensor<1x40x23x40xbf16> loc(#loc266) + } -> tensor<1x40x23x40xbf16> loc(#loc266) + xten_nn.output %472 : tensor<1x40x23x40xbf16> loc(#loc266) + } -> tensor<1x40x23x40xbf16> loc(#loc266) + %416 = xten_nn.subgraph (%arg5 = %401: tensor<1x40x23x40xbf16>, %arg6 = %415: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_389", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "817", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_389", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "833", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_389", + OutputName = "Concat_389", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc267) + xten_nn.output %472 : tensor<1x80x23x40xbf16> loc(#loc267) + } -> tensor<1x80x23x40xbf16> loc(#loc267) + %417 = xten_nn.subgraph (%arg5 = %416: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Resize_391", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "833", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Resize_391", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "838", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 80 : ui32, + config.dim_2 = 23 : ui32, + config.dim_3 = 40 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 46 : ui32, + config.output_W = 80 : ui32 + }} { + %472 = xten_nn.resize %arg5 { + LayerName = "Resize_391", + OutputName = "Resize_391", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x46x80xbf16> loc(#loc268) + xten_nn.output %472 : tensor<1x80x46x80xbf16> loc(#loc268) + } -> tensor<1x80x46x80xbf16> loc(#loc268) + %418 = xten_nn.subgraph (%arg5 = %417: tensor<1x80x46x80xbf16>) attributes { + LayerName = "Slice_397", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "838", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + OutputName = "Slice_397", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "848", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 80 : ui32, + config.dim_h = 46 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 45 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + LayerName = "Slice_397", + OutputName = "Slice_397", + size = array, + start = array} : (tensor<1x80x46x80xbf16>) -> tensor<1x80x45x80xbf16> loc(#loc269) + xten_nn.output %472 : tensor<1x80x45x80xbf16> loc(#loc269) + } -> tensor<1x80x45x80xbf16> loc(#loc269) + %419 = xten_nn.subgraph (%arg5 = %418: tensor<1x80x45x80xbf16>, %arg6 = %189: tensor<1x24x45x80xbf16>, %arg7 = %172: tensor<1x3x45x80xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_398", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "848", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "838", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "849", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_398", + OutputName = "Concat_398", + axis = 1 : i32} : (tensor<1x80x45x80xbf16>, tensor<1x24x45x80xbf16>, tensor<1x3x45x80xbf16>) -> tensor<1x107x45x80xbf16> loc(#loc270) + xten_nn.output %472 : tensor<1x107x45x80xbf16> loc(#loc270) + } -> tensor<1x107x45x80xbf16> loc(#loc270) + %420 = xten_nn.subgraph (%arg5 = %419: tensor<1x107x45x80xbf16>, %arg6 = %19: tensor<40x107x3x3xbf16>, %arg7 = %18: tensor<40xbf16>) attributes { + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "849", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Name = "848", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Name = "849", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x107x45x80xbf16>, %arg9 = %arg6: tensor<40x107x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "849", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Name = "848", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Name = "849", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<40x107x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x107xbf16> loc(#loc360) + %476 = tosa.transpose %arg8, %474 : (tensor<1x107x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x107xbf16> loc(#loc360) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_399", + PartOfOutputName = "Conv_399", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x107xbf16>, tensor<40x3x3x107xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc271) + %478 = tosa.clamp %477 { + LayerName = "Relu_400", + OutputName = "Relu_400", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc272) + %479 = tosa.transpose %478, %473 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc360) + xten_nn.output %479 : tensor<1x40x45x80xbf16> loc(#loc272) + } -> tensor<1x40x45x80xbf16> loc(#loc360) + xten_nn.output %472 : tensor<1x40x45x80xbf16> loc(#loc360) + } -> tensor<1x40x45x80xbf16> loc(#loc360) + %421 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "853", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + %422 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "853", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + %423 = xten_nn.subgraph (%arg5 = %422: tensor<1x20x45x80xbf16>, %arg6 = %163: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_402", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_402", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "855", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_402", + OutputName = "Concat_402", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc274) + xten_nn.output %472 : tensor<1x40x45x80xbf16> loc(#loc274) + } -> tensor<1x40x45x80xbf16> loc(#loc274) + %424 = xten_nn.subgraph (%arg5 = %423: tensor<1x40x45x80xbf16>, %arg6 = %17: tensor<40x40x3x3xbf16>, %arg7 = %16: tensor<40xbf16>) attributes { + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "855", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "395", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<40x40x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "855", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "395", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<40x40x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x40xbf16> loc(#loc275) + %476 = tosa.transpose %arg8, %474 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc275) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_403", + PartOfOutputName = "Conv_403", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<40x3x3x40xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc275) + %478 = tosa.transpose %477, %473 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc275) + xten_nn.output %478 : tensor<1x40x45x80xbf16> loc(#loc275) + } -> tensor<1x40x45x80xbf16> loc(#loc275) + xten_nn.output %472 : tensor<1x40x45x80xbf16> loc(#loc275) + } -> tensor<1x40x45x80xbf16> loc(#loc275) + %425 = xten_nn.subgraph (%arg5 = %424: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_404", OutputName = "Sigmoid_404"} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc276) + xten_nn.output %473 : tensor<1x40x45x80xbf16> loc(#loc276) + } -> tensor<1x40x45x80xbf16> loc(#loc276) + xten_nn.output %472 : tensor<1x40x45x80xbf16> loc(#loc276) + } -> tensor<1x40x45x80xbf16> loc(#loc276) + %426 = xten_nn.subgraph (%arg5 = %425: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + %427 = xten_nn.subgraph (%arg5 = %425: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + %428 = xten_nn.subgraph (%arg5 = %15: tensor<1x20x45x80xbf16>, %arg6 = %427: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.sub %arg7, %arg8 {LayerName = "Sub_411", OutputName = "Sub_411"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc2) + xten_nn.output %473 : tensor<1x20x45x80xbf16> loc(#loc2) + } -> tensor<1x20x45x80xbf16> loc(#loc2) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc2) + } -> tensor<1x20x45x80xbf16> loc(#loc2) + %429 = xten_nn.subgraph (%arg5 = %428: tensor<1x20x45x80xbf16>, %arg6 = %163: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_412", + OutputName = "Mul_412", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %473 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + %430 = xten_nn.subgraph (%arg5 = %426: tensor<1x20x45x80xbf16>, %arg6 = %163: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "860", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "860", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_406", + OutputName = "Mul_406", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %473 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + %431 = xten_nn.subgraph (%arg5 = %422: tensor<1x20x45x80xbf16>, %arg6 = %430: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_407", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "854", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "860", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_407", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "861", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_407", + OutputName = "Concat_407", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc280) + xten_nn.output %472 : tensor<1x40x45x80xbf16> loc(#loc280) + } -> tensor<1x40x45x80xbf16> loc(#loc280) + %432 = xten_nn.subgraph (%arg5 = %431: tensor<1x40x45x80xbf16>, %arg6 = %14: tensor<20x40x3x3xbf16>, %arg7 = %13: tensor<20xbf16>) attributes { + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "861", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "860", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<20x40x3x3xbf16>, %arg10 = %arg7: tensor<20xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "861", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "860", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<20x40x3x3xbf16>, tensor<4xi32>) -> tensor<20x3x3x40xbf16> loc(#loc281) + %476 = tosa.transpose %arg8, %474 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc281) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_408", + PartOfOutputName = "Conv_408", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<20x3x3x40xbf16>, tensor<20xbf16>) -> tensor<1x45x80x20xbf16> loc(#loc281) + %478 = tosa.transpose %477, %473 : (tensor<1x45x80x20xbf16>, tensor<4xi32>) -> tensor<1x20x45x80xbf16> loc(#loc281) + xten_nn.output %478 : tensor<1x20x45x80xbf16> loc(#loc281) + } -> tensor<1x20x45x80xbf16> loc(#loc281) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc281) + } -> tensor<1x20x45x80xbf16> loc(#loc281) + %433 = xten_nn.subgraph (%arg5 = %432: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.tanh %arg6 {LayerName = "Tanh_409", OutputName = "Tanh_409"} : (tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc282) + xten_nn.output %473 : tensor<1x20x45x80xbf16> loc(#loc282) + } -> tensor<1x20x45x80xbf16> loc(#loc282) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc282) + } -> tensor<1x20x45x80xbf16> loc(#loc282) + %434 = xten_nn.subgraph (%arg5 = %427: tensor<1x20x45x80xbf16>, %arg6 = %433: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "867", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "867", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_413", + OutputName = "Mul_413", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc283) + xten_nn.output %473 : tensor<1x20x45x80xbf16> loc(#loc283) + } -> tensor<1x20x45x80xbf16> loc(#loc283) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc283) + } -> tensor<1x20x45x80xbf16> loc(#loc283) + %435 = xten_nn.subgraph (%arg5 = %429: tensor<1x20x45x80xbf16>, %arg6 = %434: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "866", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "868", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "866", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "868", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.add %arg7, %arg8 {LayerName = "Add_414", OutputName = "Add_414"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc284) + xten_nn.output %473 : tensor<1x20x45x80xbf16> loc(#loc284) + } -> tensor<1x20x45x80xbf16> loc(#loc284) + xten_nn.output %472 : tensor<1x20x45x80xbf16> loc(#loc284) + } -> tensor<1x20x45x80xbf16> loc(#loc284) + %436 = xten_nn.subgraph (%arg5 = %421: tensor<1x20x45x80xbf16>, %arg6 = %435: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_415", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "853", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_415", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "869", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_415", + OutputName = "Concat_415", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc285) + xten_nn.output %472 : tensor<1x40x45x80xbf16> loc(#loc285) + } -> tensor<1x40x45x80xbf16> loc(#loc285) + %437 = xten_nn.subgraph (%arg5 = %436: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Resize_417", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "869", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Resize_417", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "874", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 40 : ui32, + config.dim_2 = 45 : ui32, + config.dim_3 = 80 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 90 : ui32, + config.output_W = 160 : ui32 + }} { + %472 = xten_nn.resize %arg5 { + LayerName = "Resize_417", + OutputName = "Resize_417", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x90x160xbf16> loc(#loc286) + xten_nn.output %472 : tensor<1x40x90x160xbf16> loc(#loc286) + } -> tensor<1x40x90x160xbf16> loc(#loc286) + %438 = xten_nn.subgraph (%arg5 = %437: tensor<1x40x90x160xbf16>, %arg6 = %183: tensor<1x16x90x160xbf16>, %arg7 = %171: tensor<1x3x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_418", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "874", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "869", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_418", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "875", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_418", + OutputName = "Concat_418", + axis = 1 : i32} : (tensor<1x40x90x160xbf16>, tensor<1x16x90x160xbf16>, tensor<1x3x90x160xbf16>) -> tensor<1x59x90x160xbf16> loc(#loc287) + xten_nn.output %472 : tensor<1x59x90x160xbf16> loc(#loc287) + } -> tensor<1x59x90x160xbf16> loc(#loc287) + %439 = xten_nn.subgraph (%arg5 = %438: tensor<1x59x90x160xbf16>, %arg6 = %12: tensor<32x59x3x3xbf16>, %arg7 = %11: tensor<32xbf16>) attributes { + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "875", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Name = "874", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Name = "875", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x59x90x160xbf16>, %arg9 = %arg6: tensor<32x59x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "875", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Name = "874", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Name = "875", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<32x59x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x59xbf16> loc(#loc361) + %476 = tosa.transpose %arg8, %474 : (tensor<1x59x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x59xbf16> loc(#loc361) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_419", + PartOfOutputName = "Conv_419", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x59xbf16>, tensor<32x3x3x59xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc288) + %478 = tosa.clamp %477 { + LayerName = "Relu_420", + OutputName = "Relu_420", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc289) + %479 = tosa.transpose %478, %473 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc361) + xten_nn.output %479 : tensor<1x32x90x160xbf16> loc(#loc289) + } -> tensor<1x32x90x160xbf16> loc(#loc361) + xten_nn.output %472 : tensor<1x32x90x160xbf16> loc(#loc361) + } -> tensor<1x32x90x160xbf16> loc(#loc361) + %440 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "879", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + %441 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "879", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + %442 = xten_nn.subgraph (%arg5 = %441: tensor<1x16x90x160xbf16>, %arg6 = %162: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_422", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_422", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "881", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_422", + OutputName = "Concat_422", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc291) + xten_nn.output %472 : tensor<1x32x90x160xbf16> loc(#loc291) + } -> tensor<1x32x90x160xbf16> loc(#loc291) + %443 = xten_nn.subgraph (%arg5 = %442: tensor<1x32x90x160xbf16>, %arg6 = %10: tensor<32x32x3x3xbf16>, %arg7 = %9: tensor<32xbf16>) attributes { + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "881", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "394", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<32x32x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "881", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "394", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<32x32x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x32xbf16> loc(#loc292) + %476 = tosa.transpose %arg8, %474 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc292) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_423", + PartOfOutputName = "Conv_423", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<32x3x3x32xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc292) + %478 = tosa.transpose %477, %473 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc292) + xten_nn.output %478 : tensor<1x32x90x160xbf16> loc(#loc292) + } -> tensor<1x32x90x160xbf16> loc(#loc292) + xten_nn.output %472 : tensor<1x32x90x160xbf16> loc(#loc292) + } -> tensor<1x32x90x160xbf16> loc(#loc292) + %444 = xten_nn.subgraph (%arg5 = %443: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_424", OutputName = "Sigmoid_424"} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc293) + xten_nn.output %473 : tensor<1x32x90x160xbf16> loc(#loc293) + } -> tensor<1x32x90x160xbf16> loc(#loc293) + xten_nn.output %472 : tensor<1x32x90x160xbf16> loc(#loc293) + } -> tensor<1x32x90x160xbf16> loc(#loc293) + %445 = xten_nn.subgraph (%arg5 = %444: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + %446 = xten_nn.subgraph (%arg5 = %444: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + %447 = xten_nn.subgraph (%arg5 = %8: tensor<1x16x90x160xbf16>, %arg6 = %446: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.sub %arg7, %arg8 {LayerName = "Sub_431", OutputName = "Sub_431"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc1) + xten_nn.output %473 : tensor<1x16x90x160xbf16> loc(#loc1) + } -> tensor<1x16x90x160xbf16> loc(#loc1) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc1) + } -> tensor<1x16x90x160xbf16> loc(#loc1) + %448 = xten_nn.subgraph (%arg5 = %447: tensor<1x16x90x160xbf16>, %arg6 = %162: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_432", + OutputName = "Mul_432", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %473 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + %449 = xten_nn.subgraph (%arg5 = %445: tensor<1x16x90x160xbf16>, %arg6 = %162: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "886", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "886", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_426", + OutputName = "Mul_426", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %473 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + %450 = xten_nn.subgraph (%arg5 = %441: tensor<1x16x90x160xbf16>, %arg6 = %449: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_427", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "880", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "886", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_427", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "887", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_427", + OutputName = "Concat_427", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc297) + xten_nn.output %472 : tensor<1x32x90x160xbf16> loc(#loc297) + } -> tensor<1x32x90x160xbf16> loc(#loc297) + %451 = xten_nn.subgraph (%arg5 = %450: tensor<1x32x90x160xbf16>, %arg6 = %7: tensor<16x32x3x3xbf16>, %arg7 = %6: tensor<16xbf16>) attributes { + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "887", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "886", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<16x32x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "887", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "886", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<16x32x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x32xbf16> loc(#loc298) + %476 = tosa.transpose %arg8, %474 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc298) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_428", + PartOfOutputName = "Conv_428", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<16x3x3x32xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc298) + %478 = tosa.transpose %477, %473 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc298) + xten_nn.output %478 : tensor<1x16x90x160xbf16> loc(#loc298) + } -> tensor<1x16x90x160xbf16> loc(#loc298) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc298) + } -> tensor<1x16x90x160xbf16> loc(#loc298) + %452 = xten_nn.subgraph (%arg5 = %451: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.tanh %arg6 {LayerName = "Tanh_429", OutputName = "Tanh_429"} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc299) + xten_nn.output %473 : tensor<1x16x90x160xbf16> loc(#loc299) + } -> tensor<1x16x90x160xbf16> loc(#loc299) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc299) + } -> tensor<1x16x90x160xbf16> loc(#loc299) + %453 = xten_nn.subgraph (%arg5 = %446: tensor<1x16x90x160xbf16>, %arg6 = %452: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "893", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "893", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_433", + OutputName = "Mul_433", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc300) + xten_nn.output %473 : tensor<1x16x90x160xbf16> loc(#loc300) + } -> tensor<1x16x90x160xbf16> loc(#loc300) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc300) + } -> tensor<1x16x90x160xbf16> loc(#loc300) + %454 = xten_nn.subgraph (%arg5 = %448: tensor<1x16x90x160xbf16>, %arg6 = %453: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "892", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "894", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "892", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "894", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.add %arg7, %arg8 {LayerName = "Add_434", OutputName = "Add_434"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc301) + xten_nn.output %473 : tensor<1x16x90x160xbf16> loc(#loc301) + } -> tensor<1x16x90x160xbf16> loc(#loc301) + xten_nn.output %472 : tensor<1x16x90x160xbf16> loc(#loc301) + } -> tensor<1x16x90x160xbf16> loc(#loc301) + %455 = xten_nn.subgraph (%arg5 = %454: tensor<1x16x90x160xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %472 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc302) + %473 = tosa.transpose %arg5, %472 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc302) + xten_nn.output %473 : tensor<1x90x160x16xbf16> loc(#loc302) + } -> tensor<1x90x160x16xbf16> loc(#loc302) + %456 = xten_nn.subgraph (%arg5 = %435: tensor<1x20x45x80xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %472 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc303) + %473 = tosa.transpose %arg5, %472 : (tensor<1x20x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x20xbf16> loc(#loc303) + xten_nn.output %473 : tensor<1x45x80x20xbf16> loc(#loc303) + } -> tensor<1x45x80x20xbf16> loc(#loc303) + %457 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x23x40xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %472 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc304) + %473 = tosa.transpose %arg5, %472 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc304) + xten_nn.output %473 : tensor<1x23x40x40xbf16> loc(#loc304) + } -> tensor<1x23x40x40xbf16> loc(#loc304) + %458 = xten_nn.subgraph (%arg5 = %395: tensor<1x64x12x20xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %472 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc305) + %473 = tosa.transpose %arg5, %472 : (tensor<1x64x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x64xbf16> loc(#loc305) + xten_nn.output %473 : tensor<1x12x20x64xbf16> loc(#loc305) + } -> tensor<1x12x20x64xbf16> loc(#loc305) + %459 = xten_nn.subgraph (%arg5 = %440: tensor<1x16x90x160xbf16>, %arg6 = %454: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_435", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "879", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_435", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "895", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_435", + OutputName = "Concat_435", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc306) + xten_nn.output %472 : tensor<1x32x90x160xbf16> loc(#loc306) + } -> tensor<1x32x90x160xbf16> loc(#loc306) + %460 = xten_nn.subgraph (%arg5 = %459: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Resize_437", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "895", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Resize_437", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "900", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 32 : ui32, + config.dim_2 = 90 : ui32, + config.dim_3 = 160 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 180 : ui32, + config.output_W = 320 : ui32 + }} { + %472 = xten_nn.resize %arg5 { + LayerName = "Resize_437", + OutputName = "Resize_437", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x180x320xbf16> loc(#loc307) + xten_nn.output %472 : tensor<1x32x180x320xbf16> loc(#loc307) + } -> tensor<1x32x180x320xbf16> loc(#loc307) + %461 = xten_nn.subgraph (%arg5 = %460: tensor<1x32x180x320xbf16>, %arg6 = %170: tensor<1x3x180x320xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_438", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "900", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "895", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Concat_438", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "901", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %472 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_438", + OutputName = "Concat_438", + axis = 1 : i32} : (tensor<1x32x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x35x180x320xbf16> loc(#loc308) + xten_nn.output %472 : tensor<1x35x180x320xbf16> loc(#loc308) + } -> tensor<1x35x180x320xbf16> loc(#loc308) + %462 = xten_nn.subgraph (%arg5 = %461: tensor<1x35x180x320xbf16>, %arg6 = %5: tensor<16x35x3x3xbf16>, %arg7 = %4: tensor<16xbf16>) attributes { + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "901", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Name = "900", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Name = "901", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x35x180x320xbf16>, %arg9 = %arg6: tensor<16x35x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "901", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Name = "900", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Name = "901", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<16x35x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x35xbf16> loc(#loc362) + %476 = tosa.transpose %arg8, %474 : (tensor<1x35x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x35xbf16> loc(#loc362) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_439", + PartOfOutputName = "Conv_439", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x35xbf16>, tensor<16x3x3x35xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc309) + %478 = tosa.clamp %477 { + LayerName = "Relu_440", + OutputName = "Relu_440", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc310) + %479 = tosa.transpose %478, %473 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc362) + xten_nn.output %479 : tensor<1x16x180x320xbf16> loc(#loc310) + } -> tensor<1x16x180x320xbf16> loc(#loc362) + xten_nn.output %472 : tensor<1x16x180x320xbf16> loc(#loc362) + } -> tensor<1x16x180x320xbf16> loc(#loc362) + %463 = xten_nn.subgraph (%arg5 = %462: tensor<1x16x180x320xbf16>, %arg6 = %3: tensor<16x16x3x3xbf16>, %arg7 = %2: tensor<16xbf16>) attributes { + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1078", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Name = "1082", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<16x16x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1078", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Name = "1082", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %475 = tosa.transpose %arg9, %474 : (tensor<16x16x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x16xbf16> loc(#loc363) + %476 = tosa.transpose %arg8, %474 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc363) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_441", + PartOfOutputName = "Conv_441", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<16x3x3x16xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc311) + %478 = tosa.clamp %477 { + LayerName = "Relu_442", + OutputName = "Relu_442", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc312) + %479 = tosa.transpose %478, %473 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc363) + xten_nn.output %479 : tensor<1x16x180x320xbf16> loc(#loc312) + } -> tensor<1x16x180x320xbf16> loc(#loc363) + xten_nn.output %472 : tensor<1x16x180x320xbf16> loc(#loc363) + } -> tensor<1x16x180x320xbf16> loc(#loc363) + %464 = xten_nn.subgraph (%arg5 = %463: tensor<1x16x180x320xbf16>, %arg6 = %1: tensor<4x16x1x1xbf16>, %arg7 = %0: tensor<4xbf16>) attributes { + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1081", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "project_mat.conv.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<4x16x1x1xbf16>, %arg10 = %arg7: tensor<4xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1081", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "project_mat.conv.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %473 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %474 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc313) + %475 = tosa.reshape %arg9 {new_shape = array} : (tensor<4x16x1x1xbf16>) -> tensor<4x1x1x16xbf16> loc(#loc313) + %476 = tosa.transpose %arg8, %474 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc313) + %477 = tosa.conv2d %476, %475, %arg10 { + PartOfLayerName = "Conv_443", + PartOfOutputName = "Conv_443", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<4x1x1x16xbf16>, tensor<4xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc313) + %478 = tosa.transpose %477, %473 : (tensor<1x180x320x4xbf16>, tensor<4xi32>) -> tensor<1x4x180x320xbf16> loc(#loc313) + xten_nn.output %478 : tensor<1x4x180x320xbf16> loc(#loc313) + } -> tensor<1x4x180x320xbf16> loc(#loc313) + xten_nn.output %472 : tensor<1x4x180x320xbf16> loc(#loc313) + } -> tensor<1x4x180x320xbf16> loc(#loc313) + %465 = xten_nn.subgraph (%arg5 = %464: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc314) + xten_nn.output %472 : tensor<1x3x180x320xbf16> loc(#loc314) + } -> tensor<1x3x180x320xbf16> loc(#loc314) + %466 = xten_nn.subgraph (%arg5 = %464: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 4 : ui32, + config.start = 3 : ui32, + config.step = 1 : ui32 + }} { + %472 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc314) + xten_nn.output %472 : tensor<1x1x180x320xbf16> loc(#loc314) + } -> tensor<1x1x180x320xbf16> loc(#loc314) + %467 = xten_nn.subgraph (%arg5 = %465: tensor<1x3x180x320xbf16>, %arg6 = %170: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %473 = tosa.add %arg7, %arg8 {LayerName = "Add_445", OutputName = "Add_445"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc15) + xten_nn.output %473 : tensor<1x3x180x320xbf16> loc(#loc15) + } -> tensor<1x3x180x320xbf16> loc(#loc15) + xten_nn.output %472 : tensor<1x3x180x320xbf16> loc(#loc15) + } -> tensor<1x3x180x320xbf16> loc(#loc15) + %468 = xten_nn.subgraph (%arg5 = %467: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "916", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "916", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_446", + OutputName = "Clip_446", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc315) + xten_nn.output %473 : tensor<1x3x180x320xbf16> loc(#loc315) + } -> tensor<1x3x180x320xbf16> loc(#loc315) + xten_nn.output %472 : tensor<1x3x180x320xbf16> loc(#loc315) + } -> tensor<1x3x180x320xbf16> loc(#loc315) + %469 = xten_nn.subgraph (%arg5 = %468: tensor<1x3x180x320xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %472 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc316) + %473 = tosa.transpose %arg5, %472 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc316) + xten_nn.output %473 : tensor<1x180x320x3xbf16> loc(#loc316) + } -> tensor<1x180x320x3xbf16> loc(#loc316) + %470 = xten_nn.subgraph (%arg5 = %466: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "910", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "921", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %472 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "910", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "921", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %473 = tosa.clamp %arg6 { + LayerName = "Clip_447", + OutputName = "Clip_447", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc317) + xten_nn.output %473 : tensor<1x1x180x320xbf16> loc(#loc317) + } -> tensor<1x1x180x320xbf16> loc(#loc317) + xten_nn.output %472 : tensor<1x1x180x320xbf16> loc(#loc317) + } -> tensor<1x1x180x320xbf16> loc(#loc317) + %471 = xten_nn.subgraph (%arg5 = %470: tensor<1x1x180x320xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %472 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x1x180x320xbf16>) -> tensor<1x180x320x1xbf16> loc(#loc318) + xten_nn.output %472 : tensor<1x180x320x1xbf16> loc(#loc318) + } -> tensor<1x180x320x1xbf16> loc(#loc318) + return %469, %471, %455, %456, %457, %458 : tensor<1x180x320x3xbf16>, tensor<1x180x320x1xbf16>, tensor<1x90x160x16xbf16>, tensor<1x45x80x20xbf16>, tensor<1x23x40x40xbf16>, tensor<1x12x20x64xbf16> loc(#loc) + } loc(#loc) +} loc(#loc) +#loc1 = loc("Sub_431") +#loc2 = loc("Sub_411") +#loc3 = loc("Sub_385") +#loc4 = loc("Sub_359") +#loc5 = loc("Div_16") +#loc6 = loc("Sub_14") +#loc7 = loc("Initializer_398") +#loc8 = loc("Transpose_9") +#loc9 = loc("Transpose_10") +#loc10 = loc("Transpose_11") +#loc11 = loc("Cast_0") +#loc12 = loc("Div_2") +#loc13 = loc("Slice_7") +#loc14 = loc("CompilerGeneratedLoc") +#loc15 = loc("Add_445") +#loc16 = loc("AveragePool_346") +#loc17 = loc("AveragePool_347") +#loc18 = loc("AveragePool_348") +#loc19 = loc("Transpose_12") +#loc20 = loc("Conv_17") +#loc21 = loc("Add_19") +#loc22 = loc("Clip_22") +#loc23 = loc("Div_24") +#loc24 = loc("Mul_25") +#loc25 = loc("Conv_26") +#loc26 = loc("Relu_27") +#loc27 = loc("Conv_28") +#loc28 = loc("Add_29") +#loc29 = loc("Conv_30") +#loc30 = loc("Relu_31") +#loc31 = loc("Conv_32") +#loc32 = loc("Relu_33") +#loc33 = loc("Conv_34") +#loc34 = loc("Conv_35") +#loc35 = loc("Relu_36") +#loc36 = loc("Conv_37") +#loc37 = loc("Relu_38") +#loc38 = loc("Conv_39") +#loc39 = loc("Add_40") +#loc40 = loc("Conv_41") +#loc41 = loc("Relu_42") +#loc42 = loc("Conv_43") +#loc43 = loc("Relu_44") +#loc44 = loc("GlobalAveragePool_45") +#loc45 = loc("Conv_46") +#loc46 = loc("Relu_47") +#loc47 = loc("Conv_48") +#loc48 = loc("Add_50") +#loc49 = loc("Clip_53") +#loc50 = loc("Div_55") +#loc51 = loc("Mul_56") +#loc52 = loc("Conv_57") +#loc53 = loc("Conv_58") +#loc54 = loc("Relu_59") +#loc55 = loc("Conv_60") +#loc56 = loc("Relu_61") +#loc57 = loc("GlobalAveragePool_62") +#loc58 = loc("Conv_63") +#loc59 = loc("Relu_64") +#loc60 = loc("Conv_65") +#loc61 = loc("Add_67") +#loc62 = loc("Clip_70") +#loc63 = loc("Div_72") +#loc64 = loc("Mul_73") +#loc65 = loc("Conv_74") +#loc66 = loc("Add_75") +#loc67 = loc("Conv_76") +#loc68 = loc("Relu_77") +#loc69 = loc("Conv_78") +#loc70 = loc("Relu_79") +#loc71 = loc("GlobalAveragePool_80") +#loc72 = loc("Conv_81") +#loc73 = loc("Relu_82") +#loc74 = loc("Conv_83") +#loc75 = loc("Add_85") +#loc76 = loc("Clip_88") +#loc77 = loc("Div_90") +#loc78 = loc("Mul_91") +#loc79 = loc("Conv_92") +#loc80 = loc("Add_93") +#loc81 = loc("Conv_94") +#loc82 = loc("Add_96") +#loc83 = loc("Clip_99") +#loc84 = loc("Div_101") +#loc85 = loc("Mul_102") +#loc86 = loc("Conv_103") +#loc87 = loc("Add_105") +#loc88 = loc("Clip_108") +#loc89 = loc("Div_110") +#loc90 = loc("Mul_111") +#loc91 = loc("Conv_112") +#loc92 = loc("Conv_113") +#loc93 = loc("Add_115") +#loc94 = loc("Clip_118") +#loc95 = loc("Div_120") +#loc96 = loc("Mul_121") +#loc97 = loc("Conv_122") +#loc98 = loc("Add_124") +#loc99 = loc("Clip_127") +#loc100 = loc("Div_129") +#loc101 = loc("Mul_130") +#loc102 = loc("Conv_131") +#loc103 = loc("Add_132") +#loc104 = loc("Conv_133") +#loc105 = loc("Add_135") +#loc106 = loc("Clip_138") +#loc107 = loc("Div_140") +#loc108 = loc("Mul_141") +#loc109 = loc("Conv_142") +#loc110 = loc("Add_144") +#loc111 = loc("Clip_147") +#loc112 = loc("Div_149") +#loc113 = loc("Mul_150") +#loc114 = loc("Conv_151") +#loc115 = loc("Add_152") +#loc116 = loc("Conv_153") +#loc117 = loc("Add_155") +#loc118 = loc("Clip_158") +#loc119 = loc("Div_160") +#loc120 = loc("Mul_161") +#loc121 = loc("Conv_162") +#loc122 = loc("Add_164") +#loc123 = loc("Clip_167") +#loc124 = loc("Div_169") +#loc125 = loc("Mul_170") +#loc126 = loc("Conv_171") +#loc127 = loc("Add_172") +#loc128 = loc("Conv_173") +#loc129 = loc("Add_175") +#loc130 = loc("Clip_178") +#loc131 = loc("Div_180") +#loc132 = loc("Mul_181") +#loc133 = loc("Conv_182") +#loc134 = loc("Add_184") +#loc135 = loc("Clip_187") +#loc136 = loc("Div_189") +#loc137 = loc("Mul_190") +#loc138 = loc("GlobalAveragePool_191") +#loc139 = loc("Conv_192") +#loc140 = loc("Relu_193") +#loc141 = loc("Conv_194") +#loc142 = loc("Add_196") +#loc143 = loc("Clip_199") +#loc144 = loc("Div_201") +#loc145 = loc("Mul_202") +#loc146 = loc("Conv_203") +#loc147 = loc("Conv_204") +#loc148 = loc("Add_206") +#loc149 = loc("Clip_209") +#loc150 = loc("Div_211") +#loc151 = loc("Mul_212") +#loc152 = loc("Conv_213") +#loc153 = loc("Add_215") +#loc154 = loc("Clip_218") +#loc155 = loc("Div_220") +#loc156 = loc("Mul_221") +#loc157 = loc("GlobalAveragePool_222") +#loc158 = loc("Conv_223") +#loc159 = loc("Relu_224") +#loc160 = loc("Conv_225") +#loc161 = loc("Add_227") +#loc162 = loc("Clip_230") +#loc163 = loc("Div_232") +#loc164 = loc("Mul_233") +#loc165 = loc("Conv_234") +#loc166 = loc("Add_235") +#loc167 = loc("Conv_236") +#loc168 = loc("Add_238") +#loc169 = loc("Clip_241") +#loc170 = loc("Div_243") +#loc171 = loc("Mul_244") +#loc172 = loc("Conv_245") +#loc173 = loc("Add_247") +#loc174 = loc("Clip_250") +#loc175 = loc("Div_252") +#loc176 = loc("Mul_253") +#loc177 = loc("GlobalAveragePool_254") +#loc178 = loc("Conv_255") +#loc179 = loc("Relu_256") +#loc180 = loc("Conv_257") +#loc181 = loc("Add_259") +#loc182 = loc("Clip_262") +#loc183 = loc("Div_264") +#loc184 = loc("Mul_265") +#loc185 = loc("Conv_266") +#loc186 = loc("Conv_267") +#loc187 = loc("Add_269") +#loc188 = loc("Clip_272") +#loc189 = loc("Div_274") +#loc190 = loc("Mul_275") +#loc191 = loc("Conv_276") +#loc192 = loc("Add_278") +#loc193 = loc("Clip_281") +#loc194 = loc("Div_283") +#loc195 = loc("Mul_284") +#loc196 = loc("GlobalAveragePool_285") +#loc197 = loc("Conv_286") +#loc198 = loc("Relu_287") +#loc199 = loc("Conv_288") +#loc200 = loc("Add_290") +#loc201 = loc("Clip_293") +#loc202 = loc("Div_295") +#loc203 = loc("Mul_296") +#loc204 = loc("Conv_297") +#loc205 = loc("Add_298") +#loc206 = loc("Conv_299") +#loc207 = loc("Add_301") +#loc208 = loc("Clip_304") +#loc209 = loc("Div_306") +#loc210 = loc("Mul_307") +#loc211 = loc("Conv_308") +#loc212 = loc("Add_310") +#loc213 = loc("Clip_313") +#loc214 = loc("Div_315") +#loc215 = loc("Mul_316") +#loc216 = loc("GlobalAveragePool_317") +#loc217 = loc("Conv_318") +#loc218 = loc("Relu_319") +#loc219 = loc("Conv_320") +#loc220 = loc("Add_322") +#loc221 = loc("Clip_325") +#loc222 = loc("Div_327") +#loc223 = loc("Mul_328") +#loc224 = loc("Conv_329") +#loc225 = loc("Add_330") +#loc226 = loc("Conv_331") +#loc227 = loc("Add_333") +#loc228 = loc("Clip_336") +#loc229 = loc("Div_338") +#loc230 = loc("Mul_339") +#loc231 = loc("GlobalAveragePool_342") +#loc232 = loc("Conv_343") +#loc233 = loc("Sigmoid_344") +#loc234 = loc("Mul_345") +#loc235 = loc("Conv_340") +#loc236 = loc("Relu_341") +#loc237 = loc("Split_349") +#loc238 = loc("Concat_350") +#loc239 = loc("Conv_351") +#loc240 = loc("Sigmoid_352") +#loc241 = loc("Split_353") +#loc242 = loc("Mul_360") +#loc243 = loc("Mul_354") +#loc244 = loc("Concat_355") +#loc245 = loc("Conv_356") +#loc246 = loc("Tanh_357") +#loc247 = loc("Mul_361") +#loc248 = loc("Add_362") +#loc249 = loc("Concat_363") +#loc250 = loc("Resize_365") +#loc251 = loc("Slice_371") +#loc252 = loc("Concat_372") +#loc253 = loc("Conv_373") +#loc254 = loc("Relu_374") +#loc255 = loc("Split_375") +#loc256 = loc("Concat_376") +#loc257 = loc("Conv_377") +#loc258 = loc("Sigmoid_378") +#loc259 = loc("Split_379") +#loc260 = loc("Mul_386") +#loc261 = loc("Mul_380") +#loc262 = loc("Concat_381") +#loc263 = loc("Conv_382") +#loc264 = loc("Tanh_383") +#loc265 = loc("Mul_387") +#loc266 = loc("Add_388") +#loc267 = loc("Concat_389") +#loc268 = loc("Resize_391") +#loc269 = loc("Slice_397") +#loc270 = loc("Concat_398") +#loc271 = loc("Conv_399") +#loc272 = loc("Relu_400") +#loc273 = loc("Split_401") +#loc274 = loc("Concat_402") +#loc275 = loc("Conv_403") +#loc276 = loc("Sigmoid_404") +#loc277 = loc("Split_405") +#loc278 = loc("Mul_412") +#loc279 = loc("Mul_406") +#loc280 = loc("Concat_407") +#loc281 = loc("Conv_408") +#loc282 = loc("Tanh_409") +#loc283 = loc("Mul_413") +#loc284 = loc("Add_414") +#loc285 = loc("Concat_415") +#loc286 = loc("Resize_417") +#loc287 = loc("Concat_418") +#loc288 = loc("Conv_419") +#loc289 = loc("Relu_420") +#loc290 = loc("Split_421") +#loc291 = loc("Concat_422") +#loc292 = loc("Conv_423") +#loc293 = loc("Sigmoid_424") +#loc294 = loc("Split_425") +#loc295 = loc("Mul_432") +#loc296 = loc("Mul_426") +#loc297 = loc("Concat_427") +#loc298 = loc("Conv_428") +#loc299 = loc("Tanh_429") +#loc300 = loc("Mul_433") +#loc301 = loc("Add_434") +#loc302 = loc("Transpose_448") +#loc303 = loc("Transpose_449") +#loc304 = loc("Transpose_450") +#loc305 = loc("Transpose_451") +#loc306 = loc("Concat_435") +#loc307 = loc("Resize_437") +#loc308 = loc("Concat_438") +#loc309 = loc("Conv_439") +#loc310 = loc("Relu_440") +#loc311 = loc("Conv_441") +#loc312 = loc("Relu_442") +#loc313 = loc("Conv_443") +#loc314 = loc("Split_444") +#loc315 = loc("Clip_446") +#loc316 = loc("Transpose_452") +#loc317 = loc("Clip_447") +#loc318 = loc("Transpose_453") +#loc319 = loc(fused[#loc6, #loc7]) +#loc320 = loc(fused[#loc15, #loc13, #loc16]) +#loc321 = loc(fused[#loc13, #loc16, #loc15]) +#loc322 = loc(fused[#loc25, #loc26]) +#loc323 = loc(fused[#loc27, #loc28]) +#loc324 = loc(fused[#loc29, #loc30]) +#loc325 = loc(fused[#loc31, #loc32]) +#loc326 = loc(fused[#loc34, #loc35]) +#loc327 = loc(fused[#loc36, #loc37]) +#loc328 = loc(fused[#loc38, #loc39]) +#loc329 = loc(fused[#loc40, #loc41]) +#loc330 = loc(fused[#loc42, #loc43]) +#loc331 = loc(fused[#loc45, #loc46]) +#loc332 = loc(fused[#loc53, #loc54]) +#loc333 = loc(fused[#loc55, #loc56]) +#loc334 = loc(fused[#loc58, #loc59]) +#loc335 = loc(fused[#loc65, #loc66]) +#loc336 = loc(fused[#loc67, #loc68]) +#loc337 = loc(fused[#loc69, #loc70]) +#loc338 = loc(fused[#loc72, #loc73]) +#loc339 = loc(fused[#loc79, #loc80]) +#loc340 = loc(fused[#loc102, #loc103]) +#loc341 = loc(fused[#loc114, #loc115]) +#loc342 = loc(fused[#loc126, #loc127]) +#loc343 = loc(fused[#loc137, #loc138]) +#loc344 = loc(fused[#loc139, #loc140]) +#loc345 = loc(fused[#loc156, #loc157]) +#loc346 = loc(fused[#loc158, #loc159]) +#loc347 = loc(fused[#loc165, #loc166]) +#loc348 = loc(fused[#loc176, #loc177]) +#loc349 = loc(fused[#loc178, #loc179]) +#loc350 = loc(fused[#loc195, #loc196]) +#loc351 = loc(fused[#loc197, #loc198]) +#loc352 = loc(fused[#loc204, #loc205]) +#loc353 = loc(fused[#loc215, #loc216]) +#loc354 = loc(fused[#loc217, #loc218]) +#loc355 = loc(fused[#loc224, #loc225]) +#loc356 = loc(fused[#loc230, #loc231]) +#loc357 = loc(fused[#loc235, #loc236, #loc234]) +#loc358 = loc(fused[#loc235, #loc236]) +#loc359 = loc(fused[#loc253, #loc254]) +#loc360 = loc(fused[#loc271, #loc272]) +#loc361 = loc(fused[#loc288, #loc289]) +#loc362 = loc(fused[#loc309, #loc310]) +#loc363 = loc(fused[#loc311, #loc312]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.onnx.mlir b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.onnx.mlir new file mode 100644 index 0000000000000000000000000000000000000000..c29a5976ab3b6811592e178ea52a76b985a068c8 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.onnx.mlir @@ -0,0 +1,1627 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func @main_graph(%arg0: tensor<1x180x320x4xui8> {onnx.name = "src"} loc(unknown), %arg1: tensor<1x90x160x16xf32> {onnx.name = "r1i"} loc(unknown), %arg2: tensor<1x45x80x20xf32> {onnx.name = "r2i"} loc(unknown), %arg3: tensor<1x23x40x40xf32> {onnx.name = "r3i"} loc(unknown), %arg4: tensor<1x12x20x64xf32> {onnx.name = "r4i"} loc(unknown)) -> (tensor<1x180x320x3xf32> {onnx.name = "fgr"}, tensor<1x180x320x1xf32> {onnx.name = "pha"}, tensor<1x90x160x16xf32> {onnx.name = "r1o"}, tensor<1x45x80x20xf32> {onnx.name = "r2o"}, tensor<1x23x40x40xf32> {onnx.name = "r3o"}, tensor<1x12x20x64xf32> {onnx.name = "r4o"}) { + %0 = onnx.Constant dense<[[[-4.850000e-01]], [[-4.560000e-01]], [[-4.060000e-01]]]> : tensor<3x1x1xf32> loc(#loc) + %1 = onnx.Constant dense<[3, 1]> : tensor<2xi64> loc(#loc) + %2 = onnx.Constant dense<16> : tensor<2xi64> loc(#loc) + %3 = onnx.Constant dense<20> : tensor<2xi64> loc(#loc) + %4 = onnx.Constant dense<40> : tensor<2xi64> loc(#loc) + %5 = onnx.Constant dense<64> : tensor<2xi64> loc(#loc) + %6 = "onnx.NoValue"() {onnx_node_name = "onnx.NoValue_26", value} : () -> none loc(#loc) + %7 = onnx.Constant dense_resource<__elided__> : tensor<112xf32> loc(#loc1) + %8 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc2) + %9 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc3) + %10 = onnx.Constant dense_resource<__elided__> : tensor<24xf32> loc(#loc4) + %11 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc5) + %12 = onnx.Constant dense_resource<__elided__> : tensor<184xf32> loc(#loc6) + %13 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc7) + %14 = onnx.Constant dense<3> : tensor<1xi64> loc(#loc8) + %15 = onnx.Constant dense_resource<__elided__> : tensor<128x960x1x1xf32> loc(#loc9) + %16 = onnx.Constant dense<2.550000e+02> : tensor loc(#loc10) + %17 = onnx.Constant dense_resource<__elided__> : tensor<32xf32> loc(#loc11) + %18 = onnx.Constant dense<[-0.257117867, -0.332320929, -0.342930794, 0.882337093, -0.811691761, 1.04650748, 1.993430e-01, 0.471133053, 0.0722430944, 0.554342687, 1.3374486, 0.48697716, 1.31853354, 0.714223623, 1.16618729, 0.738572299]> : tensor<16xf32> loc(#loc12) + %19 = onnx.Constant dense_resource<__elided__> : tensor<16x16x3x3xf32> loc(#loc13) + %20 = onnx.Constant dense<[0.281471759, -0.0896756947, 0.0517602414, -0.266139954, 0.132527292, 0.684469878, -0.0511226803, 0.859402895, 0.504835129, 0.569725394, 0.217058718, -0.0543790609, -0.30986914, 0.451566547, 0.166573063, 0.415171683]> : tensor<16xf32> loc(#loc14) + %21 = onnx.Constant dense_resource<__elided__> : tensor<24x64x1x1xf32> loc(#loc15) + %22 = onnx.Constant dense_resource<__elided__> : tensor<16x35x3x3xf32> loc(#loc16) + %23 = onnx.Constant dense_resource<__elided__> : tensor<32xf32> loc(#loc17) + %24 = onnx.Constant dense_resource<__elided__> : tensor<32x59x3x3xf32> loc(#loc18) + %25 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc19) + %26 = onnx.Constant dense_resource<__elided__> : tensor<128x960x1x1xf32> loc(#loc20) + %27 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc21) + %28 = onnx.Constant dense_resource<__elided__> : tensor<80x171x3x3xf32> loc(#loc22) + %29 = onnx.Constant dense<[[[2.290000e-01]], [[2.240000e-01]], [[2.250000e-01]]]> : tensor<3x1x1xf32> loc(#loc23) + %30 = onnx.Constant dense_resource<__elided__> : tensor<960x160x1x1xf32> loc(#loc24) + %31 = onnx.Constant dense_resource<__elided__> : tensor<160x960x1x1xf32> loc(#loc25) + %32 = onnx.Constant dense_resource<__elided__> : tensor<960x1x5x5xf32> loc(#loc26) + %33 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc27) + %34 = onnx.Constant dense_resource<__elided__> : tensor<960x240x1x1xf32> loc(#loc28) + %35 = onnx.Constant dense_resource<__elided__> : tensor<160xf32> loc(#loc29) + %36 = onnx.Constant dense_resource<__elided__> : tensor<960x240x1x1xf32> loc(#loc30) + %37 = onnx.Constant dense_resource<__elided__> : tensor<960x1x5x5xf32> loc(#loc31) + %38 = onnx.Constant dense_resource<__elided__> : tensor<160xf32> loc(#loc32) + %39 = onnx.Constant dense_resource<__elided__> : tensor<160x672x1x1xf32> loc(#loc33) + %40 = onnx.Constant dense_resource<__elided__> : tensor<24xf32> loc(#loc34) + %41 = onnx.Constant dense_resource<__elided__> : tensor<120x40x1x1xf32> loc(#loc35) + %42 = onnx.Constant dense_resource<__elided__> : tensor<72x1x5x5xf32> loc(#loc36) + %43 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc37) + %44 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc38) + %45 = onnx.Constant dense_resource<__elided__> : tensor<80x200x1x1xf32> loc(#loc39) + %46 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc40) + %47 = onnx.Constant dense_resource<__elided__> : tensor<240x40x1x1xf32> loc(#loc41) + %48 = onnx.Constant dense_resource<__elided__> : tensor<240x1x3x3xf32> loc(#loc42) + %49 = onnx.Constant dense_resource<__elided__> : tensor<480x120x1x1xf32> loc(#loc43) + %50 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc44) + %51 = onnx.Constant dense_resource<__elided__> : tensor<184xf32> loc(#loc45) + %52 = onnx.Constant dense_resource<__elided__> : tensor<32x32x3x3xf32> loc(#loc46) + %53 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc47) + %54 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc48) + %55 = onnx.Constant dense_resource<__elided__> : tensor<200x80x1x1xf32> loc(#loc49) + %56 = onnx.Constant dense_resource<__elided__> : tensor<184x1x3x3xf32> loc(#loc50) + %57 = onnx.Constant dense_resource<__elided__> : tensor<200x1x3x3xf32> loc(#loc51) + %58 = onnx.Constant dense_resource<__elided__> : tensor<72xf32> loc(#loc52) + %59 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc53) + %60 = onnx.Constant dense_resource<__elided__> : tensor<480xf32> loc(#loc54) + %61 = onnx.Constant dense_resource<__elided__> : tensor<16x16x1x1xf32> loc(#loc55) + %62 = onnx.Constant dense_resource<__elided__> : tensor<64x128x3x3xf32> loc(#loc56) + %63 = onnx.Constant dense_resource<__elided__> : tensor<480xf32> loc(#loc57) + %64 = onnx.Constant dense_resource<__elided__> : tensor<40x120x1x1xf32> loc(#loc58) + %65 = onnx.Constant dense<[1.000000e+00, 1.000000e+00, 2.000000e+00, 2.000000e+00]> : tensor<4xf32> loc(#loc59) + %66 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc60) + %67 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc61) + %68 = onnx.Constant dense_resource<__elided__> : tensor<16x32x3x3xf32> loc(#loc62) + %69 = onnx.Constant dense<0.000000e+00> : tensor loc(#loc63) + %70 = onnx.Constant dense<[0.00409470545, 0.00284675183, 0.00200544903, 0.124928087]> : tensor<4xf32> loc(#loc64) + %71 = onnx.Constant dense<3.000000e+00> : tensor loc(#loc65) + %72 = onnx.Constant dense_resource<__elided__> : tensor<72xf32> loc(#loc66) + %73 = onnx.Constant dense_resource<__elided__> : tensor<200xf32> loc(#loc67) + %74 = onnx.Constant dense_resource<__elided__> : tensor<72x1x3x3xf32> loc(#loc68) + %75 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc69) + %76 = onnx.Constant dense_resource<__elided__> : tensor<120x1x5x5xf32> loc(#loc70) + %77 = onnx.Constant dense_resource<__elided__> : tensor<240xf32> loc(#loc71) + %78 = onnx.Constant dense_resource<__elided__> : tensor<240xf32> loc(#loc72) + %79 = onnx.Constant dense_resource<__elided__> : tensor<672x112x1x1xf32> loc(#loc73) + %80 = onnx.Constant dense_resource<__elided__> : tensor<120x40x1x1xf32> loc(#loc74) + %81 = onnx.Constant dense<23> : tensor<1xi64> loc(#loc75) + %82 = onnx.Constant dense_resource<__elided__> : tensor<16x3x3x3xf32> loc(#loc76) + %83 = onnx.Constant dense_resource<__elided__> : tensor<40x120x1x1xf32> loc(#loc77) + %84 = onnx.Constant dense<2> : tensor<1xi64> loc(#loc78) + %85 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc79) + %86 = onnx.Constant dense_resource<__elided__> : tensor<64xf32> loc(#loc80) + %87 = onnx.Constant dense_resource<__elided__> : tensor<184x1x3x3xf32> loc(#loc81) + %88 = onnx.Constant dense_resource<__elided__> : tensor<120xf32> loc(#loc82) + %89 = onnx.Constant dense<1> : tensor<1xi64> loc(#loc83) + %90 = onnx.Constant dense_resource<__elided__> : tensor<40x107x3x3xf32> loc(#loc84) + %91 = onnx.Constant dense_resource<__elided__> : tensor<672x168x1x1xf32> loc(#loc85) + %92 = onnx.Constant dense<6.000000e+00> : tensor loc(#loc86) + %93 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc87) + %94 = onnx.Constant dense_resource<__elided__> : tensor<168xf32> loc(#loc88) + %95 = onnx.Constant dense_resource<__elided__> : tensor<112x672x1x1xf32> loc(#loc89) + %96 = onnx.Constant dense_resource<__elided__> : tensor<112x480x1x1xf32> loc(#loc90) + %97 = onnx.Constant dense_resource<__elided__> : tensor<112xf32> loc(#loc91) + %98 = onnx.Constant dense<1.000000e+00> : tensor loc(#loc92) + %99 = onnx.Constant dense_resource<__elided__> : tensor<24x72x1x1xf32> loc(#loc93) + %100 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc94) + %101 = onnx.Constant dense_resource<__elided__> : tensor<120x32x1x1xf32> loc(#loc95) + %102 = onnx.Constant dense_resource<__elided__> : tensor<32xf32> loc(#loc96) + %103 = onnx.Constant dense_resource<__elided__> : tensor<184xf32> loc(#loc97) + %104 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc98) + %105 = onnx.Constant dense<0> : tensor<1xi64> loc(#loc99) + %106 = onnx.Constant dense_resource<__elided__> : tensor<120x480x1x1xf32> loc(#loc100) + %107 = onnx.Constant dense_resource<__elided__> : tensor<672x1x5x5xf32> loc(#loc101) + %108 = onnx.Constant dense_resource<__elided__> : tensor<184xf32> loc(#loc102) + %109 = onnx.Constant dense_resource<__elided__> : tensor<184x80x1x1xf32> loc(#loc103) + %110 = onnx.Constant dense_resource<__elided__> : tensor<480x80x1x1xf32> loc(#loc104) + %111 = onnx.Constant dense_resource<__elided__> : tensor<80x184x1x1xf32> loc(#loc105) + %112 = onnx.Constant dense_resource<__elided__> : tensor<672x168x1x1xf32> loc(#loc106) + %113 = onnx.Constant dense_resource<__elided__> : tensor<240xf32> loc(#loc107) + %114 = onnx.Constant dense_resource<__elided__> : tensor<80x80x3x3xf32> loc(#loc108) + %115 = onnx.Constant dense_resource<__elided__> : tensor<4x16x1x1xf32> loc(#loc109) + %116 = onnx.Constant dense_resource<__elided__> : tensor<72x24x1x1xf32> loc(#loc110) + %117 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc111) + %118 = onnx.Constant dense_resource<__elided__> : tensor<64x1x3x3xf32> loc(#loc112) + %119 = onnx.Constant dense_resource<__elided__> : tensor<240x960x1x1xf32> loc(#loc113) + %120 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc114) + %121 = onnx.Constant dense_resource<__elided__> : tensor<64xf32> loc(#loc115) + %122 = onnx.Constant dense_resource<__elided__> : tensor<80x184x1x1xf32> loc(#loc116) + %123 = onnx.Constant dense_resource<__elided__> : tensor<480x1x3x3xf32> loc(#loc117) + %124 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc118) + %125 = onnx.Constant dense<[-1.31068802, 0.586562276, 5.67538071, 0.551027656, 2.19523954, 3.83854461, 0.0600251146, -2.18778157, -1.5404067, 2.044780e+00, -4.23846388, 0.703142225, -8.39978456E-5, 3.50620365, -0.531753063, -5.91183185]> : tensor<16xf32> loc(#loc119) + %126 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc120) + %127 = onnx.Constant dense_resource<__elided__> : tensor<128x128x3x3xf32> loc(#loc121) + %128 = onnx.Constant dense_resource<__elided__> : tensor<672x112x1x1xf32> loc(#loc122) + %129 = onnx.Constant dense_resource<__elided__> : tensor<32x120x1x1xf32> loc(#loc123) + %130 = onnx.Constant dense_resource<__elided__> : tensor<32xf32> loc(#loc124) + %131 = onnx.Constant dense_resource<__elided__> : tensor<64x16x1x1xf32> loc(#loc125) + %132 = onnx.Constant dense_resource<__elided__> : tensor<672x1x3x3xf32> loc(#loc126) + %133 = onnx.Constant dense_resource<__elided__> : tensor<168x672x1x1xf32> loc(#loc127) + %134 = onnx.Constant dense_resource<__elided__> : tensor<168x672x1x1xf32> loc(#loc128) + %135 = onnx.Constant dense_resource<__elided__> : tensor<200xf32> loc(#loc129) + %136 = onnx.Constant dense_resource<__elided__> : tensor<480xf32> loc(#loc130) + %137 = onnx.Constant dense_resource<__elided__> : tensor<120x32x1x1xf32> loc(#loc131) + %138 = onnx.Constant dense_resource<__elided__> : tensor<672xf32> loc(#loc132) + %139 = onnx.Constant dense<45> : tensor<1xi64> loc(#loc133) + %140 = onnx.Constant dense_resource<__elided__> : tensor<80xf32> loc(#loc134) + %141 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc135) + %142 = onnx.Constant dense_resource<__elided__> : tensor<24x72x1x1xf32> loc(#loc136) + %143 = onnx.Constant dense_resource<__elided__> : tensor<240xf32> loc(#loc137) + %144 = onnx.Constant dense_resource<__elided__> : tensor<40x72x1x1xf32> loc(#loc138) + %145 = onnx.Constant dense_resource<__elided__> : tensor<168xf32> loc(#loc139) + %146 = onnx.Constant dense_resource<__elided__> : tensor<32x120x1x1xf32> loc(#loc140) + %147 = onnx.Constant dense_resource<__elided__> : tensor<128xf32> loc(#loc141) + %148 = onnx.Constant dense_resource<__elided__> : tensor<128xf32> loc(#loc142) + %149 = onnx.Constant dense_resource<__elided__> : tensor<72xf32> loc(#loc143) + %150 = onnx.Constant dense_resource<__elided__> : tensor<72xf32> loc(#loc144) + %151 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc145) + %152 = onnx.Constant dense_resource<__elided__> : tensor<160xf32> loc(#loc146) + %153 = onnx.Constant dense_resource<__elided__> : tensor<80x240x1x1xf32> loc(#loc147) + %154 = onnx.Constant dense_resource<__elided__> : tensor<20x40x3x3xf32> loc(#loc148) + %155 = onnx.Constant dense_resource<__elided__> : tensor<184x80x1x1xf32> loc(#loc149) + %156 = onnx.Constant dense_resource<__elided__> : tensor<160x960x1x1xf32> loc(#loc150) + %157 = onnx.Constant dense_resource<__elided__> : tensor<24xf32> loc(#loc151) + %158 = onnx.Constant dense_resource<__elided__> : tensor<960xf32> loc(#loc152) + %159 = onnx.Constant dense_resource<__elided__> : tensor<40x80x3x3xf32> loc(#loc153) + %160 = onnx.Constant dense_resource<__elided__> : tensor<72x24x1x1xf32> loc(#loc154) + %161 = onnx.Constant dense<[-4.38406658, -1.06764766E-8, -0.704851329, -1.05036237E-8, -4.89120433E-9, 1.53249037, -0.0617836975, 2.16366434, 0.0416259095, -4.12739087E-9, -3.50249429E-9, -7.75795516E-9, -4.04315559E-9, 0.292217016, -0.010752866, 1.63358212]> : tensor<16xf32> loc(#loc155) + %162 = onnx.Constant dense_resource<__elided__> : tensor<40xf32> loc(#loc156) + %163 = onnx.Constant dense_resource<__elided__> : tensor<40x40x3x3xf32> loc(#loc157) + %164 = onnx.Constant dense_resource<__elided__> : tensor<960x160x1x1xf32> loc(#loc158) + %165 = onnx.Constant dense_resource<__elided__> : tensor<16x1x3x3xf32> loc(#loc159) + %166 = onnx.Constant dense_resource<__elided__> : tensor<72x24x1x1xf32> loc(#loc160) + %167 = onnx.Constant dense_resource<__elided__> : tensor<240x960x1x1xf32> loc(#loc161) + %168 = onnx.Constant dense_resource<__elided__> : tensor<20xf32> loc(#loc162) + %169 = onnx.Constant dense_resource<__elided__> : tensor<72xf32> loc(#loc163) + %170 = onnx.Constant dense<[0.00544366054, 0.154367775, 0.115729354, 0.171141103, -0.168815523, 0.0456937179, 0.188233331, 0.0151384082, 0.242783383, -0.139173314, -0.24988465, -9.479440e-02, -0.055940561, -0.0512795448, -0.0738077834, 0.0476587117]> : tensor<16xf32> loc(#loc164) + %171 = onnx.Constant dense_resource<__elided__> : tensor<120x1x5x5xf32> loc(#loc165) + %172 = onnx.Constant dense_resource<__elided__> : tensor<960x160x1x1xf32> loc(#loc166) + %173 = onnx.Constant dense<[2.98861408, -1.22985208, 2.43826318, -3.98499513, 4.62797928, 2.54142761, 2.45345306, 2.64061832, 2.13576674, 2.30800247, -0.198341176, -0.427822977, -1.09159482, 4.85548782, 2.70597649, 2.6902504]> : tensor<16xf32> loc(#loc167) + %174 = onnx.Constant dense_resource<__elided__> : tensor<64xf32> loc(#loc168) + %175 = "onnx.Transpose"(%arg1) {onnx_node_name = "Transpose_9", perm = [0, 3, 1, 2]} : (tensor<1x90x160x16xf32>) -> tensor<1x16x90x160xf32> loc(#loc169) + %176 = "onnx.Transpose"(%arg2) {onnx_node_name = "Transpose_10", perm = [0, 3, 1, 2]} : (tensor<1x45x80x20xf32>) -> tensor<1x20x45x80xf32> loc(#loc170) + %177 = "onnx.Transpose"(%arg3) {onnx_node_name = "Transpose_11", perm = [0, 3, 1, 2]} : (tensor<1x23x40x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc171) + %178 = "onnx.Cast"(%arg0) { + onnx_node_name = "Cast_0", + saturate = 1 : si64, + to = f32} : (tensor<1x180x320x4xui8>) -> tensor<1x180x320x4xf32> loc(#loc172) + %179 = "onnx.Div"(%178, %16) {onnx_node_name = "Div_2"} : (tensor<1x180x320x4xf32>, tensor) -> tensor<1x180x320x4xf32> loc(#loc173) + %180 = "onnx.Slice"(%179, %105, %14, %14, %89) {onnx_node_name = "Slice_7"} : (tensor<1x180x320x4xf32>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>) -> tensor<1x180x320x3xf32> loc(#loc174) + %181 = "onnx.Transpose"(%180) {onnx_node_name = "Transpose_8", perm = [0, 3, 1, 2]} : (tensor<1x180x320x3xf32>) -> tensor<1x3x180x320xf32> loc(#loc175) + %182 = "onnx.AveragePool"(%181) { + auto_pad = "NOTSET", + ceil_mode = 1 : si64, + count_include_pad = 0 : si64, + kernel_shape = [2, 2], + onnx_node_name = "AveragePool_346", + pads = [0, 0, 0, 0], + strides = [2, 2]} : (tensor<1x3x180x320xf32>) -> tensor<1x3x90x160xf32> loc(#loc176) + %183 = "onnx.AveragePool"(%182) { + auto_pad = "NOTSET", + ceil_mode = 1 : si64, + count_include_pad = 0 : si64, + kernel_shape = [2, 2], + onnx_node_name = "AveragePool_347", + pads = [0, 0, 0, 0], + strides = [2, 2]} : (tensor<1x3x90x160xf32>) -> tensor<1x3x45x80xf32> loc(#loc177) + %184 = "onnx.AveragePool"(%183) { + auto_pad = "NOTSET", + ceil_mode = 1 : si64, + count_include_pad = 0 : si64, + kernel_shape = [2, 2], + onnx_node_name = "AveragePool_348", + pads = [0, 0, 0, 0], + strides = [2, 2]} : (tensor<1x3x45x80xf32>) -> tensor<1x3x23x40xf32> loc(#loc178) + %185 = "onnx.Transpose"(%arg4) {onnx_node_name = "Transpose_12", perm = [0, 3, 1, 2]} : (tensor<1x12x20x64xf32>) -> tensor<1x64x12x20xf32> loc(#loc179) + %186 = "onnx.Add"(%181, %0) {onnx_node_name = "Sub_14-Initializer_398_48"} : (tensor<1x3x180x320xf32>, tensor<3x1x1xf32>) -> tensor<1x3x180x320xf32> loc(#loc487) + %187 = "onnx.Div"(%186, %29) {onnx_node_name = "Div_16"} : (tensor<1x3x180x320xf32>, tensor<3x1x1xf32>) -> tensor<1x3x180x320xf32> loc(#loc182) + %188 = "onnx.Conv"(%187, %82, %173) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_17", + pads = [1, 1, 1, 1], + strides = [2, 2]} : (tensor<1x3x180x320xf32>, tensor<16x3x3x3xf32>, tensor<16xf32>) -> tensor<1x16x90x160xf32> loc(#loc183) + %189 = "onnx.Add"(%188, %71) {onnx_node_name = "Add_19"} : (tensor<1x16x90x160xf32>, tensor) -> tensor<1x16x90x160xf32> loc(#loc184) + %190 = "onnx.Clip"(%189, %69, %92) {onnx_node_name = "Clip_22_50"} : (tensor<1x16x90x160xf32>, tensor, tensor) -> tensor<1x16x90x160xf32> loc(#loc185) + %191 = "onnx.Div"(%190, %92) {onnx_node_name = "Div_24"} : (tensor<1x16x90x160xf32>, tensor) -> tensor<1x16x90x160xf32> loc(#loc186) + %192 = "onnx.Mul"(%188, %191) {onnx_node_name = "Mul_25"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc187) + %193 = "onnx.Conv"(%192, %165, %161) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 16 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_26", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x16x90x160xf32>, tensor<16x1x3x3xf32>, tensor<16xf32>) -> tensor<1x16x90x160xf32> loc(#loc188) + %194 = "onnx.Relu"(%193) {onnx_node_name = "Relu_27"} : (tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc189) + %195 = "onnx.Conv"(%194, %61, %125) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_28", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x16x90x160xf32>, tensor<16x16x1x1xf32>, tensor<16xf32>) -> tensor<1x16x90x160xf32> loc(#loc190) + %196 = "onnx.Add"(%195, %192) {onnx_node_name = "Add_29"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc191) + %197 = "onnx.Conv"(%196, %131, %174) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_30", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x16x90x160xf32>, tensor<64x16x1x1xf32>, tensor<64xf32>) -> tensor<1x64x90x160xf32> loc(#loc192) + %198 = "onnx.Relu"(%197) {onnx_node_name = "Relu_31"} : (tensor<1x64x90x160xf32>) -> tensor<1x64x90x160xf32> loc(#loc193) + %199 = "onnx.Conv"(%198, %118, %86) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 64 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_32", + pads = [1, 1, 1, 1], + strides = [2, 2]} : (tensor<1x64x90x160xf32>, tensor<64x1x3x3xf32>, tensor<64xf32>) -> tensor<1x64x45x80xf32> loc(#loc194) + %200 = "onnx.Relu"(%199) {onnx_node_name = "Relu_33"} : (tensor<1x64x45x80xf32>) -> tensor<1x64x45x80xf32> loc(#loc195) + %201 = "onnx.Conv"(%200, %21, %10) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_34", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x64x45x80xf32>, tensor<24x64x1x1xf32>, tensor<24xf32>) -> tensor<1x24x45x80xf32> loc(#loc196) + %202 = "onnx.Conv"(%201, %116, %72) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_35", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x24x45x80xf32>, tensor<72x24x1x1xf32>, tensor<72xf32>) -> tensor<1x72x45x80xf32> loc(#loc197) + %203 = "onnx.Relu"(%202) {onnx_node_name = "Relu_36"} : (tensor<1x72x45x80xf32>) -> tensor<1x72x45x80xf32> loc(#loc198) + %204 = "onnx.Conv"(%203, %74, %58) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 72 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_37", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x72x45x80xf32>, tensor<72x1x3x3xf32>, tensor<72xf32>) -> tensor<1x72x45x80xf32> loc(#loc199) + %205 = "onnx.Relu"(%204) {onnx_node_name = "Relu_38"} : (tensor<1x72x45x80xf32>) -> tensor<1x72x45x80xf32> loc(#loc200) + %206 = "onnx.Conv"(%205, %99, %40) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_39", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x72x45x80xf32>, tensor<24x72x1x1xf32>, tensor<24xf32>) -> tensor<1x24x45x80xf32> loc(#loc201) + %207 = "onnx.Add"(%206, %201) {onnx_node_name = "Add_40"} : (tensor<1x24x45x80xf32>, tensor<1x24x45x80xf32>) -> tensor<1x24x45x80xf32> loc(#loc202) + %208 = "onnx.Conv"(%207, %166, %169) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_41", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x24x45x80xf32>, tensor<72x24x1x1xf32>, tensor<72xf32>) -> tensor<1x72x45x80xf32> loc(#loc203) + %209 = "onnx.Relu"(%208) {onnx_node_name = "Relu_42"} : (tensor<1x72x45x80xf32>) -> tensor<1x72x45x80xf32> loc(#loc204) + %210 = "onnx.Conv"(%209, %42, %149) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 72 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_43", + pads = [2, 2, 2, 2], + strides = [2, 2]} : (tensor<1x72x45x80xf32>, tensor<72x1x5x5xf32>, tensor<72xf32>) -> tensor<1x72x23x40xf32> loc(#loc205) + %211 = "onnx.Relu"(%210) {onnx_node_name = "Relu_44"} : (tensor<1x72x23x40xf32>) -> tensor<1x72x23x40xf32> loc(#loc206) + %212 = "onnx.ReduceMeanV13"(%211) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_45_12"} : (tensor<1x72x23x40xf32>) -> tensor<1x72x1x1xf32> loc(#loc207) + %213 = "onnx.Conv"(%212, %142, %157) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_46", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x72x1x1xf32>, tensor<24x72x1x1xf32>, tensor<24xf32>) -> tensor<1x24x1x1xf32> loc(#loc208) + %214 = "onnx.Relu"(%213) {onnx_node_name = "Relu_47"} : (tensor<1x24x1x1xf32>) -> tensor<1x24x1x1xf32> loc(#loc209) + %215 = "onnx.Conv"(%214, %160, %150) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_48", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x24x1x1xf32>, tensor<72x24x1x1xf32>, tensor<72xf32>) -> tensor<1x72x1x1xf32> loc(#loc210) + %216 = "onnx.Add"(%215, %71) {onnx_node_name = "Add_50"} : (tensor<1x72x1x1xf32>, tensor) -> tensor<1x72x1x1xf32> loc(#loc211) + %217 = "onnx.Clip"(%216, %69, %92) {onnx_node_name = "Clip_53_39"} : (tensor<1x72x1x1xf32>, tensor, tensor) -> tensor<1x72x1x1xf32> loc(#loc212) + %218 = "onnx.Div"(%217, %92) {onnx_node_name = "Div_55"} : (tensor<1x72x1x1xf32>, tensor) -> tensor<1x72x1x1xf32> loc(#loc213) + %219 = "onnx.Mul"(%218, %211) {onnx_node_name = "Mul_56"} : (tensor<1x72x1x1xf32>, tensor<1x72x23x40xf32>) -> tensor<1x72x23x40xf32> loc(#loc214) + %220 = "onnx.Conv"(%219, %144, %9) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_57", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x72x23x40xf32>, tensor<40x72x1x1xf32>, tensor<40xf32>) -> tensor<1x40x23x40xf32> loc(#loc215) + %221 = "onnx.Conv"(%220, %80, %11) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_58", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x40x23x40xf32>, tensor<120x40x1x1xf32>, tensor<120xf32>) -> tensor<1x120x23x40xf32> loc(#loc216) + %222 = "onnx.Relu"(%221) {onnx_node_name = "Relu_59"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc217) + %223 = "onnx.Conv"(%222, %171, %75) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 120 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_60", + pads = [2, 2, 2, 2], + strides = [1, 1]} : (tensor<1x120x23x40xf32>, tensor<120x1x5x5xf32>, tensor<120xf32>) -> tensor<1x120x23x40xf32> loc(#loc218) + %224 = "onnx.Relu"(%223) {onnx_node_name = "Relu_61"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc219) + %225 = "onnx.ReduceMeanV13"(%224) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_62_5"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x1x1xf32> loc(#loc220) + %226 = "onnx.Conv"(%225, %129, %17) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_63", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x120x1x1xf32>, tensor<32x120x1x1xf32>, tensor<32xf32>) -> tensor<1x32x1x1xf32> loc(#loc221) + %227 = "onnx.Relu"(%226) {onnx_node_name = "Relu_64"} : (tensor<1x32x1x1xf32>) -> tensor<1x32x1x1xf32> loc(#loc222) + %228 = "onnx.Conv"(%227, %137, %44) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_65", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x32x1x1xf32>, tensor<120x32x1x1xf32>, tensor<120xf32>) -> tensor<1x120x1x1xf32> loc(#loc223) + %229 = "onnx.Add"(%228, %71) {onnx_node_name = "Add_67"} : (tensor<1x120x1x1xf32>, tensor) -> tensor<1x120x1x1xf32> loc(#loc224) + %230 = "onnx.Clip"(%229, %69, %92) {onnx_node_name = "Clip_70_45"} : (tensor<1x120x1x1xf32>, tensor, tensor) -> tensor<1x120x1x1xf32> loc(#loc225) + %231 = "onnx.Div"(%230, %92) {onnx_node_name = "Div_72"} : (tensor<1x120x1x1xf32>, tensor) -> tensor<1x120x1x1xf32> loc(#loc226) + %232 = "onnx.Mul"(%231, %224) {onnx_node_name = "Mul_73"} : (tensor<1x120x1x1xf32>, tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc227) + %233 = "onnx.Conv"(%232, %64, %43) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_74", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x120x23x40xf32>, tensor<40x120x1x1xf32>, tensor<40xf32>) -> tensor<1x40x23x40xf32> loc(#loc228) + %234 = "onnx.Add"(%233, %220) {onnx_node_name = "Add_75"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc229) + %235 = "onnx.Conv"(%234, %41, %88) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_76", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x40x23x40xf32>, tensor<120x40x1x1xf32>, tensor<120xf32>) -> tensor<1x120x23x40xf32> loc(#loc230) + %236 = "onnx.Relu"(%235) {onnx_node_name = "Relu_77"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc231) + %237 = "onnx.Conv"(%236, %76, %53) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 120 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_78", + pads = [2, 2, 2, 2], + strides = [1, 1]} : (tensor<1x120x23x40xf32>, tensor<120x1x5x5xf32>, tensor<120xf32>) -> tensor<1x120x23x40xf32> loc(#loc232) + %238 = "onnx.Relu"(%237) {onnx_node_name = "Relu_79"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc233) + %239 = "onnx.ReduceMeanV13"(%238) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_80_6"} : (tensor<1x120x23x40xf32>) -> tensor<1x120x1x1xf32> loc(#loc234) + %240 = "onnx.Conv"(%239, %146, %130) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_81", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x120x1x1xf32>, tensor<32x120x1x1xf32>, tensor<32xf32>) -> tensor<1x32x1x1xf32> loc(#loc235) + %241 = "onnx.Relu"(%240) {onnx_node_name = "Relu_82"} : (tensor<1x32x1x1xf32>) -> tensor<1x32x1x1xf32> loc(#loc236) + %242 = "onnx.Conv"(%241, %101, %8) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_83", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x32x1x1xf32>, tensor<120x32x1x1xf32>, tensor<120xf32>) -> tensor<1x120x1x1xf32> loc(#loc237) + %243 = "onnx.Add"(%242, %71) {onnx_node_name = "Add_85"} : (tensor<1x120x1x1xf32>, tensor) -> tensor<1x120x1x1xf32> loc(#loc238) + %244 = "onnx.Clip"(%243, %69, %92) {onnx_node_name = "Clip_88_52"} : (tensor<1x120x1x1xf32>, tensor, tensor) -> tensor<1x120x1x1xf32> loc(#loc239) + %245 = "onnx.Div"(%244, %92) {onnx_node_name = "Div_90"} : (tensor<1x120x1x1xf32>, tensor) -> tensor<1x120x1x1xf32> loc(#loc240) + %246 = "onnx.Mul"(%245, %238) {onnx_node_name = "Mul_91"} : (tensor<1x120x1x1xf32>, tensor<1x120x23x40xf32>) -> tensor<1x120x23x40xf32> loc(#loc241) + %247 = "onnx.Conv"(%246, %83, %104) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_92", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x120x23x40xf32>, tensor<40x120x1x1xf32>, tensor<40xf32>) -> tensor<1x40x23x40xf32> loc(#loc242) + %248 = "onnx.Add"(%247, %234) {onnx_node_name = "Add_93"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc243) + %249 = "onnx.Conv"(%248, %47, %113) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_94", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x40x23x40xf32>, tensor<240x40x1x1xf32>, tensor<240xf32>) -> tensor<1x240x23x40xf32> loc(#loc244) + %250 = "onnx.Add"(%249, %71) {onnx_node_name = "Add_96"} : (tensor<1x240x23x40xf32>, tensor) -> tensor<1x240x23x40xf32> loc(#loc245) + %251 = "onnx.Clip"(%250, %69, %92) {onnx_node_name = "Clip_99_43"} : (tensor<1x240x23x40xf32>, tensor, tensor) -> tensor<1x240x23x40xf32> loc(#loc246) + %252 = "onnx.Div"(%251, %92) {onnx_node_name = "Div_101"} : (tensor<1x240x23x40xf32>, tensor) -> tensor<1x240x23x40xf32> loc(#loc247) + %253 = "onnx.Mul"(%249, %252) {onnx_node_name = "Mul_102"} : (tensor<1x240x23x40xf32>, tensor<1x240x23x40xf32>) -> tensor<1x240x23x40xf32> loc(#loc248) + %254 = "onnx.Conv"(%253, %48, %77) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 240 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_103", + pads = [1, 1, 1, 1], + strides = [2, 2]} : (tensor<1x240x23x40xf32>, tensor<240x1x3x3xf32>, tensor<240xf32>) -> tensor<1x240x12x20xf32> loc(#loc249) + %255 = "onnx.Add"(%254, %71) {onnx_node_name = "Add_105"} : (tensor<1x240x12x20xf32>, tensor) -> tensor<1x240x12x20xf32> loc(#loc250) + %256 = "onnx.Clip"(%255, %69, %92) {onnx_node_name = "Clip_108_25"} : (tensor<1x240x12x20xf32>, tensor, tensor) -> tensor<1x240x12x20xf32> loc(#loc251) + %257 = "onnx.Div"(%256, %92) {onnx_node_name = "Div_110"} : (tensor<1x240x12x20xf32>, tensor) -> tensor<1x240x12x20xf32> loc(#loc252) + %258 = "onnx.Mul"(%254, %257) {onnx_node_name = "Mul_111"} : (tensor<1x240x12x20xf32>, tensor<1x240x12x20xf32>) -> tensor<1x240x12x20xf32> loc(#loc253) + %259 = "onnx.Conv"(%258, %153, %67) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_112", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x240x12x20xf32>, tensor<80x240x1x1xf32>, tensor<80xf32>) -> tensor<1x80x12x20xf32> loc(#loc254) + %260 = "onnx.Conv"(%259, %55, %135) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_113", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x80x12x20xf32>, tensor<200x80x1x1xf32>, tensor<200xf32>) -> tensor<1x200x12x20xf32> loc(#loc255) + %261 = "onnx.Add"(%260, %71) {onnx_node_name = "Add_115"} : (tensor<1x200x12x20xf32>, tensor) -> tensor<1x200x12x20xf32> loc(#loc256) + %262 = "onnx.Clip"(%261, %69, %92) {onnx_node_name = "Clip_118_44"} : (tensor<1x200x12x20xf32>, tensor, tensor) -> tensor<1x200x12x20xf32> loc(#loc257) + %263 = "onnx.Div"(%262, %92) {onnx_node_name = "Div_120"} : (tensor<1x200x12x20xf32>, tensor) -> tensor<1x200x12x20xf32> loc(#loc258) + %264 = "onnx.Mul"(%260, %263) {onnx_node_name = "Mul_121"} : (tensor<1x200x12x20xf32>, tensor<1x200x12x20xf32>) -> tensor<1x200x12x20xf32> loc(#loc259) + %265 = "onnx.Conv"(%264, %57, %73) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 200 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_122", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x200x12x20xf32>, tensor<200x1x3x3xf32>, tensor<200xf32>) -> tensor<1x200x12x20xf32> loc(#loc260) + %266 = "onnx.Add"(%265, %71) {onnx_node_name = "Add_124"} : (tensor<1x200x12x20xf32>, tensor) -> tensor<1x200x12x20xf32> loc(#loc261) + %267 = "onnx.Clip"(%266, %69, %92) {onnx_node_name = "Clip_127_32"} : (tensor<1x200x12x20xf32>, tensor, tensor) -> tensor<1x200x12x20xf32> loc(#loc262) + %268 = "onnx.Div"(%267, %92) {onnx_node_name = "Div_129"} : (tensor<1x200x12x20xf32>, tensor) -> tensor<1x200x12x20xf32> loc(#loc263) + %269 = "onnx.Mul"(%265, %268) {onnx_node_name = "Mul_130"} : (tensor<1x200x12x20xf32>, tensor<1x200x12x20xf32>) -> tensor<1x200x12x20xf32> loc(#loc264) + %270 = "onnx.Conv"(%269, %45, %59) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_131", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x200x12x20xf32>, tensor<80x200x1x1xf32>, tensor<80xf32>) -> tensor<1x80x12x20xf32> loc(#loc265) + %271 = "onnx.Add"(%270, %259) {onnx_node_name = "Add_132"} : (tensor<1x80x12x20xf32>, tensor<1x80x12x20xf32>) -> tensor<1x80x12x20xf32> loc(#loc266) + %272 = "onnx.Conv"(%271, %109, %103) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_133", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x80x12x20xf32>, tensor<184x80x1x1xf32>, tensor<184xf32>) -> tensor<1x184x12x20xf32> loc(#loc267) + %273 = "onnx.Add"(%272, %71) {onnx_node_name = "Add_135"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc268) + %274 = "onnx.Clip"(%273, %69, %92) {onnx_node_name = "Clip_138_1"} : (tensor<1x184x12x20xf32>, tensor, tensor) -> tensor<1x184x12x20xf32> loc(#loc269) + %275 = "onnx.Div"(%274, %92) {onnx_node_name = "Div_140"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc270) + %276 = "onnx.Mul"(%272, %275) {onnx_node_name = "Mul_141"} : (tensor<1x184x12x20xf32>, tensor<1x184x12x20xf32>) -> tensor<1x184x12x20xf32> loc(#loc271) + %277 = "onnx.Conv"(%276, %87, %108) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 184 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_142", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x184x12x20xf32>, tensor<184x1x3x3xf32>, tensor<184xf32>) -> tensor<1x184x12x20xf32> loc(#loc272) + %278 = "onnx.Add"(%277, %71) {onnx_node_name = "Add_144"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc273) + %279 = "onnx.Clip"(%278, %69, %92) {onnx_node_name = "Clip_147_11"} : (tensor<1x184x12x20xf32>, tensor, tensor) -> tensor<1x184x12x20xf32> loc(#loc274) + %280 = "onnx.Div"(%279, %92) {onnx_node_name = "Div_149"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc275) + %281 = "onnx.Mul"(%277, %280) {onnx_node_name = "Mul_150"} : (tensor<1x184x12x20xf32>, tensor<1x184x12x20xf32>) -> tensor<1x184x12x20xf32> loc(#loc276) + %282 = "onnx.Conv"(%281, %111, %93) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_151", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x184x12x20xf32>, tensor<80x184x1x1xf32>, tensor<80xf32>) -> tensor<1x80x12x20xf32> loc(#loc277) + %283 = "onnx.Add"(%282, %271) {onnx_node_name = "Add_152"} : (tensor<1x80x12x20xf32>, tensor<1x80x12x20xf32>) -> tensor<1x80x12x20xf32> loc(#loc278) + %284 = "onnx.Conv"(%283, %155, %12) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_153", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x80x12x20xf32>, tensor<184x80x1x1xf32>, tensor<184xf32>) -> tensor<1x184x12x20xf32> loc(#loc279) + %285 = "onnx.Add"(%284, %71) {onnx_node_name = "Add_155"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc280) + %286 = "onnx.Clip"(%285, %69, %92) {onnx_node_name = "Clip_158_33"} : (tensor<1x184x12x20xf32>, tensor, tensor) -> tensor<1x184x12x20xf32> loc(#loc281) + %287 = "onnx.Div"(%286, %92) {onnx_node_name = "Div_160"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc282) + %288 = "onnx.Mul"(%284, %287) {onnx_node_name = "Mul_161"} : (tensor<1x184x12x20xf32>, tensor<1x184x12x20xf32>) -> tensor<1x184x12x20xf32> loc(#loc283) + %289 = "onnx.Conv"(%288, %56, %51) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 184 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_162", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x184x12x20xf32>, tensor<184x1x3x3xf32>, tensor<184xf32>) -> tensor<1x184x12x20xf32> loc(#loc284) + %290 = "onnx.Add"(%289, %71) {onnx_node_name = "Add_164"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc285) + %291 = "onnx.Clip"(%290, %69, %92) {onnx_node_name = "Clip_167_46"} : (tensor<1x184x12x20xf32>, tensor, tensor) -> tensor<1x184x12x20xf32> loc(#loc286) + %292 = "onnx.Div"(%291, %92) {onnx_node_name = "Div_169"} : (tensor<1x184x12x20xf32>, tensor) -> tensor<1x184x12x20xf32> loc(#loc287) + %293 = "onnx.Mul"(%289, %292) {onnx_node_name = "Mul_170"} : (tensor<1x184x12x20xf32>, tensor<1x184x12x20xf32>) -> tensor<1x184x12x20xf32> loc(#loc288) + %294 = "onnx.Conv"(%293, %122, %13) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_171", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x184x12x20xf32>, tensor<80x184x1x1xf32>, tensor<80xf32>) -> tensor<1x80x12x20xf32> loc(#loc289) + %295 = "onnx.Add"(%294, %283) {onnx_node_name = "Add_172"} : (tensor<1x80x12x20xf32>, tensor<1x80x12x20xf32>) -> tensor<1x80x12x20xf32> loc(#loc290) + %296 = "onnx.Conv"(%295, %110, %136) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_173", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x80x12x20xf32>, tensor<480x80x1x1xf32>, tensor<480xf32>) -> tensor<1x480x12x20xf32> loc(#loc291) + %297 = "onnx.Add"(%296, %71) {onnx_node_name = "Add_175"} : (tensor<1x480x12x20xf32>, tensor) -> tensor<1x480x12x20xf32> loc(#loc292) + %298 = "onnx.Clip"(%297, %69, %92) {onnx_node_name = "Clip_178_24"} : (tensor<1x480x12x20xf32>, tensor, tensor) -> tensor<1x480x12x20xf32> loc(#loc293) + %299 = "onnx.Div"(%298, %92) {onnx_node_name = "Div_180"} : (tensor<1x480x12x20xf32>, tensor) -> tensor<1x480x12x20xf32> loc(#loc294) + %300 = "onnx.Mul"(%296, %299) {onnx_node_name = "Mul_181"} : (tensor<1x480x12x20xf32>, tensor<1x480x12x20xf32>) -> tensor<1x480x12x20xf32> loc(#loc295) + %301 = "onnx.Conv"(%300, %123, %60) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 480 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_182", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x480x12x20xf32>, tensor<480x1x3x3xf32>, tensor<480xf32>) -> tensor<1x480x12x20xf32> loc(#loc296) + %302 = "onnx.Add"(%301, %71) {onnx_node_name = "Add_184"} : (tensor<1x480x12x20xf32>, tensor) -> tensor<1x480x12x20xf32> loc(#loc297) + %303 = "onnx.Clip"(%302, %69, %92) {onnx_node_name = "Clip_187_23"} : (tensor<1x480x12x20xf32>, tensor, tensor) -> tensor<1x480x12x20xf32> loc(#loc298) + %304 = "onnx.Div"(%303, %92) {onnx_node_name = "Div_189"} : (tensor<1x480x12x20xf32>, tensor) -> tensor<1x480x12x20xf32> loc(#loc299) + %305 = "onnx.Mul"(%301, %304) {onnx_node_name = "Mul_190"} : (tensor<1x480x12x20xf32>, tensor<1x480x12x20xf32>) -> tensor<1x480x12x20xf32> loc(#loc300) + %306 = "onnx.ReduceMeanV13"(%305) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_191_37"} : (tensor<1x480x12x20xf32>) -> tensor<1x480x1x1xf32> loc(#loc301) + %307 = "onnx.Conv"(%306, %106, %50) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_192", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x480x1x1xf32>, tensor<120x480x1x1xf32>, tensor<120xf32>) -> tensor<1x120x1x1xf32> loc(#loc302) + %308 = "onnx.Relu"(%307) {onnx_node_name = "Relu_193"} : (tensor<1x120x1x1xf32>) -> tensor<1x120x1x1xf32> loc(#loc303) + %309 = "onnx.Conv"(%308, %49, %63) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_194", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x120x1x1xf32>, tensor<480x120x1x1xf32>, tensor<480xf32>) -> tensor<1x480x1x1xf32> loc(#loc304) + %310 = "onnx.Add"(%309, %71) {onnx_node_name = "Add_196"} : (tensor<1x480x1x1xf32>, tensor) -> tensor<1x480x1x1xf32> loc(#loc305) + %311 = "onnx.Clip"(%310, %69, %92) {onnx_node_name = "Clip_199_41"} : (tensor<1x480x1x1xf32>, tensor, tensor) -> tensor<1x480x1x1xf32> loc(#loc306) + %312 = "onnx.Div"(%311, %92) {onnx_node_name = "Div_201"} : (tensor<1x480x1x1xf32>, tensor) -> tensor<1x480x1x1xf32> loc(#loc307) + %313 = "onnx.Mul"(%312, %305) {onnx_node_name = "Mul_202"} : (tensor<1x480x1x1xf32>, tensor<1x480x12x20xf32>) -> tensor<1x480x12x20xf32> loc(#loc308) + %314 = "onnx.Conv"(%313, %96, %7) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_203", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x480x12x20xf32>, tensor<112x480x1x1xf32>, tensor<112xf32>) -> tensor<1x112x12x20xf32> loc(#loc309) + %315 = "onnx.Conv"(%314, %79, %138) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_204", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x112x12x20xf32>, tensor<672x112x1x1xf32>, tensor<672xf32>) -> tensor<1x672x12x20xf32> loc(#loc310) + %316 = "onnx.Add"(%315, %71) {onnx_node_name = "Add_206"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc311) + %317 = "onnx.Clip"(%316, %69, %92) {onnx_node_name = "Clip_209_0"} : (tensor<1x672x12x20xf32>, tensor, tensor) -> tensor<1x672x12x20xf32> loc(#loc312) + %318 = "onnx.Div"(%317, %92) {onnx_node_name = "Div_211"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc313) + %319 = "onnx.Mul"(%315, %318) {onnx_node_name = "Mul_212"} : (tensor<1x672x12x20xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc314) + %320 = "onnx.Conv"(%319, %132, %100) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 672 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_213", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x672x12x20xf32>, tensor<672x1x3x3xf32>, tensor<672xf32>) -> tensor<1x672x12x20xf32> loc(#loc315) + %321 = "onnx.Add"(%320, %71) {onnx_node_name = "Add_215"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc316) + %322 = "onnx.Clip"(%321, %69, %92) {onnx_node_name = "Clip_218_51"} : (tensor<1x672x12x20xf32>, tensor, tensor) -> tensor<1x672x12x20xf32> loc(#loc317) + %323 = "onnx.Div"(%322, %92) {onnx_node_name = "Div_220"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc318) + %324 = "onnx.Mul"(%320, %323) {onnx_node_name = "Mul_221"} : (tensor<1x672x12x20xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc319) + %325 = "onnx.ReduceMeanV13"(%324) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_222_2"} : (tensor<1x672x12x20xf32>) -> tensor<1x672x1x1xf32> loc(#loc320) + %326 = "onnx.Conv"(%325, %133, %145) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_223", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x672x1x1xf32>, tensor<168x672x1x1xf32>, tensor<168xf32>) -> tensor<1x168x1x1xf32> loc(#loc321) + %327 = "onnx.Relu"(%326) {onnx_node_name = "Relu_224"} : (tensor<1x168x1x1xf32>) -> tensor<1x168x1x1xf32> loc(#loc322) + %328 = "onnx.Conv"(%327, %91, %124) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_225", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x168x1x1xf32>, tensor<672x168x1x1xf32>, tensor<672xf32>) -> tensor<1x672x1x1xf32> loc(#loc323) + %329 = "onnx.Add"(%328, %71) {onnx_node_name = "Add_227"} : (tensor<1x672x1x1xf32>, tensor) -> tensor<1x672x1x1xf32> loc(#loc324) + %330 = "onnx.Clip"(%329, %69, %92) {onnx_node_name = "Clip_230_53"} : (tensor<1x672x1x1xf32>, tensor, tensor) -> tensor<1x672x1x1xf32> loc(#loc325) + %331 = "onnx.Div"(%330, %92) {onnx_node_name = "Div_232"} : (tensor<1x672x1x1xf32>, tensor) -> tensor<1x672x1x1xf32> loc(#loc326) + %332 = "onnx.Mul"(%331, %324) {onnx_node_name = "Mul_233"} : (tensor<1x672x1x1xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc327) + %333 = "onnx.Conv"(%332, %95, %97) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_234", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x672x12x20xf32>, tensor<112x672x1x1xf32>, tensor<112xf32>) -> tensor<1x112x12x20xf32> loc(#loc328) + %334 = "onnx.Add"(%333, %314) {onnx_node_name = "Add_235"} : (tensor<1x112x12x20xf32>, tensor<1x112x12x20xf32>) -> tensor<1x112x12x20xf32> loc(#loc329) + %335 = "onnx.Conv"(%334, %128, %120) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_236", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x112x12x20xf32>, tensor<672x112x1x1xf32>, tensor<672xf32>) -> tensor<1x672x12x20xf32> loc(#loc330) + %336 = "onnx.Add"(%335, %71) {onnx_node_name = "Add_238"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc331) + %337 = "onnx.Clip"(%336, %69, %92) {onnx_node_name = "Clip_241_17"} : (tensor<1x672x12x20xf32>, tensor, tensor) -> tensor<1x672x12x20xf32> loc(#loc332) + %338 = "onnx.Div"(%337, %92) {onnx_node_name = "Div_243"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc333) + %339 = "onnx.Mul"(%335, %338) {onnx_node_name = "Mul_244"} : (tensor<1x672x12x20xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc334) + %340 = "onnx.Conv"(%339, %107, %46) { + auto_pad = "NOTSET", + dilations = [2, 2], + group = 672 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_245", + pads = [4, 4, 4, 4], + strides = [1, 1]} : (tensor<1x672x12x20xf32>, tensor<672x1x5x5xf32>, tensor<672xf32>) -> tensor<1x672x12x20xf32> loc(#loc335) + %341 = "onnx.Add"(%340, %71) {onnx_node_name = "Add_247"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc336) + %342 = "onnx.Clip"(%341, %69, %92) {onnx_node_name = "Clip_250_15"} : (tensor<1x672x12x20xf32>, tensor, tensor) -> tensor<1x672x12x20xf32> loc(#loc337) + %343 = "onnx.Div"(%342, %92) {onnx_node_name = "Div_252"} : (tensor<1x672x12x20xf32>, tensor) -> tensor<1x672x12x20xf32> loc(#loc338) + %344 = "onnx.Mul"(%340, %343) {onnx_node_name = "Mul_253"} : (tensor<1x672x12x20xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc339) + %345 = "onnx.ReduceMeanV13"(%344) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_254_36"} : (tensor<1x672x12x20xf32>) -> tensor<1x672x1x1xf32> loc(#loc340) + %346 = "onnx.Conv"(%345, %134, %94) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_255", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x672x1x1xf32>, tensor<168x672x1x1xf32>, tensor<168xf32>) -> tensor<1x168x1x1xf32> loc(#loc341) + %347 = "onnx.Relu"(%346) {onnx_node_name = "Relu_256"} : (tensor<1x168x1x1xf32>) -> tensor<1x168x1x1xf32> loc(#loc342) + %348 = "onnx.Conv"(%347, %112, %54) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_257", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x168x1x1xf32>, tensor<672x168x1x1xf32>, tensor<672xf32>) -> tensor<1x672x1x1xf32> loc(#loc343) + %349 = "onnx.Add"(%348, %71) {onnx_node_name = "Add_259"} : (tensor<1x672x1x1xf32>, tensor) -> tensor<1x672x1x1xf32> loc(#loc344) + %350 = "onnx.Clip"(%349, %69, %92) {onnx_node_name = "Clip_262_20"} : (tensor<1x672x1x1xf32>, tensor, tensor) -> tensor<1x672x1x1xf32> loc(#loc345) + %351 = "onnx.Div"(%350, %92) {onnx_node_name = "Div_264"} : (tensor<1x672x1x1xf32>, tensor) -> tensor<1x672x1x1xf32> loc(#loc346) + %352 = "onnx.Mul"(%351, %344) {onnx_node_name = "Mul_265"} : (tensor<1x672x1x1xf32>, tensor<1x672x12x20xf32>) -> tensor<1x672x12x20xf32> loc(#loc347) + %353 = "onnx.Conv"(%352, %39, %38) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_266", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x672x12x20xf32>, tensor<160x672x1x1xf32>, tensor<160xf32>) -> tensor<1x160x12x20xf32> loc(#loc348) + %354 = "onnx.Conv"(%353, %164, %141) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_267", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x160x12x20xf32>, tensor<960x160x1x1xf32>, tensor<960xf32>) -> tensor<1x960x12x20xf32> loc(#loc349) + %355 = "onnx.Add"(%354, %71) {onnx_node_name = "Add_269"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc350) + %356 = "onnx.Clip"(%355, %69, %92) {onnx_node_name = "Clip_272_34"} : (tensor<1x960x12x20xf32>, tensor, tensor) -> tensor<1x960x12x20xf32> loc(#loc351) + %357 = "onnx.Div"(%356, %92) {onnx_node_name = "Div_274"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc352) + %358 = "onnx.Mul"(%354, %357) {onnx_node_name = "Mul_275"} : (tensor<1x960x12x20xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc353) + %359 = "onnx.Conv"(%358, %37, %85) { + auto_pad = "NOTSET", + dilations = [2, 2], + group = 960 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_276", + pads = [4, 4, 4, 4], + strides = [1, 1]} : (tensor<1x960x12x20xf32>, tensor<960x1x5x5xf32>, tensor<960xf32>) -> tensor<1x960x12x20xf32> loc(#loc354) + %360 = "onnx.Add"(%359, %71) {onnx_node_name = "Add_278"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc355) + %361 = "onnx.Clip"(%360, %69, %92) {onnx_node_name = "Clip_281_18"} : (tensor<1x960x12x20xf32>, tensor, tensor) -> tensor<1x960x12x20xf32> loc(#loc356) + %362 = "onnx.Div"(%361, %92) {onnx_node_name = "Div_283"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc357) + %363 = "onnx.Mul"(%359, %362) {onnx_node_name = "Mul_284"} : (tensor<1x960x12x20xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc358) + %364 = "onnx.ReduceMeanV13"(%363) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_285_38"} : (tensor<1x960x12x20xf32>) -> tensor<1x960x1x1xf32> loc(#loc359) + %365 = "onnx.Conv"(%364, %167, %78) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_286", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x1x1xf32>, tensor<240x960x1x1xf32>, tensor<240xf32>) -> tensor<1x240x1x1xf32> loc(#loc360) + %366 = "onnx.Relu"(%365) {onnx_node_name = "Relu_287"} : (tensor<1x240x1x1xf32>) -> tensor<1x240x1x1xf32> loc(#loc361) + %367 = "onnx.Conv"(%366, %36, %66) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_288", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x240x1x1xf32>, tensor<960x240x1x1xf32>, tensor<960xf32>) -> tensor<1x960x1x1xf32> loc(#loc362) + %368 = "onnx.Add"(%367, %71) {onnx_node_name = "Add_290"} : (tensor<1x960x1x1xf32>, tensor) -> tensor<1x960x1x1xf32> loc(#loc363) + %369 = "onnx.Clip"(%368, %69, %92) {onnx_node_name = "Clip_293_3"} : (tensor<1x960x1x1xf32>, tensor, tensor) -> tensor<1x960x1x1xf32> loc(#loc364) + %370 = "onnx.Div"(%369, %92) {onnx_node_name = "Div_295"} : (tensor<1x960x1x1xf32>, tensor) -> tensor<1x960x1x1xf32> loc(#loc365) + %371 = "onnx.Mul"(%370, %363) {onnx_node_name = "Mul_296"} : (tensor<1x960x1x1xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc366) + %372 = "onnx.Conv"(%371, %156, %35) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_297", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x12x20xf32>, tensor<160x960x1x1xf32>, tensor<160xf32>) -> tensor<1x160x12x20xf32> loc(#loc367) + %373 = "onnx.Add"(%372, %353) {onnx_node_name = "Add_298"} : (tensor<1x160x12x20xf32>, tensor<1x160x12x20xf32>) -> tensor<1x160x12x20xf32> loc(#loc368) + %374 = "onnx.Conv"(%373, %172, %33) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_299", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x160x12x20xf32>, tensor<960x160x1x1xf32>, tensor<960xf32>) -> tensor<1x960x12x20xf32> loc(#loc369) + %375 = "onnx.Add"(%374, %71) {onnx_node_name = "Add_301"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc370) + %376 = "onnx.Clip"(%375, %69, %92) {onnx_node_name = "Clip_304_8"} : (tensor<1x960x12x20xf32>, tensor, tensor) -> tensor<1x960x12x20xf32> loc(#loc371) + %377 = "onnx.Div"(%376, %92) {onnx_node_name = "Div_306"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc372) + %378 = "onnx.Mul"(%374, %377) {onnx_node_name = "Mul_307"} : (tensor<1x960x12x20xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc373) + %379 = "onnx.Conv"(%378, %32, %117) { + auto_pad = "NOTSET", + dilations = [2, 2], + group = 960 : si64, + kernel_shape = [5, 5], + onnx_node_name = "Conv_308", + pads = [4, 4, 4, 4], + strides = [1, 1]} : (tensor<1x960x12x20xf32>, tensor<960x1x5x5xf32>, tensor<960xf32>) -> tensor<1x960x12x20xf32> loc(#loc374) + %380 = "onnx.Add"(%379, %71) {onnx_node_name = "Add_310"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc375) + %381 = "onnx.Clip"(%380, %69, %92) {onnx_node_name = "Clip_313_40"} : (tensor<1x960x12x20xf32>, tensor, tensor) -> tensor<1x960x12x20xf32> loc(#loc376) + %382 = "onnx.Div"(%381, %92) {onnx_node_name = "Div_315"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc377) + %383 = "onnx.Mul"(%379, %382) {onnx_node_name = "Mul_316"} : (tensor<1x960x12x20xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc378) + %384 = "onnx.ReduceMeanV13"(%383) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_317_7"} : (tensor<1x960x12x20xf32>) -> tensor<1x960x1x1xf32> loc(#loc379) + %385 = "onnx.Conv"(%384, %119, %143) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_318", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x1x1xf32>, tensor<240x960x1x1xf32>, tensor<240xf32>) -> tensor<1x240x1x1xf32> loc(#loc380) + %386 = "onnx.Relu"(%385) {onnx_node_name = "Relu_319"} : (tensor<1x240x1x1xf32>) -> tensor<1x240x1x1xf32> loc(#loc381) + %387 = "onnx.Conv"(%386, %34, %151) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_320", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x240x1x1xf32>, tensor<960x240x1x1xf32>, tensor<960xf32>) -> tensor<1x960x1x1xf32> loc(#loc382) + %388 = "onnx.Add"(%387, %71) {onnx_node_name = "Add_322"} : (tensor<1x960x1x1xf32>, tensor) -> tensor<1x960x1x1xf32> loc(#loc383) + %389 = "onnx.Clip"(%388, %69, %92) {onnx_node_name = "Clip_325_10"} : (tensor<1x960x1x1xf32>, tensor, tensor) -> tensor<1x960x1x1xf32> loc(#loc384) + %390 = "onnx.Div"(%389, %92) {onnx_node_name = "Div_327"} : (tensor<1x960x1x1xf32>, tensor) -> tensor<1x960x1x1xf32> loc(#loc385) + %391 = "onnx.Mul"(%390, %383) {onnx_node_name = "Mul_328"} : (tensor<1x960x1x1xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc386) + %392 = "onnx.Conv"(%391, %31, %152) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_329", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x12x20xf32>, tensor<160x960x1x1xf32>, tensor<160xf32>) -> tensor<1x160x12x20xf32> loc(#loc387) + %393 = "onnx.Add"(%392, %373) {onnx_node_name = "Add_330"} : (tensor<1x160x12x20xf32>, tensor<1x160x12x20xf32>) -> tensor<1x160x12x20xf32> loc(#loc388) + %394 = "onnx.Conv"(%393, %30, %158) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_331", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x160x12x20xf32>, tensor<960x160x1x1xf32>, tensor<960xf32>) -> tensor<1x960x12x20xf32> loc(#loc389) + %395 = "onnx.Add"(%394, %71) {onnx_node_name = "Add_333"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc390) + %396 = "onnx.Clip"(%395, %69, %92) {onnx_node_name = "Clip_336_42"} : (tensor<1x960x12x20xf32>, tensor, tensor) -> tensor<1x960x12x20xf32> loc(#loc391) + %397 = "onnx.Div"(%396, %92) {onnx_node_name = "Div_338"} : (tensor<1x960x12x20xf32>, tensor) -> tensor<1x960x12x20xf32> loc(#loc392) + %398 = "onnx.Mul"(%394, %397) {onnx_node_name = "Mul_339"} : (tensor<1x960x12x20xf32>, tensor<1x960x12x20xf32>) -> tensor<1x960x12x20xf32> loc(#loc393) + %399 = "onnx.Conv"(%398, %26, %148) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_340", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x12x20xf32>, tensor<128x960x1x1xf32>, tensor<128xf32>) -> tensor<1x128x12x20xf32> loc(#loc394) + %400 = "onnx.Relu"(%399) {onnx_node_name = "Relu_341"} : (tensor<1x128x12x20xf32>) -> tensor<1x128x12x20xf32> loc(#loc395) + %401 = "onnx.ReduceMeanV13"(%398) { + axes = [2, 3], + keepdims = 1 : si64, + onnx_node_name = "GlobalAveragePool_342_49"} : (tensor<1x960x12x20xf32>) -> tensor<1x960x1x1xf32> loc(#loc396) + %402 = "onnx.Conv"(%401, %15, %6) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_343", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x960x1x1xf32>, tensor<128x960x1x1xf32>, none) -> tensor<1x128x1x1xf32> loc(#loc397) + %403 = "onnx.Sigmoid"(%402) {onnx_node_name = "Sigmoid_344"} : (tensor<1x128x1x1xf32>) -> tensor<1x128x1x1xf32> loc(#loc398) + %404 = "onnx.Mul"(%400, %403) {onnx_node_name = "Mul_345"} : (tensor<1x128x12x20xf32>, tensor<1x128x1x1xf32>) -> tensor<1x128x12x20xf32> loc(#loc399) + %405:2 = "onnx.Split"(%404, %5) {axis = 1 : si64, onnx_node_name = "Split_349_29"} : (tensor<1x128x12x20xf32>, tensor<2xi64>) -> (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) loc(#loc400) + %406 = "onnx.Concat"(%405#1, %185) {axis = 1 : si64, onnx_node_name = "Concat_350"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x128x12x20xf32> loc(#loc401) + %407 = "onnx.Conv"(%406, %127, %147) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_351", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x128x12x20xf32>, tensor<128x128x3x3xf32>, tensor<128xf32>) -> tensor<1x128x12x20xf32> loc(#loc402) + %408 = "onnx.Sigmoid"(%407) {onnx_node_name = "Sigmoid_352"} : (tensor<1x128x12x20xf32>) -> tensor<1x128x12x20xf32> loc(#loc403) + %409:2 = "onnx.Split"(%408, %5) {axis = 1 : si64, onnx_node_name = "Split_353_13"} : (tensor<1x128x12x20xf32>, tensor<2xi64>) -> (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) loc(#loc404) + %410 = "onnx.Sub"(%98, %409#1) {onnx_node_name = "Sub_359"} : (tensor, tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc405) + %411 = "onnx.Mul"(%410, %185) {onnx_node_name = "Mul_360"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc406) + %412 = "onnx.Mul"(%409#0, %185) {onnx_node_name = "Mul_354"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc407) + %413 = "onnx.Concat"(%405#1, %412) {axis = 1 : si64, onnx_node_name = "Concat_355"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x128x12x20xf32> loc(#loc408) + %414 = "onnx.Conv"(%413, %62, %121) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_356", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x128x12x20xf32>, tensor<64x128x3x3xf32>, tensor<64xf32>) -> tensor<1x64x12x20xf32> loc(#loc409) + %415 = "onnx.Tanh"(%414) {onnx_node_name = "Tanh_357"} : (tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc410) + %416 = "onnx.Mul"(%409#1, %415) {onnx_node_name = "Mul_361"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc411) + %417 = "onnx.Add"(%411, %416) {onnx_node_name = "Add_362"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x64x12x20xf32> loc(#loc412) + %418 = "onnx.Concat"(%405#0, %417) {axis = 1 : si64, onnx_node_name = "Concat_363"} : (tensor<1x64x12x20xf32>, tensor<1x64x12x20xf32>) -> tensor<1x128x12x20xf32> loc(#loc413) + %419 = "onnx.Resize"(%418, %6, %65, %6) { + antialias = 0 : si64, + coordinate_transformation_mode = "pytorch_half_pixel", + cubic_coeff_a = -7.500000e-01 : f32, + exclude_outside = 0 : si64, + extrapolation_value = 0.000000e+00 : f32, + keep_aspect_ratio_policy = "stretch", + mode = "linear", + nearest_mode = "floor", + onnx_node_name = "Resize_365_21"} : (tensor<1x128x12x20xf32>, none, tensor<4xf32>, none) -> tensor<1x128x24x40xf32> loc(#loc414) + %420 = "onnx.Slice"(%419, %105, %81, %84, %89) {onnx_node_name = "Slice_371"} : (tensor<1x128x24x40xf32>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>) -> tensor<1x128x23x40xf32> loc(#loc415) + %421 = "onnx.Concat"(%420, %248, %184) {axis = 1 : si64, onnx_node_name = "Concat_372"} : (tensor<1x128x23x40xf32>, tensor<1x40x23x40xf32>, tensor<1x3x23x40xf32>) -> tensor<1x171x23x40xf32> loc(#loc416) + %422 = "onnx.Conv"(%421, %28, %27) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_373", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x171x23x40xf32>, tensor<80x171x3x3xf32>, tensor<80xf32>) -> tensor<1x80x23x40xf32> loc(#loc417) + %423 = "onnx.Relu"(%422) {onnx_node_name = "Relu_374"} : (tensor<1x80x23x40xf32>) -> tensor<1x80x23x40xf32> loc(#loc418) + %424:2 = "onnx.Split"(%423, %4) {axis = 1 : si64, onnx_node_name = "Split_375_19"} : (tensor<1x80x23x40xf32>, tensor<2xi64>) -> (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) loc(#loc419) + %425 = "onnx.Concat"(%424#1, %177) {axis = 1 : si64, onnx_node_name = "Concat_376"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x80x23x40xf32> loc(#loc420) + %426 = "onnx.Conv"(%425, %114, %140) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_377", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x80x23x40xf32>, tensor<80x80x3x3xf32>, tensor<80xf32>) -> tensor<1x80x23x40xf32> loc(#loc421) + %427 = "onnx.Sigmoid"(%426) {onnx_node_name = "Sigmoid_378"} : (tensor<1x80x23x40xf32>) -> tensor<1x80x23x40xf32> loc(#loc422) + %428:2 = "onnx.Split"(%427, %4) {axis = 1 : si64, onnx_node_name = "Split_379_35"} : (tensor<1x80x23x40xf32>, tensor<2xi64>) -> (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) loc(#loc423) + %429 = "onnx.Sub"(%98, %428#1) {onnx_node_name = "Sub_385"} : (tensor, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc424) + %430 = "onnx.Mul"(%429, %177) {onnx_node_name = "Mul_386"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc425) + %431 = "onnx.Mul"(%428#0, %177) {onnx_node_name = "Mul_380"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc426) + %432 = "onnx.Concat"(%424#1, %431) {axis = 1 : si64, onnx_node_name = "Concat_381"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x80x23x40xf32> loc(#loc427) + %433 = "onnx.Conv"(%432, %159, %162) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_382", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x80x23x40xf32>, tensor<40x80x3x3xf32>, tensor<40xf32>) -> tensor<1x40x23x40xf32> loc(#loc428) + %434 = "onnx.Tanh"(%433) {onnx_node_name = "Tanh_383"} : (tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc429) + %435 = "onnx.Mul"(%428#1, %434) {onnx_node_name = "Mul_387"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc430) + %436 = "onnx.Add"(%430, %435) {onnx_node_name = "Add_388"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x40x23x40xf32> loc(#loc431) + %437 = "onnx.Concat"(%424#0, %436) {axis = 1 : si64, onnx_node_name = "Concat_389"} : (tensor<1x40x23x40xf32>, tensor<1x40x23x40xf32>) -> tensor<1x80x23x40xf32> loc(#loc432) + %438 = "onnx.Resize"(%437, %6, %65, %6) { + antialias = 0 : si64, + coordinate_transformation_mode = "pytorch_half_pixel", + cubic_coeff_a = -7.500000e-01 : f32, + exclude_outside = 0 : si64, + extrapolation_value = 0.000000e+00 : f32, + keep_aspect_ratio_policy = "stretch", + mode = "linear", + nearest_mode = "floor", + onnx_node_name = "Resize_391_4"} : (tensor<1x80x23x40xf32>, none, tensor<4xf32>, none) -> tensor<1x80x46x80xf32> loc(#loc433) + %439 = "onnx.Slice"(%438, %105, %139, %84, %89) {onnx_node_name = "Slice_397"} : (tensor<1x80x46x80xf32>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>, tensor<1xi64>) -> tensor<1x80x45x80xf32> loc(#loc434) + %440 = "onnx.Concat"(%439, %207, %183) {axis = 1 : si64, onnx_node_name = "Concat_398"} : (tensor<1x80x45x80xf32>, tensor<1x24x45x80xf32>, tensor<1x3x45x80xf32>) -> tensor<1x107x45x80xf32> loc(#loc435) + %441 = "onnx.Conv"(%440, %90, %25) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_399", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x107x45x80xf32>, tensor<40x107x3x3xf32>, tensor<40xf32>) -> tensor<1x40x45x80xf32> loc(#loc436) + %442 = "onnx.Relu"(%441) {onnx_node_name = "Relu_400"} : (tensor<1x40x45x80xf32>) -> tensor<1x40x45x80xf32> loc(#loc437) + %443:2 = "onnx.Split"(%442, %3) {axis = 1 : si64, onnx_node_name = "Split_401_9"} : (tensor<1x40x45x80xf32>, tensor<2xi64>) -> (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) loc(#loc438) + %444 = "onnx.Concat"(%443#1, %176) {axis = 1 : si64, onnx_node_name = "Concat_402"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x40x45x80xf32> loc(#loc439) + %445 = "onnx.Conv"(%444, %163, %126) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_403", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x40x45x80xf32>, tensor<40x40x3x3xf32>, tensor<40xf32>) -> tensor<1x40x45x80xf32> loc(#loc440) + %446 = "onnx.Sigmoid"(%445) {onnx_node_name = "Sigmoid_404"} : (tensor<1x40x45x80xf32>) -> tensor<1x40x45x80xf32> loc(#loc441) + %447:2 = "onnx.Split"(%446, %3) {axis = 1 : si64, onnx_node_name = "Split_405_14"} : (tensor<1x40x45x80xf32>, tensor<2xi64>) -> (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) loc(#loc442) + %448 = "onnx.Sub"(%98, %447#1) {onnx_node_name = "Sub_411"} : (tensor, tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc443) + %449 = "onnx.Mul"(%448, %176) {onnx_node_name = "Mul_412"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc444) + %450 = "onnx.Mul"(%447#0, %176) {onnx_node_name = "Mul_406"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc445) + %451 = "onnx.Concat"(%443#1, %450) {axis = 1 : si64, onnx_node_name = "Concat_407"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x40x45x80xf32> loc(#loc446) + %452 = "onnx.Conv"(%451, %154, %168) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_408", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x40x45x80xf32>, tensor<20x40x3x3xf32>, tensor<20xf32>) -> tensor<1x20x45x80xf32> loc(#loc447) + %453 = "onnx.Tanh"(%452) {onnx_node_name = "Tanh_409"} : (tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc448) + %454 = "onnx.Mul"(%447#1, %453) {onnx_node_name = "Mul_413"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc449) + %455 = "onnx.Add"(%449, %454) {onnx_node_name = "Add_414"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x20x45x80xf32> loc(#loc450) + %456 = "onnx.Concat"(%443#0, %455) {axis = 1 : si64, onnx_node_name = "Concat_415"} : (tensor<1x20x45x80xf32>, tensor<1x20x45x80xf32>) -> tensor<1x40x45x80xf32> loc(#loc451) + %457 = "onnx.Resize"(%456, %6, %65, %6) { + antialias = 0 : si64, + coordinate_transformation_mode = "pytorch_half_pixel", + cubic_coeff_a = -7.500000e-01 : f32, + exclude_outside = 0 : si64, + extrapolation_value = 0.000000e+00 : f32, + keep_aspect_ratio_policy = "stretch", + mode = "linear", + nearest_mode = "floor", + onnx_node_name = "Resize_417_27"} : (tensor<1x40x45x80xf32>, none, tensor<4xf32>, none) -> tensor<1x40x90x160xf32> loc(#loc452) + %458 = "onnx.Concat"(%457, %196, %182) {axis = 1 : si64, onnx_node_name = "Concat_418"} : (tensor<1x40x90x160xf32>, tensor<1x16x90x160xf32>, tensor<1x3x90x160xf32>) -> tensor<1x59x90x160xf32> loc(#loc453) + %459 = "onnx.Conv"(%458, %24, %23) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_419", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x59x90x160xf32>, tensor<32x59x3x3xf32>, tensor<32xf32>) -> tensor<1x32x90x160xf32> loc(#loc454) + %460 = "onnx.Relu"(%459) {onnx_node_name = "Relu_420"} : (tensor<1x32x90x160xf32>) -> tensor<1x32x90x160xf32> loc(#loc455) + %461:2 = "onnx.Split"(%460, %2) {axis = 1 : si64, onnx_node_name = "Split_421_16"} : (tensor<1x32x90x160xf32>, tensor<2xi64>) -> (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) loc(#loc456) + %462 = "onnx.Concat"(%461#1, %175) {axis = 1 : si64, onnx_node_name = "Concat_422"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x32x90x160xf32> loc(#loc457) + %463 = "onnx.Conv"(%462, %52, %102) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_423", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x32x90x160xf32>, tensor<32x32x3x3xf32>, tensor<32xf32>) -> tensor<1x32x90x160xf32> loc(#loc458) + %464 = "onnx.Sigmoid"(%463) {onnx_node_name = "Sigmoid_424"} : (tensor<1x32x90x160xf32>) -> tensor<1x32x90x160xf32> loc(#loc459) + %465:2 = "onnx.Split"(%464, %2) {axis = 1 : si64, onnx_node_name = "Split_425_22"} : (tensor<1x32x90x160xf32>, tensor<2xi64>) -> (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) loc(#loc460) + %466 = "onnx.Sub"(%98, %465#1) {onnx_node_name = "Sub_431"} : (tensor, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc461) + %467 = "onnx.Mul"(%466, %175) {onnx_node_name = "Mul_432"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc462) + %468 = "onnx.Mul"(%465#0, %175) {onnx_node_name = "Mul_426"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc463) + %469 = "onnx.Concat"(%461#1, %468) {axis = 1 : si64, onnx_node_name = "Concat_427"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x32x90x160xf32> loc(#loc464) + %470 = "onnx.Conv"(%469, %68, %170) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_428", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x32x90x160xf32>, tensor<16x32x3x3xf32>, tensor<16xf32>) -> tensor<1x16x90x160xf32> loc(#loc465) + %471 = "onnx.Tanh"(%470) {onnx_node_name = "Tanh_429"} : (tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc466) + %472 = "onnx.Mul"(%465#1, %471) {onnx_node_name = "Mul_433"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc467) + %473 = "onnx.Add"(%467, %472) {onnx_node_name = "Add_434"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x16x90x160xf32> loc(#loc468) + %474 = "onnx.Transpose"(%473) {onnx_node_name = "Transpose_448", perm = [0, 2, 3, 1]} : (tensor<1x16x90x160xf32>) -> tensor<1x90x160x16xf32> loc(#loc469) + %475 = "onnx.Transpose"(%455) {onnx_node_name = "Transpose_449", perm = [0, 2, 3, 1]} : (tensor<1x20x45x80xf32>) -> tensor<1x45x80x20xf32> loc(#loc470) + %476 = "onnx.Transpose"(%436) {onnx_node_name = "Transpose_450", perm = [0, 2, 3, 1]} : (tensor<1x40x23x40xf32>) -> tensor<1x23x40x40xf32> loc(#loc471) + %477 = "onnx.Transpose"(%417) {onnx_node_name = "Transpose_451", perm = [0, 2, 3, 1]} : (tensor<1x64x12x20xf32>) -> tensor<1x12x20x64xf32> loc(#loc472) + %478 = "onnx.Concat"(%461#0, %473) {axis = 1 : si64, onnx_node_name = "Concat_435"} : (tensor<1x16x90x160xf32>, tensor<1x16x90x160xf32>) -> tensor<1x32x90x160xf32> loc(#loc473) + %479 = "onnx.Resize"(%478, %6, %65, %6) { + antialias = 0 : si64, + coordinate_transformation_mode = "pytorch_half_pixel", + cubic_coeff_a = -7.500000e-01 : f32, + exclude_outside = 0 : si64, + extrapolation_value = 0.000000e+00 : f32, + keep_aspect_ratio_policy = "stretch", + mode = "linear", + nearest_mode = "floor", + onnx_node_name = "Resize_437_28"} : (tensor<1x32x90x160xf32>, none, tensor<4xf32>, none) -> tensor<1x32x180x320xf32> loc(#loc474) + %480 = "onnx.Concat"(%479, %181) {axis = 1 : si64, onnx_node_name = "Concat_438"} : (tensor<1x32x180x320xf32>, tensor<1x3x180x320xf32>) -> tensor<1x35x180x320xf32> loc(#loc475) + %481 = "onnx.Conv"(%480, %22, %20) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_439", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x35x180x320xf32>, tensor<16x35x3x3xf32>, tensor<16xf32>) -> tensor<1x16x180x320xf32> loc(#loc476) + %482 = "onnx.Relu"(%481) {onnx_node_name = "Relu_440"} : (tensor<1x16x180x320xf32>) -> tensor<1x16x180x320xf32> loc(#loc477) + %483 = "onnx.Conv"(%482, %19, %18) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [3, 3], + onnx_node_name = "Conv_441", + pads = [1, 1, 1, 1], + strides = [1, 1]} : (tensor<1x16x180x320xf32>, tensor<16x16x3x3xf32>, tensor<16xf32>) -> tensor<1x16x180x320xf32> loc(#loc478) + %484 = "onnx.Relu"(%483) {onnx_node_name = "Relu_442"} : (tensor<1x16x180x320xf32>) -> tensor<1x16x180x320xf32> loc(#loc479) + %485 = "onnx.Conv"(%484, %115, %70) { + auto_pad = "NOTSET", + dilations = [1, 1], + group = 1 : si64, + kernel_shape = [1, 1], + onnx_node_name = "Conv_443", + pads = [0, 0, 0, 0], + strides = [1, 1]} : (tensor<1x16x180x320xf32>, tensor<4x16x1x1xf32>, tensor<4xf32>) -> tensor<1x4x180x320xf32> loc(#loc480) + %486:2 = "onnx.Split"(%485, %1) {axis = 1 : si64, onnx_node_name = "Split_444_31"} : (tensor<1x4x180x320xf32>, tensor<2xi64>) -> (tensor<1x3x180x320xf32>, tensor<1x1x180x320xf32>) loc(#loc481) + %487 = "onnx.Add"(%486#0, %181) {onnx_node_name = "Add_445"} : (tensor<1x3x180x320xf32>, tensor<1x3x180x320xf32>) -> tensor<1x3x180x320xf32> loc(#loc482) + %488 = "onnx.Clip"(%487, %69, %98) {onnx_node_name = "Clip_446_47"} : (tensor<1x3x180x320xf32>, tensor, tensor) -> tensor<1x3x180x320xf32> loc(#loc483) + %489 = "onnx.Transpose"(%488) {onnx_node_name = "Transpose_452", perm = [0, 2, 3, 1]} : (tensor<1x3x180x320xf32>) -> tensor<1x180x320x3xf32> loc(#loc484) + %490 = "onnx.Clip"(%486#1, %69, %98) {onnx_node_name = "Clip_447_30"} : (tensor<1x1x180x320xf32>, tensor, tensor) -> tensor<1x1x180x320xf32> loc(#loc485) + %491 = "onnx.Transpose"(%490) {onnx_node_name = "Transpose_453", perm = [0, 2, 3, 1]} : (tensor<1x1x180x320xf32>) -> tensor<1x180x320x1xf32> loc(#loc486) + return %489, %491, %474, %475, %476, %477 : tensor<1x180x320x3xf32>, tensor<1x180x320x1xf32>, tensor<1x90x160x16xf32>, tensor<1x45x80x20xf32>, tensor<1x23x40x40xf32>, tensor<1x12x20x64xf32> loc(#loc) + } loc(#loc) + "onnx.EntryPoint"() {func = @main_graph} : () -> () loc(#loc) +} loc(#loc) +#loc1 = loc("Initializer_1026") +#loc2 = loc("Initializer_backbone.features.6.block.2.fc2.bias") +#loc3 = loc("Initializer_963") +#loc4 = loc("Initializer_945") +#loc5 = loc("Initializer_966") +#loc6 = loc("Initializer_1011") +#loc7 = loc("Initializer_1017") +#loc8 = loc("Initializer_388") +#loc9 = loc("Initializer_aspp.aspp2.1.weight") +#loc10 = loc("Initializer_386") +#loc11 = loc("Initializer_backbone.features.5.block.2.fc1.bias") +#loc12 = loc("Initializer_1083") +#loc13 = loc("Initializer_1082") +#loc14 = loc("Initializer_1080") +#loc15 = loc("Initializer_944") +#loc16 = loc("Initializer_1079") +#loc17 = loc("Initializer_1077") +#loc18 = loc("Initializer_1076") +#loc19 = loc("Initializer_1074") +#loc20 = loc("Initializer_1067") +#loc21 = loc("Initializer_1071") +#loc22 = loc("Initializer_1070") +#loc23 = loc("Initializer_400") +#loc24 = loc("Initializer_1064") +#loc25 = loc("Initializer_1061") +#loc26 = loc("Initializer_1058") +#loc27 = loc("Initializer_1056") +#loc28 = loc("Initializer_backbone.features.15.block.2.fc2.weight") +#loc29 = loc("Initializer_1053") +#loc30 = loc("Initializer_backbone.features.14.block.2.fc2.weight") +#loc31 = loc("Initializer_1049") +#loc32 = loc("Initializer_1044") +#loc33 = loc("Initializer_1043") +#loc34 = loc("Initializer_954") +#loc35 = loc("Initializer_974") +#loc36 = loc("Initializer_959") +#loc37 = loc("Initializer_972") +#loc38 = loc("Initializer_backbone.features.5.block.2.fc2.bias") +#loc39 = loc("Initializer_998") +#loc40 = loc("Initializer_1041") +#loc41 = loc("Initializer_983") +#loc42 = loc("Initializer_986") +#loc43 = loc("Initializer_backbone.features.11.block.2.fc2.weight") +#loc44 = loc("Initializer_backbone.features.11.block.2.fc1.bias") +#loc45 = loc("Initializer_1014") +#loc46 = loc("Initializer_decoder.decode1.gru.ih.0.weight") +#loc47 = loc("Initializer_978") +#loc48 = loc("Initializer_backbone.features.13.block.2.fc2.bias") +#loc49 = loc("Initializer_992") +#loc50 = loc("Initializer_1013") +#loc51 = loc("Initializer_995") +#loc52 = loc("Initializer_951") +#loc53 = loc("Initializer_999") +#loc54 = loc("Initializer_1023") +#loc55 = loc("Initializer_935") +#loc56 = loc("Initializer_decoder.decode4.gru.hh.0.weight") +#loc57 = loc("Initializer_backbone.features.11.block.2.fc2.bias") +#loc58 = loc("Initializer_971") +#loc59 = loc("Initializer_1090") +#loc60 = loc("Initializer_backbone.features.14.block.2.fc2.bias") +#loc61 = loc("Initializer_990") +#loc62 = loc("Initializer_decoder.decode1.gru.hh.0.weight") +#loc63 = loc("Initializer_752") +#loc64 = loc("Initializer_project_mat.conv.bias") +#loc65 = loc("Initializer_763") +#loc66 = loc("Initializer_948") +#loc67 = loc("Initializer_996") +#loc68 = loc("Initializer_950") +#loc69 = loc("Initializer_969") +#loc70 = loc("Initializer_977") +#loc71 = loc("Initializer_987") +#loc72 = loc("Initializer_backbone.features.14.block.2.fc1.bias") +#loc73 = loc("Initializer_1028") +#loc74 = loc("Initializer_965") +#loc75 = loc("Initializer_809") +#loc76 = loc("Initializer_929") +#loc77 = loc("Initializer_980") +#loc78 = loc("Initializer_1086") +#loc79 = loc("Initializer_1050") +#loc80 = loc("Initializer_942") +#loc81 = loc("Initializer_1004") +#loc82 = loc("Initializer_975") +#loc83 = loc("Initializer_847") +#loc84 = loc("Initializer_1073") +#loc85 = loc("Initializer_backbone.features.12.block.2.fc2.weight") +#loc86 = loc("Initializer_755") +#loc87 = loc("Initializer_1008") +#loc88 = loc("Initializer_backbone.features.13.block.2.fc1.bias") +#loc89 = loc("Initializer_1034") +#loc90 = loc("Initializer_1025") +#loc91 = loc("Initializer_1035") +#loc92 = loc("Initializer_890") +#loc93 = loc("Initializer_953") +#loc94 = loc("Initializer_1032") +#loc95 = loc("Initializer_backbone.features.6.block.2.fc2.weight") +#loc96 = loc("Initializer_decoder.decode1.gru.ih.0.bias") +#loc97 = loc("Initializer_1002") +#loc98 = loc("Initializer_981") +#loc99 = loc("Initializer_389") +#loc100 = loc("Initializer_backbone.features.11.block.2.fc1.weight") +#loc101 = loc("Initializer_1040") +#loc102 = loc("Initializer_1005") +#loc103 = loc("Initializer_1001") +#loc104 = loc("Initializer_1019") +#loc105 = loc("Initializer_1007") +#loc106 = loc("Initializer_backbone.features.13.block.2.fc2.weight") +#loc107 = loc("Initializer_984") +#loc108 = loc("Initializer_decoder.decode3.gru.ih.0.weight") +#loc109 = loc("Initializer_project_mat.conv.weight") +#loc110 = loc("Initializer_947") +#loc111 = loc("Initializer_1059") +#loc112 = loc("Initializer_941") +#loc113 = loc("Initializer_backbone.features.15.block.2.fc1.weight") +#loc114 = loc("Initializer_1038") +#loc115 = loc("Initializer_decoder.decode4.gru.hh.0.bias") +#loc116 = loc("Initializer_1016") +#loc117 = loc("Initializer_1022") +#loc118 = loc("Initializer_backbone.features.12.block.2.fc2.bias") +#loc119 = loc("Initializer_936") +#loc120 = loc("Initializer_decoder.decode2.gru.ih.0.bias") +#loc121 = loc("Initializer_decoder.decode4.gru.ih.0.weight") +#loc122 = loc("Initializer_1037") +#loc123 = loc("Initializer_backbone.features.5.block.2.fc1.weight") +#loc124 = loc("Initializer_backbone.features.6.block.2.fc1.bias") +#loc125 = loc("Initializer_938") +#loc126 = loc("Initializer_1031") +#loc127 = loc("Initializer_backbone.features.12.block.2.fc1.weight") +#loc128 = loc("Initializer_backbone.features.13.block.2.fc1.weight") +#loc129 = loc("Initializer_993") +#loc130 = loc("Initializer_1020") +#loc131 = loc("Initializer_backbone.features.5.block.2.fc2.weight") +#loc132 = loc("Initializer_1029") +#loc133 = loc("Initializer_845") +#loc134 = loc("Initializer_decoder.decode3.gru.ih.0.bias") +#loc135 = loc("Initializer_1047") +#loc136 = loc("Initializer_backbone.features.4.block.2.fc1.weight") +#loc137 = loc("Initializer_backbone.features.15.block.2.fc1.bias") +#loc138 = loc("Initializer_962") +#loc139 = loc("Initializer_backbone.features.12.block.2.fc1.bias") +#loc140 = loc("Initializer_backbone.features.6.block.2.fc1.weight") +#loc141 = loc("Initializer_decoder.decode4.gru.ih.0.bias") +#loc142 = loc("Initializer_1068") +#loc143 = loc("Initializer_960") +#loc144 = loc("Initializer_backbone.features.4.block.2.fc2.bias") +#loc145 = loc("Initializer_backbone.features.15.block.2.fc2.bias") +#loc146 = loc("Initializer_1062") +#loc147 = loc("Initializer_989") +#loc148 = loc("Initializer_decoder.decode2.gru.hh.0.weight") +#loc149 = loc("Initializer_1010") +#loc150 = loc("Initializer_1052") +#loc151 = loc("Initializer_backbone.features.4.block.2.fc1.bias") +#loc152 = loc("Initializer_1065") +#loc153 = loc("Initializer_decoder.decode3.gru.hh.0.weight") +#loc154 = loc("Initializer_backbone.features.4.block.2.fc2.weight") +#loc155 = loc("Initializer_933") +#loc156 = loc("Initializer_decoder.decode3.gru.hh.0.bias") +#loc157 = loc("Initializer_decoder.decode2.gru.ih.0.weight") +#loc158 = loc("Initializer_1046") +#loc159 = loc("Initializer_932") +#loc160 = loc("Initializer_956") +#loc161 = loc("Initializer_backbone.features.14.block.2.fc1.weight") +#loc162 = loc("Initializer_decoder.decode2.gru.hh.0.bias") +#loc163 = loc("Initializer_957") +#loc164 = loc("Initializer_decoder.decode1.gru.hh.0.bias") +#loc165 = loc("Initializer_968") +#loc166 = loc("Initializer_1055") +#loc167 = loc("Initializer_930") +#loc168 = loc("Initializer_939") +#loc169 = loc("Transpose_9") +#loc170 = loc("Transpose_10") +#loc171 = loc("Transpose_11") +#loc172 = loc("Cast_0") +#loc173 = loc("Div_2") +#loc174 = loc("Slice_7") +#loc175 = loc("Transpose_8") +#loc176 = loc("AveragePool_346") +#loc177 = loc("AveragePool_347") +#loc178 = loc("AveragePool_348") +#loc179 = loc("Transpose_12") +#loc180 = loc("Sub_14") +#loc181 = loc("Initializer_398") +#loc182 = loc("Div_16") +#loc183 = loc("Conv_17") +#loc184 = loc("Add_19") +#loc185 = loc("Clip_22") +#loc186 = loc("Div_24") +#loc187 = loc("Mul_25") +#loc188 = loc("Conv_26") +#loc189 = loc("Relu_27") +#loc190 = loc("Conv_28") +#loc191 = loc("Add_29") +#loc192 = loc("Conv_30") +#loc193 = loc("Relu_31") +#loc194 = loc("Conv_32") +#loc195 = loc("Relu_33") +#loc196 = loc("Conv_34") +#loc197 = loc("Conv_35") +#loc198 = loc("Relu_36") +#loc199 = loc("Conv_37") +#loc200 = loc("Relu_38") +#loc201 = loc("Conv_39") +#loc202 = loc("Add_40") +#loc203 = loc("Conv_41") +#loc204 = loc("Relu_42") +#loc205 = loc("Conv_43") +#loc206 = loc("Relu_44") +#loc207 = loc("GlobalAveragePool_45") +#loc208 = loc("Conv_46") +#loc209 = loc("Relu_47") +#loc210 = loc("Conv_48") +#loc211 = loc("Add_50") +#loc212 = loc("Clip_53") +#loc213 = loc("Div_55") +#loc214 = loc("Mul_56") +#loc215 = loc("Conv_57") +#loc216 = loc("Conv_58") +#loc217 = loc("Relu_59") +#loc218 = loc("Conv_60") +#loc219 = loc("Relu_61") +#loc220 = loc("GlobalAveragePool_62") +#loc221 = loc("Conv_63") +#loc222 = loc("Relu_64") +#loc223 = loc("Conv_65") +#loc224 = loc("Add_67") +#loc225 = loc("Clip_70") +#loc226 = loc("Div_72") +#loc227 = loc("Mul_73") +#loc228 = loc("Conv_74") +#loc229 = loc("Add_75") +#loc230 = loc("Conv_76") +#loc231 = loc("Relu_77") +#loc232 = loc("Conv_78") +#loc233 = loc("Relu_79") +#loc234 = loc("GlobalAveragePool_80") +#loc235 = loc("Conv_81") +#loc236 = loc("Relu_82") +#loc237 = loc("Conv_83") +#loc238 = loc("Add_85") +#loc239 = loc("Clip_88") +#loc240 = loc("Div_90") +#loc241 = loc("Mul_91") +#loc242 = loc("Conv_92") +#loc243 = loc("Add_93") +#loc244 = loc("Conv_94") +#loc245 = loc("Add_96") +#loc246 = loc("Clip_99") +#loc247 = loc("Div_101") +#loc248 = loc("Mul_102") +#loc249 = loc("Conv_103") +#loc250 = loc("Add_105") +#loc251 = loc("Clip_108") +#loc252 = loc("Div_110") +#loc253 = loc("Mul_111") +#loc254 = loc("Conv_112") +#loc255 = loc("Conv_113") +#loc256 = loc("Add_115") +#loc257 = loc("Clip_118") +#loc258 = loc("Div_120") +#loc259 = loc("Mul_121") +#loc260 = loc("Conv_122") +#loc261 = loc("Add_124") +#loc262 = loc("Clip_127") +#loc263 = loc("Div_129") +#loc264 = loc("Mul_130") +#loc265 = loc("Conv_131") +#loc266 = loc("Add_132") +#loc267 = loc("Conv_133") +#loc268 = loc("Add_135") +#loc269 = loc("Clip_138") +#loc270 = loc("Div_140") +#loc271 = loc("Mul_141") +#loc272 = loc("Conv_142") +#loc273 = loc("Add_144") +#loc274 = loc("Clip_147") +#loc275 = loc("Div_149") +#loc276 = loc("Mul_150") +#loc277 = loc("Conv_151") +#loc278 = loc("Add_152") +#loc279 = loc("Conv_153") +#loc280 = loc("Add_155") +#loc281 = loc("Clip_158") +#loc282 = loc("Div_160") +#loc283 = loc("Mul_161") +#loc284 = loc("Conv_162") +#loc285 = loc("Add_164") +#loc286 = loc("Clip_167") +#loc287 = loc("Div_169") +#loc288 = loc("Mul_170") +#loc289 = loc("Conv_171") +#loc290 = loc("Add_172") +#loc291 = loc("Conv_173") +#loc292 = loc("Add_175") +#loc293 = loc("Clip_178") +#loc294 = loc("Div_180") +#loc295 = loc("Mul_181") +#loc296 = loc("Conv_182") +#loc297 = loc("Add_184") +#loc298 = loc("Clip_187") +#loc299 = loc("Div_189") +#loc300 = loc("Mul_190") +#loc301 = loc("GlobalAveragePool_191") +#loc302 = loc("Conv_192") +#loc303 = loc("Relu_193") +#loc304 = loc("Conv_194") +#loc305 = loc("Add_196") +#loc306 = loc("Clip_199") +#loc307 = loc("Div_201") +#loc308 = loc("Mul_202") +#loc309 = loc("Conv_203") +#loc310 = loc("Conv_204") +#loc311 = loc("Add_206") +#loc312 = loc("Clip_209") +#loc313 = loc("Div_211") +#loc314 = loc("Mul_212") +#loc315 = loc("Conv_213") +#loc316 = loc("Add_215") +#loc317 = loc("Clip_218") +#loc318 = loc("Div_220") +#loc319 = loc("Mul_221") +#loc320 = loc("GlobalAveragePool_222") +#loc321 = loc("Conv_223") +#loc322 = loc("Relu_224") +#loc323 = loc("Conv_225") +#loc324 = loc("Add_227") +#loc325 = loc("Clip_230") +#loc326 = loc("Div_232") +#loc327 = loc("Mul_233") +#loc328 = loc("Conv_234") +#loc329 = loc("Add_235") +#loc330 = loc("Conv_236") +#loc331 = loc("Add_238") +#loc332 = loc("Clip_241") +#loc333 = loc("Div_243") +#loc334 = loc("Mul_244") +#loc335 = loc("Conv_245") +#loc336 = loc("Add_247") +#loc337 = loc("Clip_250") +#loc338 = loc("Div_252") +#loc339 = loc("Mul_253") +#loc340 = loc("GlobalAveragePool_254") +#loc341 = loc("Conv_255") +#loc342 = loc("Relu_256") +#loc343 = loc("Conv_257") +#loc344 = loc("Add_259") +#loc345 = loc("Clip_262") +#loc346 = loc("Div_264") +#loc347 = loc("Mul_265") +#loc348 = loc("Conv_266") +#loc349 = loc("Conv_267") +#loc350 = loc("Add_269") +#loc351 = loc("Clip_272") +#loc352 = loc("Div_274") +#loc353 = loc("Mul_275") +#loc354 = loc("Conv_276") +#loc355 = loc("Add_278") +#loc356 = loc("Clip_281") +#loc357 = loc("Div_283") +#loc358 = loc("Mul_284") +#loc359 = loc("GlobalAveragePool_285") +#loc360 = loc("Conv_286") +#loc361 = loc("Relu_287") +#loc362 = loc("Conv_288") +#loc363 = loc("Add_290") +#loc364 = loc("Clip_293") +#loc365 = loc("Div_295") +#loc366 = loc("Mul_296") +#loc367 = loc("Conv_297") +#loc368 = loc("Add_298") +#loc369 = loc("Conv_299") +#loc370 = loc("Add_301") +#loc371 = loc("Clip_304") +#loc372 = loc("Div_306") +#loc373 = loc("Mul_307") +#loc374 = loc("Conv_308") +#loc375 = loc("Add_310") +#loc376 = loc("Clip_313") +#loc377 = loc("Div_315") +#loc378 = loc("Mul_316") +#loc379 = loc("GlobalAveragePool_317") +#loc380 = loc("Conv_318") +#loc381 = loc("Relu_319") +#loc382 = loc("Conv_320") +#loc383 = loc("Add_322") +#loc384 = loc("Clip_325") +#loc385 = loc("Div_327") +#loc386 = loc("Mul_328") +#loc387 = loc("Conv_329") +#loc388 = loc("Add_330") +#loc389 = loc("Conv_331") +#loc390 = loc("Add_333") +#loc391 = loc("Clip_336") +#loc392 = loc("Div_338") +#loc393 = loc("Mul_339") +#loc394 = loc("Conv_340") +#loc395 = loc("Relu_341") +#loc396 = loc("GlobalAveragePool_342") +#loc397 = loc("Conv_343") +#loc398 = loc("Sigmoid_344") +#loc399 = loc("Mul_345") +#loc400 = loc("Split_349") +#loc401 = loc("Concat_350") +#loc402 = loc("Conv_351") +#loc403 = loc("Sigmoid_352") +#loc404 = loc("Split_353") +#loc405 = loc("Sub_359") +#loc406 = loc("Mul_360") +#loc407 = loc("Mul_354") +#loc408 = loc("Concat_355") +#loc409 = loc("Conv_356") +#loc410 = loc("Tanh_357") +#loc411 = loc("Mul_361") +#loc412 = loc("Add_362") +#loc413 = loc("Concat_363") +#loc414 = loc("Resize_365") +#loc415 = loc("Slice_371") +#loc416 = loc("Concat_372") +#loc417 = loc("Conv_373") +#loc418 = loc("Relu_374") +#loc419 = loc("Split_375") +#loc420 = loc("Concat_376") +#loc421 = loc("Conv_377") +#loc422 = loc("Sigmoid_378") +#loc423 = loc("Split_379") +#loc424 = loc("Sub_385") +#loc425 = loc("Mul_386") +#loc426 = loc("Mul_380") +#loc427 = loc("Concat_381") +#loc428 = loc("Conv_382") +#loc429 = loc("Tanh_383") +#loc430 = loc("Mul_387") +#loc431 = loc("Add_388") +#loc432 = loc("Concat_389") +#loc433 = loc("Resize_391") +#loc434 = loc("Slice_397") +#loc435 = loc("Concat_398") +#loc436 = loc("Conv_399") +#loc437 = loc("Relu_400") +#loc438 = loc("Split_401") +#loc439 = loc("Concat_402") +#loc440 = loc("Conv_403") +#loc441 = loc("Sigmoid_404") +#loc442 = loc("Split_405") +#loc443 = loc("Sub_411") +#loc444 = loc("Mul_412") +#loc445 = loc("Mul_406") +#loc446 = loc("Concat_407") +#loc447 = loc("Conv_408") +#loc448 = loc("Tanh_409") +#loc449 = loc("Mul_413") +#loc450 = loc("Add_414") +#loc451 = loc("Concat_415") +#loc452 = loc("Resize_417") +#loc453 = loc("Concat_418") +#loc454 = loc("Conv_419") +#loc455 = loc("Relu_420") +#loc456 = loc("Split_421") +#loc457 = loc("Concat_422") +#loc458 = loc("Conv_423") +#loc459 = loc("Sigmoid_424") +#loc460 = loc("Split_425") +#loc461 = loc("Sub_431") +#loc462 = loc("Mul_432") +#loc463 = loc("Mul_426") +#loc464 = loc("Concat_427") +#loc465 = loc("Conv_428") +#loc466 = loc("Tanh_429") +#loc467 = loc("Mul_433") +#loc468 = loc("Add_434") +#loc469 = loc("Transpose_448") +#loc470 = loc("Transpose_449") +#loc471 = loc("Transpose_450") +#loc472 = loc("Transpose_451") +#loc473 = loc("Concat_435") +#loc474 = loc("Resize_437") +#loc475 = loc("Concat_438") +#loc476 = loc("Conv_439") +#loc477 = loc("Relu_440") +#loc478 = loc("Conv_441") +#loc479 = loc("Relu_442") +#loc480 = loc("Conv_443") +#loc481 = loc("Split_444") +#loc482 = loc("Add_445") +#loc483 = loc("Clip_446") +#loc484 = loc("Transpose_452") +#loc485 = loc("Clip_447") +#loc486 = loc("Transpose_453") +#loc487 = loc(fused[#loc180, #loc181]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.par.x86.unwrapped.mlir b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.par.x86.unwrapped.mlir new file mode 100644 index 0000000000000000000000000000000000000000..5c915cd330108ef2e1b1f0a4e6cd25db750d7971 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.par.x86.unwrapped.mlir @@ -0,0 +1,348 @@ +#loc = loc(unknown) +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func private @forward_outlined_part_0(tensor<1x180x320x4xbf16>, tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) attributes {aie_partition = 0 : i32, kernel} loc(#loc319) + func.func @forward(%arg0: tensor<1x180x320x4xui8> {onnx.name = "src"} loc(unknown), %arg1: tensor<1x90x160x16xbf16> {onnx.name = "r1i"} loc(unknown), %arg2: tensor<1x45x80x20xbf16> {onnx.name = "r2i"} loc(unknown), %arg3: tensor<1x23x40x40xbf16> {onnx.name = "r3i"} loc(unknown), %arg4: tensor<1x12x20x64xbf16> {onnx.name = "r4i"} loc(unknown)) -> (tensor<1x180x320x3xbf16> {onnx.name = "fgr"}, tensor<1x180x320x1xbf16> {onnx.name = "pha"}, tensor<1x90x160x16xbf16> {onnx.name = "r1o"}, tensor<1x45x80x20xbf16> {onnx.name = "r2o"}, tensor<1x23x40x40xbf16> {onnx.name = "r3o"}, tensor<1x12x20x64xbf16> {onnx.name = "r4o"}) { + %0 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %1 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %2 = tosa.transpose %arg1, %1 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc308) + %3 = tosa.transpose %arg2, %1 : (tensor<1x45x80x20xbf16>, tensor<4xi32>) -> tensor<1x20x45x80xbf16> loc(#loc309) + %4 = tosa.transpose %arg3, %1 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc310) + %5 = tosa.cast %arg0 {LayerName = "Cast_0", OutputName = "Cast_0"} : (tensor<1x180x320x4xui8>) -> tensor<1x180x320x4xbf16> loc(#loc311) + %6 = tosa.transpose %arg4, %1 : (tensor<1x12x20x64xbf16>, tensor<4xi32>) -> tensor<1x64x12x20xbf16> loc(#loc312) + %7:6 = call @forward_outlined_part_0(%5, %2, %3, %4, %6) : (tensor<1x180x320x4xbf16>, tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) loc(#loc319) + %8 = tosa.transpose %7#4, %0 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc313) + %9 = tosa.transpose %7#3, %0 : (tensor<1x64x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x64xbf16> loc(#loc314) + %10 = tosa.transpose %7#2, %0 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc315) + %11 = tosa.transpose %7#1, %0 : (tensor<1x20x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x20xbf16> loc(#loc316) + %12 = tosa.transpose %7#0, %0 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc317) + %13 = tosa.reshape %7#5 {new_shape = array} : (tensor<1x1x180x320xbf16>) -> tensor<1x180x320x1xbf16> loc(#loc318) + return %8, %13, %12, %11, %10, %9 : tensor<1x180x320x3xbf16>, tensor<1x180x320x1xbf16>, tensor<1x90x160x16xbf16>, tensor<1x45x80x20xbf16>, tensor<1x23x40x40xbf16>, tensor<1x12x20x64xbf16> loc(#loc) + } loc(#loc) +} loc(#loc) +#loc1 = loc("Div_2") +#loc2 = loc("Sub_431") +#loc3 = loc("Sub_411") +#loc4 = loc("Sub_385") +#loc5 = loc("Sub_359") +#loc6 = loc("Div_16") +#loc7 = loc("Sub_14") +#loc8 = loc("Initializer_398") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("AveragePool_347") +#loc14 = loc("AveragePool_348") +#loc15 = loc("Conv_17") +#loc16 = loc("Add_19") +#loc17 = loc("Clip_22") +#loc18 = loc("Div_24") +#loc19 = loc("Mul_25") +#loc20 = loc("Conv_26") +#loc21 = loc("Relu_27") +#loc22 = loc("Conv_28") +#loc23 = loc("Add_29") +#loc24 = loc("Conv_30") +#loc25 = loc("Relu_31") +#loc26 = loc("Conv_32") +#loc27 = loc("Relu_33") +#loc28 = loc("Conv_34") +#loc29 = loc("Conv_35") +#loc30 = loc("Relu_36") +#loc31 = loc("Conv_37") +#loc32 = loc("Relu_38") +#loc33 = loc("Conv_39") +#loc34 = loc("Add_40") +#loc35 = loc("Conv_41") +#loc36 = loc("Relu_42") +#loc37 = loc("Conv_43") +#loc38 = loc("Relu_44") +#loc39 = loc("GlobalAveragePool_45") +#loc40 = loc("Conv_46") +#loc41 = loc("Relu_47") +#loc42 = loc("Conv_48") +#loc43 = loc("Add_50") +#loc44 = loc("Clip_53") +#loc45 = loc("Div_55") +#loc46 = loc("Mul_56") +#loc47 = loc("Conv_57") +#loc48 = loc("Conv_58") +#loc49 = loc("Relu_59") +#loc50 = loc("Conv_60") +#loc51 = loc("Relu_61") +#loc52 = loc("GlobalAveragePool_62") +#loc53 = loc("Conv_63") +#loc54 = loc("Relu_64") +#loc55 = loc("Conv_65") +#loc56 = loc("Add_67") +#loc57 = loc("Clip_70") +#loc58 = loc("Div_72") +#loc59 = loc("Mul_73") +#loc60 = loc("Conv_74") +#loc61 = loc("Add_75") +#loc62 = loc("Conv_76") +#loc63 = loc("Relu_77") +#loc64 = loc("Conv_78") +#loc65 = loc("Relu_79") +#loc66 = loc("GlobalAveragePool_80") +#loc67 = loc("Conv_81") +#loc68 = loc("Relu_82") +#loc69 = loc("Conv_83") +#loc70 = loc("Add_85") +#loc71 = loc("Clip_88") +#loc72 = loc("Div_90") +#loc73 = loc("Mul_91") +#loc74 = loc("Conv_92") +#loc75 = loc("Add_93") +#loc76 = loc("Conv_94") +#loc77 = loc("Add_96") +#loc78 = loc("Clip_99") +#loc79 = loc("Div_101") +#loc80 = loc("Mul_102") +#loc81 = loc("Conv_103") +#loc82 = loc("Add_105") +#loc83 = loc("Clip_108") +#loc84 = loc("Div_110") +#loc85 = loc("Mul_111") +#loc86 = loc("Conv_112") +#loc87 = loc("Conv_113") +#loc88 = loc("Add_115") +#loc89 = loc("Clip_118") +#loc90 = loc("Div_120") +#loc91 = loc("Mul_121") +#loc92 = loc("Conv_122") +#loc93 = loc("Add_124") +#loc94 = loc("Clip_127") +#loc95 = loc("Div_129") +#loc96 = loc("Mul_130") +#loc97 = loc("Conv_131") +#loc98 = loc("Add_132") +#loc99 = loc("Conv_133") +#loc100 = loc("Add_135") +#loc101 = loc("Clip_138") +#loc102 = loc("Div_140") +#loc103 = loc("Mul_141") +#loc104 = loc("Conv_142") +#loc105 = loc("Add_144") +#loc106 = loc("Clip_147") +#loc107 = loc("Div_149") +#loc108 = loc("Mul_150") +#loc109 = loc("Conv_151") +#loc110 = loc("Add_152") +#loc111 = loc("Conv_153") +#loc112 = loc("Add_155") +#loc113 = loc("Clip_158") +#loc114 = loc("Div_160") +#loc115 = loc("Mul_161") +#loc116 = loc("Conv_162") +#loc117 = loc("Add_164") +#loc118 = loc("Clip_167") +#loc119 = loc("Div_169") +#loc120 = loc("Mul_170") +#loc121 = loc("Conv_171") +#loc122 = loc("Add_172") +#loc123 = loc("Conv_173") +#loc124 = loc("Add_175") +#loc125 = loc("Clip_178") +#loc126 = loc("Div_180") +#loc127 = loc("Mul_181") +#loc128 = loc("Conv_182") +#loc129 = loc("Add_184") +#loc130 = loc("Clip_187") +#loc131 = loc("Div_189") +#loc132 = loc("Mul_190") +#loc133 = loc("GlobalAveragePool_191") +#loc134 = loc("Conv_192") +#loc135 = loc("Relu_193") +#loc136 = loc("Conv_194") +#loc137 = loc("Add_196") +#loc138 = loc("Clip_199") +#loc139 = loc("Div_201") +#loc140 = loc("Mul_202") +#loc141 = loc("Conv_203") +#loc142 = loc("Conv_204") +#loc143 = loc("Add_206") +#loc144 = loc("Clip_209") +#loc145 = loc("Div_211") +#loc146 = loc("Mul_212") +#loc147 = loc("Conv_213") +#loc148 = loc("Add_215") +#loc149 = loc("Clip_218") +#loc150 = loc("Div_220") +#loc151 = loc("Mul_221") +#loc152 = loc("GlobalAveragePool_222") +#loc153 = loc("Conv_223") +#loc154 = loc("Relu_224") +#loc155 = loc("Conv_225") +#loc156 = loc("Add_227") +#loc157 = loc("Clip_230") +#loc158 = loc("Div_232") +#loc159 = loc("Mul_233") +#loc160 = loc("Conv_234") +#loc161 = loc("Add_235") +#loc162 = loc("Conv_236") +#loc163 = loc("Add_238") +#loc164 = loc("Clip_241") +#loc165 = loc("Div_243") +#loc166 = loc("Mul_244") +#loc167 = loc("Conv_245") +#loc168 = loc("Add_247") +#loc169 = loc("Clip_250") +#loc170 = loc("Div_252") +#loc171 = loc("Mul_253") +#loc172 = loc("GlobalAveragePool_254") +#loc173 = loc("Conv_255") +#loc174 = loc("Relu_256") +#loc175 = loc("Conv_257") +#loc176 = loc("Add_259") +#loc177 = loc("Clip_262") +#loc178 = loc("Div_264") +#loc179 = loc("Mul_265") +#loc180 = loc("Conv_266") +#loc181 = loc("Conv_267") +#loc182 = loc("Add_269") +#loc183 = loc("Clip_272") +#loc184 = loc("Div_274") +#loc185 = loc("Mul_275") +#loc186 = loc("Conv_276") +#loc187 = loc("Add_278") +#loc188 = loc("Clip_281") +#loc189 = loc("Div_283") +#loc190 = loc("Mul_284") +#loc191 = loc("GlobalAveragePool_285") +#loc192 = loc("Conv_286") +#loc193 = loc("Relu_287") +#loc194 = loc("Conv_288") +#loc195 = loc("Add_290") +#loc196 = loc("Clip_293") +#loc197 = loc("Div_295") +#loc198 = loc("Mul_296") +#loc199 = loc("Conv_297") +#loc200 = loc("Add_298") +#loc201 = loc("Conv_299") +#loc202 = loc("Add_301") +#loc203 = loc("Clip_304") +#loc204 = loc("Div_306") +#loc205 = loc("Mul_307") +#loc206 = loc("Conv_308") +#loc207 = loc("Add_310") +#loc208 = loc("Clip_313") +#loc209 = loc("Div_315") +#loc210 = loc("Mul_316") +#loc211 = loc("GlobalAveragePool_317") +#loc212 = loc("Conv_318") +#loc213 = loc("Relu_319") +#loc214 = loc("Conv_320") +#loc215 = loc("Add_322") +#loc216 = loc("Clip_325") +#loc217 = loc("Div_327") +#loc218 = loc("Mul_328") +#loc219 = loc("Conv_329") +#loc220 = loc("Add_330") +#loc221 = loc("Conv_331") +#loc222 = loc("Add_333") +#loc223 = loc("Clip_336") +#loc224 = loc("Div_338") +#loc225 = loc("Mul_339") +#loc226 = loc("GlobalAveragePool_342") +#loc227 = loc("Conv_343") +#loc228 = loc("Sigmoid_344") +#loc229 = loc("Mul_345") +#loc230 = loc("Conv_340") +#loc231 = loc("Relu_341") +#loc232 = loc("Split_349") +#loc233 = loc("Concat_350") +#loc234 = loc("Conv_351") +#loc235 = loc("Sigmoid_352") +#loc236 = loc("Split_353") +#loc237 = loc("Mul_360") +#loc238 = loc("Mul_354") +#loc239 = loc("Concat_355") +#loc240 = loc("Conv_356") +#loc241 = loc("Tanh_357") +#loc242 = loc("Mul_361") +#loc243 = loc("Add_362") +#loc244 = loc("Concat_363") +#loc245 = loc("Resize_365") +#loc246 = loc("Slice_371") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_386") +#loc256 = loc("Mul_380") +#loc257 = loc("Concat_381") +#loc258 = loc("Conv_382") +#loc259 = loc("Tanh_383") +#loc260 = loc("Mul_387") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_412") +#loc274 = loc("Mul_406") +#loc275 = loc("Concat_407") +#loc276 = loc("Conv_408") +#loc277 = loc("Tanh_409") +#loc278 = loc("Mul_413") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_432") +#loc291 = loc("Mul_426") +#loc292 = loc("Concat_427") +#loc293 = loc("Conv_428") +#loc294 = loc("Tanh_429") +#loc295 = loc("Mul_433") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_446") +#loc307 = loc("Clip_447") +#loc308 = loc("Transpose_9") +#loc309 = loc("Transpose_10") +#loc310 = loc("Transpose_11") +#loc311 = loc("Cast_0") +#loc312 = loc("Transpose_12") +#loc313 = loc("Transpose_452") +#loc314 = loc("Transpose_451") +#loc315 = loc("Transpose_450") +#loc316 = loc("Transpose_449") +#loc317 = loc("Transpose_448") +#loc318 = loc("Transpose_453") +#loc319 = loc(fused[#loc1, #loc2, #loc3, #loc4, #loc5, #loc6, #loc7, #loc8, #loc9, #loc10, #loc11, #loc12, #loc13, #loc14, #loc15, #loc16, #loc17, #loc18, #loc19, #loc20, #loc21, #loc22, #loc23, #loc24, #loc25, #loc26, #loc27, #loc28, #loc29, #loc30, #loc31, #loc32, #loc33, #loc34, #loc35, #loc36, #loc37, #loc38, #loc39, #loc40, #loc41, #loc42, #loc43, #loc44, #loc45, #loc46, #loc47, #loc48, #loc49, #loc50, #loc51, #loc52, #loc53, #loc54, #loc55, #loc56, #loc57, #loc58, #loc59, #loc60, #loc61, #loc62, #loc63, #loc64, #loc65, #loc66, #loc67, #loc68, #loc69, #loc70, #loc71, #loc72, #loc73, #loc74, #loc75, #loc76, #loc77, #loc78, #loc79, #loc80, #loc81, #loc82, #loc83, #loc84, #loc85, #loc86, #loc87, #loc88, #loc89, #loc90, #loc91, #loc92, #loc93, #loc94, #loc95, #loc96, #loc97, #loc98, #loc99, #loc100, #loc101, #loc102, #loc103, #loc104, #loc105, #loc106, #loc107, #loc108, #loc109, #loc110, #loc111, #loc112, #loc113, #loc114, #loc115, #loc116, #loc117, #loc118, #loc119, #loc120, #loc121, #loc122, #loc123, #loc124, #loc125, #loc126, #loc127, #loc128, #loc129, #loc130, #loc131, #loc132, #loc133, #loc134, #loc135, #loc136, #loc137, #loc138, #loc139, #loc140, #loc141, #loc142, #loc143, #loc144, #loc145, #loc146, #loc147, #loc148, #loc149, #loc150, #loc151, #loc152, #loc153, #loc154, #loc155, #loc156, #loc157, #loc158, #loc159, #loc160, #loc161, #loc162, #loc163, #loc164, #loc165, #loc166, #loc167, #loc168, #loc169, #loc170, #loc171, #loc172, #loc173, #loc174, #loc175, #loc176, #loc177, #loc178, #loc179, #loc180, #loc181, #loc182, #loc183, #loc184, #loc185, #loc186, #loc187, #loc188, #loc189, #loc190, #loc191, #loc192, #loc193, #loc194, #loc195, #loc196, #loc197, #loc198, #loc199, #loc200, #loc201, #loc202, #loc203, #loc204, #loc205, #loc206, #loc207, #loc208, #loc209, #loc210, #loc211, #loc212, #loc213, #loc214, #loc215, #loc216, #loc217, #loc218, #loc219, #loc220, #loc221, #loc222, #loc223, #loc224, #loc225, #loc226, #loc227, #loc228, #loc229, #loc230, #loc231, #loc232, #loc233, #loc234, #loc235, #loc236, #loc237, #loc238, #loc239, #loc240, #loc241, #loc242, #loc243, #loc244, #loc245, #loc246, #loc247, #loc248, #loc249, #loc250, #loc251, #loc252, #loc253, #loc254, #loc255, #loc256, #loc257, #loc258, #loc259, #loc260, #loc261, #loc262, #loc263, #loc264, #loc265, #loc266, #loc267, #loc268, #loc269, #loc270, #loc271, #loc272, #loc273, #loc274, #loc275, #loc276, #loc277, #loc278, #loc279, #loc280, #loc281, #loc282, #loc283, #loc284, #loc285, #loc286, #loc287, #loc288, #loc289, #loc290, #loc291, #loc292, #loc293, #loc294, #loc295, #loc296, #loc297, #loc298, #loc299, #loc300, #loc301, #loc302, #loc303, #loc304, #loc305, #loc306, #loc307]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.par.x86.wrapped.mlir b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.par.x86.wrapped.mlir new file mode 100644 index 0000000000000000000000000000000000000000..d65af80f089691f434bf38e8f42d568d7f124773 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/OnnxModel.par.x86.wrapped.mlir @@ -0,0 +1,26478 @@ +#loc = loc(unknown) +#loc308 = loc("Cast_0") +#loc309 = loc("Transpose_9") +#loc310 = loc("Transpose_10") +#loc311 = loc("Transpose_11") +#loc312 = loc("Transpose_12") +module attributes { + llvm.data_layout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-i128:128-f80:128-n8:16:32:64-S128", + llvm.target_triple = "x86_64-unknown-linux-gnu", + "onnx-mlir.symbol-postfix" = "onnxmodel.onnx.mlir", + vaimlconf.device = "stx", + vaimlconf.device_models = "${vaimlconf.install_dir}/data/deviceModels", + vaimlconf.install_dir = "/usr/local/lib/python3.10/dist-packages/flexml/flexml_extras", + vaimlconf.library_metadata = ["${vaimlconf.install_dir}/data/libraryMetadata/L1", "${vaimlconf.install_dir}/data/libraryMetadata/L2", "${vaimlconf.install_dir}/../../vitis_mllib/L1/metadata", "${vaimlconf.install_dir}/../../vitis_mllib/L2/metadata", "${vaimlconf.install_dir}/share/microkernel-tiling/tiling-recipe-specs"], + vaimlconf.single_core_compiler = "chess"} { + func.func private @forward_outlined_part_0(%arg0: tensor<1x180x320x4xbf16> loc("Cast_0"), %arg1: tensor<1x16x90x160xbf16> loc("Transpose_9"), %arg2: tensor<1x20x45x80xbf16> loc("Transpose_10"), %arg3: tensor<1x40x23x40xbf16> loc("Transpose_11"), %arg4: tensor<1x64x12x20xbf16> loc("Transpose_12")) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) attributes {aie_partition = 0 : i32, kernel} { + %0 = xten_nn.subgraph (%arg5 = %arg0: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "386", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Div_2", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "386", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Div_2", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.906250e-03 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.906250e-03> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_2", + OutputName = "Div_2", + shift = 0 : i8} : (tensor<1x180x320x4xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %463 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + xten_nn.output %461 : tensor<1x180x320x4xbf16> loc(#loc1) + } -> tensor<1x180x320x4xbf16> loc(#loc1) + %1 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/biases"} -> tensor<4xbf16> loc(#loc) + %2 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_443/weights"} -> tensor<4x16x1x1xbf16> loc(#loc) + %3 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/biases"} -> tensor<16xbf16> loc(#loc) + %4 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_441/weights"} -> tensor<16x16x3x3xbf16> loc(#loc) + %5 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/biases"} -> tensor<16xbf16> loc(#loc) + %6 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_439/weights"} -> tensor<16x35x3x3xbf16> loc(#loc) + %7 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/biases"} -> tensor<16xbf16> loc(#loc) + %8 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_428/weights"} -> tensor<16x32x3x3xbf16> loc(#loc) + %9 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_431/Constant_0_0"} -> tensor<1x16x90x160xbf16> loc(#loc2) + %10 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/biases"} -> tensor<32xbf16> loc(#loc) + %11 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_423/weights"} -> tensor<32x32x3x3xbf16> loc(#loc) + %12 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/biases"} -> tensor<32xbf16> loc(#loc) + %13 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_419/weights"} -> tensor<32x59x3x3xbf16> loc(#loc) + %14 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/biases"} -> tensor<20xbf16> loc(#loc) + %15 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_408/weights"} -> tensor<20x40x3x3xbf16> loc(#loc) + %16 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_411/Constant_0_0"} -> tensor<1x20x45x80xbf16> loc(#loc3) + %17 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/biases"} -> tensor<40xbf16> loc(#loc) + %18 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_403/weights"} -> tensor<40x40x3x3xbf16> loc(#loc) + %19 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/biases"} -> tensor<40xbf16> loc(#loc) + %20 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_399/weights"} -> tensor<40x107x3x3xbf16> loc(#loc) + %21 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/biases"} -> tensor<40xbf16> loc(#loc) + %22 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_382/weights"} -> tensor<40x80x3x3xbf16> loc(#loc) + %23 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_385/Constant_0_0"} -> tensor<1x40x23x40xbf16> loc(#loc4) + %24 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/biases"} -> tensor<80xbf16> loc(#loc) + %25 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_377/weights"} -> tensor<80x80x3x3xbf16> loc(#loc) + %26 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/biases"} -> tensor<80xbf16> loc(#loc) + %27 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_373/weights"} -> tensor<80x171x3x3xbf16> loc(#loc) + %28 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/biases"} -> tensor<64xbf16> loc(#loc) + %29 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_356/weights"} -> tensor<64x128x3x3xbf16> loc(#loc) + %30 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_359/Constant_0_0"} -> tensor<1x64x12x20xbf16> loc(#loc5) + %31 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/biases"} -> tensor<128xbf16> loc(#loc) + %32 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_351/weights"} -> tensor<128x128x3x3xbf16> loc(#loc) + %33 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/biases"} -> tensor<128xbf16> loc(#loc) + %34 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_340/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %35 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/biases"} -> tensor<128xbf16> loc(#loc) + %36 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_343/weights"} -> tensor<128x960x1x1xbf16> loc(#loc) + %37 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/biases"} -> tensor<960xbf16> loc(#loc) + %38 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_331/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %39 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/biases"} -> tensor<160xbf16> loc(#loc) + %40 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_329/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %41 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/biases"} -> tensor<960xbf16> loc(#loc) + %42 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_320/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %43 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/biases"} -> tensor<240xbf16> loc(#loc) + %44 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_318/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %45 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/biases"} -> tensor<960xbf16> loc(#loc) + %46 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_308/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %47 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/biases"} -> tensor<960xbf16> loc(#loc) + %48 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_299/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %49 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/biases"} -> tensor<160xbf16> loc(#loc) + %50 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_297/weights"} -> tensor<160x960x1x1xbf16> loc(#loc) + %51 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/biases"} -> tensor<960xbf16> loc(#loc) + %52 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_288/weights"} -> tensor<960x240x1x1xbf16> loc(#loc) + %53 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/biases"} -> tensor<240xbf16> loc(#loc) + %54 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_286/weights"} -> tensor<240x960x1x1xbf16> loc(#loc) + %55 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/biases"} -> tensor<960xbf16> loc(#loc) + %56 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_276/weights"} -> tensor<960x1x9x9xbf16> loc(#loc) + %57 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/biases"} -> tensor<960xbf16> loc(#loc) + %58 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_267/weights"} -> tensor<960x160x1x1xbf16> loc(#loc) + %59 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/biases"} -> tensor<160xbf16> loc(#loc) + %60 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_266/weights"} -> tensor<160x672x1x1xbf16> loc(#loc) + %61 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/biases"} -> tensor<672xbf16> loc(#loc) + %62 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_257/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %63 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/biases"} -> tensor<168xbf16> loc(#loc) + %64 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_255/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %65 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/biases"} -> tensor<672xbf16> loc(#loc) + %66 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_245/weights"} -> tensor<672x1x9x9xbf16> loc(#loc) + %67 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/biases"} -> tensor<672xbf16> loc(#loc) + %68 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_236/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %69 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/biases"} -> tensor<112xbf16> loc(#loc) + %70 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_234/weights"} -> tensor<112x672x1x1xbf16> loc(#loc) + %71 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/biases"} -> tensor<672xbf16> loc(#loc) + %72 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_225/weights"} -> tensor<672x168x1x1xbf16> loc(#loc) + %73 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/biases"} -> tensor<168xbf16> loc(#loc) + %74 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_223/weights"} -> tensor<168x672x1x1xbf16> loc(#loc) + %75 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/biases"} -> tensor<672xbf16> loc(#loc) + %76 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_213/weights"} -> tensor<672x1x3x3xbf16> loc(#loc) + %77 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/biases"} -> tensor<672xbf16> loc(#loc) + %78 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_204/weights"} -> tensor<672x112x1x1xbf16> loc(#loc) + %79 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/biases"} -> tensor<112xbf16> loc(#loc) + %80 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_203/weights"} -> tensor<112x480x1x1xbf16> loc(#loc) + %81 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/biases"} -> tensor<480xbf16> loc(#loc) + %82 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_194/weights"} -> tensor<480x120x1x1xbf16> loc(#loc) + %83 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/biases"} -> tensor<120xbf16> loc(#loc) + %84 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_192/weights"} -> tensor<120x480x1x1xbf16> loc(#loc) + %85 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/biases"} -> tensor<480xbf16> loc(#loc) + %86 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_182/weights"} -> tensor<480x1x3x3xbf16> loc(#loc) + %87 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/biases"} -> tensor<480xbf16> loc(#loc) + %88 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_173/weights"} -> tensor<480x80x1x1xbf16> loc(#loc) + %89 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/biases"} -> tensor<80xbf16> loc(#loc) + %90 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_171/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %91 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/biases"} -> tensor<184xbf16> loc(#loc) + %92 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_162/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %93 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/biases"} -> tensor<184xbf16> loc(#loc) + %94 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_153/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %95 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/biases"} -> tensor<80xbf16> loc(#loc) + %96 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_151/weights"} -> tensor<80x184x1x1xbf16> loc(#loc) + %97 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/biases"} -> tensor<184xbf16> loc(#loc) + %98 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_142/weights"} -> tensor<184x1x3x3xbf16> loc(#loc) + %99 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/biases"} -> tensor<184xbf16> loc(#loc) + %100 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_133/weights"} -> tensor<184x80x1x1xbf16> loc(#loc) + %101 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/biases"} -> tensor<80xbf16> loc(#loc) + %102 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_131/weights"} -> tensor<80x200x1x1xbf16> loc(#loc) + %103 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/biases"} -> tensor<200xbf16> loc(#loc) + %104 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_122/weights"} -> tensor<200x1x3x3xbf16> loc(#loc) + %105 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/biases"} -> tensor<200xbf16> loc(#loc) + %106 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_113/weights"} -> tensor<200x80x1x1xbf16> loc(#loc) + %107 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/biases"} -> tensor<80xbf16> loc(#loc) + %108 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_112/weights"} -> tensor<80x240x1x1xbf16> loc(#loc) + %109 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/biases"} -> tensor<240xbf16> loc(#loc) + %110 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_103/weights"} -> tensor<240x1x3x3xbf16> loc(#loc) + %111 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/biases"} -> tensor<240xbf16> loc(#loc) + %112 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_94/weights"} -> tensor<240x40x1x1xbf16> loc(#loc) + %113 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/biases"} -> tensor<40xbf16> loc(#loc) + %114 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_92/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %115 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/biases"} -> tensor<120xbf16> loc(#loc) + %116 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_83/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %117 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/biases"} -> tensor<32xbf16> loc(#loc) + %118 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_81/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %119 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/biases"} -> tensor<120xbf16> loc(#loc) + %120 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_78/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %121 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/biases"} -> tensor<120xbf16> loc(#loc) + %122 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_76/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %123 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/biases"} -> tensor<40xbf16> loc(#loc) + %124 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_74/weights"} -> tensor<40x120x1x1xbf16> loc(#loc) + %125 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/biases"} -> tensor<120xbf16> loc(#loc) + %126 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_65/weights"} -> tensor<120x32x1x1xbf16> loc(#loc) + %127 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/biases"} -> tensor<32xbf16> loc(#loc) + %128 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_63/weights"} -> tensor<32x120x1x1xbf16> loc(#loc) + %129 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/biases"} -> tensor<120xbf16> loc(#loc) + %130 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_60/weights"} -> tensor<120x1x5x5xbf16> loc(#loc) + %131 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/biases"} -> tensor<120xbf16> loc(#loc) + %132 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_58/weights"} -> tensor<120x40x1x1xbf16> loc(#loc) + %133 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/biases"} -> tensor<40xbf16> loc(#loc) + %134 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_57/weights"} -> tensor<40x72x1x1xbf16> loc(#loc) + %135 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/biases"} -> tensor<72xbf16> loc(#loc) + %136 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_48/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %137 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/biases"} -> tensor<24xbf16> loc(#loc) + %138 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_46/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %139 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/biases"} -> tensor<72xbf16> loc(#loc) + %140 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_43/weights"} -> tensor<72x1x5x5xbf16> loc(#loc) + %141 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/biases"} -> tensor<72xbf16> loc(#loc) + %142 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_41/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %143 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/biases"} -> tensor<24xbf16> loc(#loc) + %144 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_39/weights"} -> tensor<24x72x1x1xbf16> loc(#loc) + %145 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/biases"} -> tensor<72xbf16> loc(#loc) + %146 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_37/weights"} -> tensor<72x1x3x3xbf16> loc(#loc) + %147 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/biases"} -> tensor<72xbf16> loc(#loc) + %148 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_35/weights"} -> tensor<72x24x1x1xbf16> loc(#loc) + %149 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/biases"} -> tensor<24xbf16> loc(#loc) + %150 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_34/weights"} -> tensor<24x64x1x1xbf16> loc(#loc) + %151 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/biases"} -> tensor<64xbf16> loc(#loc) + %152 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_32/weights"} -> tensor<64x1x3x3xbf16> loc(#loc) + %153 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/biases"} -> tensor<64xbf16> loc(#loc) + %154 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_30/weights"} -> tensor<64x16x1x1xbf16> loc(#loc) + %155 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/biases"} -> tensor<16xbf16> loc(#loc) + %156 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_28/weights"} -> tensor<16x16x1x1xbf16> loc(#loc) + %157 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/biases"} -> tensor<16xbf16> loc(#loc) + %158 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_26/weights"} -> tensor<16x1x3x3xbf16> loc(#loc) + %159 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/biases"} -> tensor<16xbf16> loc(#loc) + %160 = xten_nn.load_external_const {file = "constants.h5", key = "Conv_17/weights"} -> tensor<16x3x3x3xbf16> loc(#loc) + %161 = xten_nn.load_external_const {file = "constants.h5", key = "Div_16/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc6) + %162 = xten_nn.load_external_const {file = "constants.h5", key = "Sub_14/Constant_1_0"} -> tensor<1x3x180x320xbf16> loc(#loc320) + %163 = xten_nn.subgraph (%arg5 = %0: tensor<1x180x320x4xbf16>) attributes { + LayerName = "Slice_7_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 4]> : vector<4xindex> + } + ], + OutputName = "Slice_7", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "392", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "W", + config.dim_c = 184 : ui32, + config.dim_h = 320 : ui32, + config.dim_w = 4 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_7", + OutputName = "Slice_7", + size = array, + start = array} : (tensor<1x180x320x4xbf16>) -> tensor<1x180x320x3xbf16> loc(#loc9) + xten_nn.output %461 : tensor<1x180x320x3xbf16> loc(#loc9) + } -> tensor<1x180x320x3xbf16> loc(#loc9) + %164 = xten_nn.subgraph (%arg5 = %163: tensor<1x180x320x3xbf16>) attributes { + LayerName = "CompilerGenerated_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "CompilerGenerated_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + Specializes = "BufferPadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_0_padded = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_1_padded = 23 : ui32, + config.dim_2 = 3 : ui32, + config.dim_2_padded = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_padded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x180x320x3xbf16> loc(#loc10) + } -> tensor<1x180x320x3xbf16> loc(#loc10) + %165 = xten_nn.subgraph (%arg5 = %164: tensor<1x180x320x3xbf16>) attributes { + LayerName = "Slice_7_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "387", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 5]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 180, 320, 3]> : vector<4xindex> + } + ], + OutputName = "Add_445_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 320 : ui32, + config.dim_1 = 23 : ui32, + config.dim_2 = 8 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 10 : ui32 + }} { + %461 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc11) + %462 = tosa.transpose %arg5, %461 : (tensor<1x180x320x3xbf16>, tensor<4xi32>) -> tensor<1x3x180x320xbf16> loc(#loc322) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc322) + } -> tensor<1x3x180x320xbf16> loc(#loc321) + %166 = xten_nn.subgraph (%arg5 = %165: tensor<1x3x180x320xbf16>) attributes { + LayerName = "CompilerGenerated_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 4, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "CompilerGenerated_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "BufferUnpadAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 184 : ui32, + config.dim_0_unpadded = 180 : ui32, + config.dim_1 = 1 : ui32, + config.dim_1_unpadded = 1 : ui32, + config.dim_2 = 320 : ui32, + config.dim_2_unpadded = 320 : ui32, + config.dim_3 = 8 : ui32, + config.dim_3_unpadded = 8 : ui32, + config.dtype = "bfloat16" + }} { + xten_nn.output %arg5 : tensor<1x3x180x320xbf16> loc(#loc10) + } -> tensor<1x3x180x320xbf16> loc(#loc10) + %167 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>) attributes { + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_346", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "AveragePool_346", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc12) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc12) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_346", + PartOfOutputName = "AveragePool_346", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>) -> tensor<1x90x160x3xbf16> loc(#loc12) + %466 = tosa.transpose %465, %462 : (tensor<1x90x160x3xbf16>, tensor<4xi32>) -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %466 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + xten_nn.output %461 : tensor<1x3x90x160xbf16> loc(#loc12) + } -> tensor<1x3x90x160xbf16> loc(#loc12) + %168 = xten_nn.subgraph (%arg5 = %167: tensor<1x3x90x160xbf16>) attributes { + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x90x160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + HWPaddingNotCounted = [[0, 0], [0, 0]], + LayerName = "AveragePool_347", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "778", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "AveragePool_347", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc13) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x3xbf16> loc(#loc13) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_347", + PartOfOutputName = "AveragePool_347", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x90x160x3xbf16>) -> tensor<1x45x80x3xbf16> loc(#loc13) + %466 = tosa.transpose %465, %462 : (tensor<1x45x80x3xbf16>, tensor<4xi32>) -> tensor<1x3x45x80xbf16> loc(#loc13) + xten_nn.output %466 : tensor<1x3x45x80xbf16> loc(#loc13) + } -> tensor<1x3x45x80xbf16> loc(#loc13) + xten_nn.output %461 : tensor<1x3x45x80xbf16> loc(#loc13) + } -> tensor<1x3x45x80xbf16> loc(#loc13) + %169 = xten_nn.subgraph (%arg5 = %168: tensor<1x3x45x80xbf16>) attributes { + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "780", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x45x80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 1], [0, 0]], + HWPaddingNotCounted = [[0, 1], [0, 0]], + LayerName = "AveragePool_348", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "779", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "AveragePool_348", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "780", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AvgPool2dBf16", + With = { + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.dtype = "bfloat16", + config.ksize = 2 : ui8, + config.stride_log2 = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc14) + %464 = tosa.transpose %arg6, %463 : (tensor<1x3x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x3xbf16> loc(#loc14) + %465 = tosa.avg_pool2d %464 { + PartOfLayerName = "AveragePool_348", + PartOfOutputName = "AveragePool_348", + acc_type = f32, + kernel = array, + pad = array, + stride = array} : (tensor<1x45x80x3xbf16>) -> tensor<1x23x40x3xbf16> loc(#loc14) + %466 = tosa.transpose %465, %462 : (tensor<1x23x40x3xbf16>, tensor<4xi32>) -> tensor<1x3x23x40xbf16> loc(#loc14) + xten_nn.output %466 : tensor<1x3x23x40xbf16> loc(#loc14) + } -> tensor<1x3x23x40xbf16> loc(#loc14) + xten_nn.output %461 : tensor<1x3x23x40xbf16> loc(#loc14) + } -> tensor<1x3x23x40xbf16> loc(#loc14) + %170 = xten_nn.subgraph (%arg5 = %166: tensor<1x3x180x320xbf16>, %arg6 = %162: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "392", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Sub_14", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Sub_14", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "392", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Sub_14", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Sub_14", OutputName = "Initializer_398"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc320) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc320) + } -> tensor<1x3x180x320xbf16> loc(#loc320) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc320) + } -> tensor<1x3x180x320xbf16> loc(#loc320) + %171 = xten_nn.subgraph (%arg5 = %170: tensor<1x3x180x320xbf16>, %arg6 = %161: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Div_16", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "399", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "393", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Div_16", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + OutputName = "Div_16", + PartOfLayerName = "Div_16", + shift = 0 : i8} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc6) + } -> tensor<1x3x180x320xbf16> loc(#loc6) + %172 = xten_nn.subgraph (%arg5 = %171: tensor<1x3x180x320xbf16>, %arg6 = %160: tensor<16x3x3x3xbf16>, %arg7 = %159: tensor<16xbf16>) attributes { + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Name = "399", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Name = "929", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x3x180x320xbf16>, %arg9 = %arg6: tensor<16x3x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_17", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + Name = "399", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 3, 3, 3]> : vector<4xindex> + }, + { + Name = "929", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_17", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 2 : ui8, + config.stride_w = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x3x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x3xbf16> loc(#loc15) + %465 = tosa.transpose %arg8, %463 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc15) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_17", + PartOfOutputName = "Conv_17", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x3xbf16>, tensor<16x3x3x3xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc15) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc15) + } -> tensor<1x16x90x160xbf16> loc(#loc15) + %173 = xten_nn.subgraph (%arg5 = %172: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_19", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_19", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_19", OutputName = "Add_19"} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc16) + } -> tensor<1x16x90x160xbf16> loc(#loc16) + %174 = xten_nn.subgraph (%arg5 = %173: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Clip_22", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "405", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Clip_22", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_22", + OutputName = "Clip_22", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc17) + } -> tensor<1x16x90x160xbf16> loc(#loc17) + %175 = xten_nn.subgraph (%arg5 = %174: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "410", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Div_24", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "408", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Div_24", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "410", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_24", + OutputName = "Div_24", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc18) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc18) + } -> tensor<1x16x90x160xbf16> loc(#loc18) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc18) + } -> tensor<1x16x90x160xbf16> loc(#loc18) + %176 = xten_nn.subgraph (%arg5 = %172: tensor<1x16x90x160xbf16>, %arg6 = %175: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_25", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "401", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_25", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_25", + OutputName = "Mul_25", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc19) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc19) + } -> tensor<1x16x90x160xbf16> loc(#loc19) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc19) + } -> tensor<1x16x90x160xbf16> loc(#loc19) + %177 = xten_nn.subgraph (%arg5 = %176: tensor<1x16x90x160xbf16>, %arg6 = %158: tensor<16x1x3x3xbf16>, %arg7 = %157: tensor<16xbf16>) attributes { + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "411", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<16x1x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_26", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "411", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "928", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "411", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_27", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc323) + %465 = tosa.transpose %arg9, %464 : (tensor<16x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x16x1xbf16> loc(#loc323) + %466 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc323) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_26", + PartOfOutputName = "Conv_26", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<3x3x16x1xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc20) + %468 = tosa.clamp %467 { + LayerName = "Relu_27", + OutputName = "Relu_27", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc21) + %469 = tosa.transpose %468, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc323) + xten_nn.output %469 : tensor<1x16x90x160xbf16> loc(#loc21) + } -> tensor<1x16x90x160xbf16> loc(#loc323) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc323) + } -> tensor<1x16x90x160xbf16> loc(#loc323) + %178 = xten_nn.subgraph (%arg5 = %177: tensor<1x16x90x160xbf16>, %arg6 = %156: tensor<16x16x1x1xbf16>, %arg7 = %155: tensor<16xbf16>, %arg8 = %176: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Conv_28", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "931", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "935", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x16x90x160xbf16>, %arg10 = %arg6: tensor<16x16x1x1xbf16>, %arg11 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_28", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "931", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "935", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_28", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "934", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc22) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<16x16x1x1xbf16>) -> tensor<16x1x1x16xbf16> loc(#loc22) + %466 = tosa.transpose %arg9, %464 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc22) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_28", + PartOfOutputName = "Conv_28", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<16x1x1x16xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc22) + %468 = tosa.transpose %467, %463 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc22) + xten_nn.output %468 : tensor<1x16x90x160xbf16> loc(#loc22) + } -> tensor<1x16x90x160xbf16> loc(#loc22) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x16x90x160xbf16>, %arg10 = %arg8: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_29", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "934", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "414", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_29", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_29", OutputName = "Add_29"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc23) + xten_nn.output %463 : tensor<1x16x90x160xbf16> loc(#loc23) + } -> tensor<1x16x90x160xbf16> loc(#loc23) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc23) + } -> tensor<1x16x90x160xbf16> loc(#loc324) + %179 = xten_nn.subgraph (%arg5 = %178: tensor<1x16x90x160xbf16>, %arg6 = %154: tensor<64x16x1x1xbf16>, %arg7 = %153: tensor<64xbf16>) attributes { + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "934", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "417", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x90x160xbf16>, %arg9 = %arg6: tensor<64x16x1x1xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_30", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + Name = "934", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "417", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_31", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc325) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<64x16x1x1xbf16>) -> tensor<64x1x1x16xbf16> loc(#loc325) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc325) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_30", + PartOfOutputName = "Conv_30", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x16xbf16>, tensor<64x1x1x16xbf16>, tensor<64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc24) + %467 = tosa.clamp %466 { + LayerName = "Relu_31", + OutputName = "Relu_31", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x64xbf16>) -> tensor<1x90x160x64xbf16> loc(#loc25) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x64xbf16>, tensor<4xi32>) -> tensor<1x64x90x160xbf16> loc(#loc325) + xten_nn.output %468 : tensor<1x64x90x160xbf16> loc(#loc25) + } -> tensor<1x64x90x160xbf16> loc(#loc325) + xten_nn.output %461 : tensor<1x64x90x160xbf16> loc(#loc325) + } -> tensor<1x64x90x160xbf16> loc(#loc325) + %180 = xten_nn.subgraph (%arg5 = %179: tensor<1x64x90x160xbf16>, %arg6 = %152: tensor<64x1x3x3xbf16>, %arg7 = %151: tensor<64xbf16>) attributes { + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "937", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "941", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x90x160xbf16>, %arg9 = %arg6: tensor<64x1x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 0], [1, 0]], + LayerName = "Conv_32", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "420", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "937", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "941", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_33", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc326) + %465 = tosa.transpose %arg9, %464 : (tensor<64x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x64x1xbf16> loc(#loc326) + %466 = tosa.transpose %arg8, %463 : (tensor<1x64x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x64xbf16> loc(#loc326) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_32", + PartOfOutputName = "Conv_32", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x64xbf16>, tensor<3x3x64x1xbf16>, tensor<64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc26) + %468 = tosa.clamp %467 { + LayerName = "Relu_33", + OutputName = "Relu_33", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x64xbf16>) -> tensor<1x45x80x64xbf16> loc(#loc27) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x64xbf16>, tensor<4xi32>) -> tensor<1x64x45x80xbf16> loc(#loc326) + xten_nn.output %469 : tensor<1x64x45x80xbf16> loc(#loc27) + } -> tensor<1x64x45x80xbf16> loc(#loc326) + xten_nn.output %461 : tensor<1x64x45x80xbf16> loc(#loc326) + } -> tensor<1x64x45x80xbf16> loc(#loc326) + %181 = xten_nn.subgraph (%arg5 = %180: tensor<1x64x45x80xbf16>, %arg6 = %150: tensor<24x64x1x1xbf16>, %arg7 = %149: tensor<24xbf16>) attributes { + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Name = "940", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Name = "944", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x64x45x80xbf16>, %arg9 = %arg6: tensor<24x64x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_34", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "423", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 45, 80]> : vector<4xindex> + }, + { + Name = "940", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 64, 1, 1]> : vector<4xindex> + }, + { + Name = "944", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_34", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc28) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x64x1x1xbf16>) -> tensor<24x1x1x64xbf16> loc(#loc28) + %465 = tosa.transpose %arg8, %463 : (tensor<1x64x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x64xbf16> loc(#loc28) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_34", + PartOfOutputName = "Conv_34", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x64xbf16>, tensor<24x1x1x64xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc28) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc28) + xten_nn.output %467 : tensor<1x24x45x80xbf16> loc(#loc28) + } -> tensor<1x24x45x80xbf16> loc(#loc28) + xten_nn.output %461 : tensor<1x24x45x80xbf16> loc(#loc28) + } -> tensor<1x24x45x80xbf16> loc(#loc28) + %182 = xten_nn.subgraph (%arg5 = %181: tensor<1x24x45x80xbf16>, %arg6 = %148: tensor<72x24x1x1xbf16>, %arg7 = %147: tensor<72xbf16>) attributes { + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "423", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "947", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_35", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "943", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "423", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "947", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_36", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc327) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc327) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc327) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_35", + PartOfOutputName = "Conv_35", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc29) + %467 = tosa.clamp %466 { + LayerName = "Relu_36", + OutputName = "Relu_36", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc30) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc327) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc30) + } -> tensor<1x72x45x80xbf16> loc(#loc327) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc327) + } -> tensor<1x72x45x80xbf16> loc(#loc327) + %183 = xten_nn.subgraph (%arg5 = %182: tensor<1x72x45x80xbf16>, %arg6 = %146: tensor<72x1x3x3xbf16>, %arg7 = %145: tensor<72xbf16>) attributes { + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "946", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "950", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x3x3xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_37", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "428", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "946", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "950", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_38", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc328) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x72x1xbf16> loc(#loc328) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc328) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_37", + PartOfOutputName = "Conv_37", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<3x3x72x1xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc31) + %468 = tosa.clamp %467 { + LayerName = "Relu_38", + OutputName = "Relu_38", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc32) + %469 = tosa.transpose %468, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc328) + xten_nn.output %469 : tensor<1x72x45x80xbf16> loc(#loc32) + } -> tensor<1x72x45x80xbf16> loc(#loc328) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc328) + } -> tensor<1x72x45x80xbf16> loc(#loc328) + %184 = xten_nn.subgraph (%arg5 = %183: tensor<1x72x45x80xbf16>, %arg6 = %144: tensor<24x72x1x1xbf16>, %arg7 = %143: tensor<24xbf16>, %arg8 = %181: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Conv_39", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Name = "949", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "953", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x72x45x80xbf16>, %arg10 = %arg6: tensor<24x72x1x1xbf16>, %arg11 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_39", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + Name = "949", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "953", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_39", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "952", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc33) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc33) + %466 = tosa.transpose %arg9, %464 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc33) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_39", + PartOfOutputName = "Conv_39", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x45x80x24xbf16> loc(#loc33) + %468 = tosa.transpose %467, %463 : (tensor<1x45x80x24xbf16>, tensor<4xi32>) -> tensor<1x24x45x80xbf16> loc(#loc33) + xten_nn.output %468 : tensor<1x24x45x80xbf16> loc(#loc33) + } -> tensor<1x24x45x80xbf16> loc(#loc33) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x24x45x80xbf16>, %arg10 = %arg8: tensor<1x24x45x80xbf16>) attributes { + LayerName = "Add_40", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "952", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "431", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_40", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_40", OutputName = "Add_40"} : (tensor<1x24x45x80xbf16>, tensor<1x24x45x80xbf16>) -> tensor<1x24x45x80xbf16> loc(#loc34) + xten_nn.output %463 : tensor<1x24x45x80xbf16> loc(#loc34) + } -> tensor<1x24x45x80xbf16> loc(#loc34) + xten_nn.output %462 : tensor<1x24x45x80xbf16> loc(#loc34) + } -> tensor<1x24x45x80xbf16> loc(#loc329) + %185 = xten_nn.subgraph (%arg5 = %184: tensor<1x24x45x80xbf16>, %arg6 = %142: tensor<72x24x1x1xbf16>, %arg7 = %141: tensor<72xbf16>) attributes { + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "952", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "434", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x45x80xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_41", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + Name = "952", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "434", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_42", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc330) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc330) + %465 = tosa.transpose %arg8, %463 : (tensor<1x24x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x24xbf16> loc(#loc330) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_41", + PartOfOutputName = "Conv_41", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc35) + %467 = tosa.clamp %466 { + LayerName = "Relu_42", + OutputName = "Relu_42", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x72xbf16>) -> tensor<1x45x80x72xbf16> loc(#loc36) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x72xbf16>, tensor<4xi32>) -> tensor<1x72x45x80xbf16> loc(#loc330) + xten_nn.output %468 : tensor<1x72x45x80xbf16> loc(#loc36) + } -> tensor<1x72x45x80xbf16> loc(#loc330) + xten_nn.output %461 : tensor<1x72x45x80xbf16> loc(#loc330) + } -> tensor<1x72x45x80xbf16> loc(#loc330) + %186 = xten_nn.subgraph (%arg5 = %185: tensor<1x72x45x80xbf16>, %arg6 = %140: tensor<72x1x5x5xbf16>, %arg7 = %139: tensor<72xbf16>) attributes { + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "955", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "959", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x45x80xbf16>, %arg9 = %arg6: tensor<72x1x5x5xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 1]], + LayerName = "Conv_43", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "437", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "955", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "959", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_44", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc331) + %465 = tosa.transpose %arg9, %464 : (tensor<72x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x72x1xbf16> loc(#loc331) + %466 = tosa.transpose %arg8, %463 : (tensor<1x72x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x72xbf16> loc(#loc331) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_43", + PartOfOutputName = "Conv_43", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x72xbf16>, tensor<5x5x72x1xbf16>, tensor<72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc37) + %468 = tosa.clamp %467 { + LayerName = "Relu_44", + OutputName = "Relu_44", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x72xbf16>) -> tensor<1x23x40x72xbf16> loc(#loc38) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x72xbf16>, tensor<4xi32>) -> tensor<1x72x23x40xbf16> loc(#loc331) + xten_nn.output %469 : tensor<1x72x23x40xbf16> loc(#loc38) + } -> tensor<1x72x23x40xbf16> loc(#loc331) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc331) + } -> tensor<1x72x23x40xbf16> loc(#loc331) + %187 = xten_nn.subgraph (%arg5 = %186: tensor<1x72x23x40xbf16>) attributes { + LayerName = "GlobalAveragePool_45_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_45_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 9 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x72x23x40xbf16>) -> tensor<1x72x1x920xbf16> loc(#loc39) + xten_nn.output %461 : tensor<1x72x1x920xbf16> loc(#loc39) + } -> tensor<1x72x1x920xbf16> loc(#loc39) + %188 = xten_nn.subgraph (%arg5 = %187: tensor<1x72x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_45_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_45_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_45_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "440", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_45_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 72 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x72x1x920xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc39) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc39) + } -> tensor<1x72x1x1xbf16> loc(#loc39) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc39) + } -> tensor<1x72x1x1xbf16> loc(#loc39) + %189 = xten_nn.subgraph (%arg5 = %188: tensor<1x72x1x1xbf16>, %arg6 = %138: tensor<24x72x1x1xbf16>, %arg7 = %137: tensor<24xbf16>) attributes { + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "440", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x1x1xbf16>, %arg9 = %arg6: tensor<24x72x1x1xbf16>, %arg10 = %arg7: tensor<24xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_46", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "441", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "440", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[24, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_47", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<24x72x1x1xbf16>) -> tensor<24x1x1x72xbf16> loc(#loc332) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc332) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_46", + PartOfOutputName = "Conv_46", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x72xbf16>, tensor<24x1x1x72xbf16>, tensor<24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc40) + %465 = tosa.clamp %464 { + LayerName = "Relu_47", + OutputName = "Relu_47", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x24xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc41) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x24xbf16>) -> tensor<1x24x1x1xbf16> loc(#loc332) + xten_nn.output %466 : tensor<1x24x1x1xbf16> loc(#loc41) + } -> tensor<1x24x1x1xbf16> loc(#loc332) + xten_nn.output %461 : tensor<1x24x1x1xbf16> loc(#loc332) + } -> tensor<1x24x1x1xbf16> loc(#loc332) + %190 = xten_nn.subgraph (%arg5 = %189: tensor<1x24x1x1xbf16>, %arg6 = %136: tensor<72x24x1x1xbf16>, %arg7 = %135: tensor<72xbf16>) attributes { + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "442", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x24x1x1xbf16>, %arg9 = %arg6: tensor<72x24x1x1xbf16>, %arg10 = %arg7: tensor<72xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_48", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "443", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "442", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[72, 24, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.4.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_48", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<72x24x1x1xbf16>) -> tensor<72x1x1x24xbf16> loc(#loc42) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x24x1x1xbf16>) -> tensor<1x1x1x24xbf16> loc(#loc42) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_48", + PartOfOutputName = "Conv_48", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x24xbf16>, tensor<72x1x1x24xbf16>, tensor<72xbf16>) -> tensor<1x1x1x72xbf16> loc(#loc42) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x72xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %465 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc42) + } -> tensor<1x72x1x1xbf16> loc(#loc42) + %191 = xten_nn.subgraph (%arg5 = %190: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Add_50", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "444", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_50", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_50", OutputName = "Add_50"} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc43) + } -> tensor<1x72x1x1xbf16> loc(#loc43) + %192 = xten_nn.subgraph (%arg5 = %191: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Clip_53", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "446", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_53", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_53", + OutputName = "Clip_53", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc44) + xten_nn.output %462 : tensor<1x72x1x1xbf16> loc(#loc44) + } -> tensor<1x72x1x1xbf16> loc(#loc44) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc44) + } -> tensor<1x72x1x1xbf16> loc(#loc44) + %193 = xten_nn.subgraph (%arg5 = %192: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Div_55", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_55", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_55", + OutputName = "Div_55", + shift = 0 : i8} : (tensor<1x72x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x72x1x1xbf16> loc(#loc45) + xten_nn.output %463 : tensor<1x72x1x1xbf16> loc(#loc45) + } -> tensor<1x72x1x1xbf16> loc(#loc45) + xten_nn.output %461 : tensor<1x72x1x1xbf16> loc(#loc45) + } -> tensor<1x72x1x1xbf16> loc(#loc45) + %194 = xten_nn.subgraph (%arg5 = %193: tensor<1x72x1x1xbf16>) attributes { + LayerName = "Mul_56_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_56_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 72 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x72x1x1xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc46) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc46) + } -> tensor<1x72x23x40xbf16> loc(#loc46) + %195 = xten_nn.subgraph (%arg5 = %194: tensor<1x72x23x40xbf16>, %arg6 = %186: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x72x23x40xbf16>, %arg8 = %arg6: tensor<1x72x23x40xbf16>) attributes { + LayerName = "Mul_56_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "451", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "449", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_56_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_56", + OutputName = "Mul_56", + shift = 0 : i8} : (tensor<1x72x23x40xbf16>, tensor<1x72x23x40xbf16>) -> tensor<1x72x23x40xbf16> loc(#loc46) + xten_nn.output %462 : tensor<1x72x23x40xbf16> loc(#loc46) + } -> tensor<1x72x23x40xbf16> loc(#loc46) + xten_nn.output %461 : tensor<1x72x23x40xbf16> loc(#loc46) + } -> tensor<1x72x23x40xbf16> loc(#loc46) + %196 = xten_nn.subgraph (%arg5 = %195: tensor<1x72x23x40xbf16>, %arg6 = %134: tensor<40x72x1x1xbf16>, %arg7 = %133: tensor<40xbf16>) attributes { + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Name = "451", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "452", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x72x23x40xbf16>, %arg9 = %arg6: tensor<40x72x1x1xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_57", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "452", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 72, 23, 40]> : vector<4xindex> + }, + { + Name = "451", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 72, 1, 1]> : vector<4xindex> + }, + { + Name = "452", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_57", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc47) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<40x72x1x1xbf16>) -> tensor<40x1x1x72xbf16> loc(#loc47) + %465 = tosa.transpose %arg8, %463 : (tensor<1x72x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x72xbf16> loc(#loc47) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_57", + PartOfOutputName = "Conv_57", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x72xbf16>, tensor<40x1x1x72xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc47) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc47) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc47) + } -> tensor<1x40x23x40xbf16> loc(#loc47) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc47) + } -> tensor<1x40x23x40xbf16> loc(#loc47) + %197 = xten_nn.subgraph (%arg5 = %196: tensor<1x40x23x40xbf16>, %arg6 = %132: tensor<120x40x1x1xbf16>, %arg7 = %131: tensor<120xbf16>) attributes { + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "452", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "965", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_58", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "961", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "452", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "965", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_59", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc333) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc333) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc333) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_58", + PartOfOutputName = "Conv_58", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc48) + %467 = tosa.clamp %466 { + LayerName = "Relu_59", + OutputName = "Relu_59", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc49) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc333) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc49) + } -> tensor<1x120x23x40xbf16> loc(#loc333) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc333) + } -> tensor<1x120x23x40xbf16> loc(#loc333) + %198 = xten_nn.subgraph (%arg5 = %197: tensor<1x120x23x40xbf16>, %arg6 = %130: tensor<120x1x5x5xbf16>, %arg7 = %129: tensor<120xbf16>) attributes { + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "964", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "968", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_60", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "457", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "964", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "968", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_61", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc334) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc334) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc334) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_60", + PartOfOutputName = "Conv_60", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc50) + %468 = tosa.clamp %467 { + LayerName = "Relu_61", + OutputName = "Relu_61", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc51) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc334) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc51) + } -> tensor<1x120x23x40xbf16> loc(#loc334) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc334) + } -> tensor<1x120x23x40xbf16> loc(#loc334) + %199 = xten_nn.subgraph (%arg5 = %198: tensor<1x120x23x40xbf16>) attributes { + LayerName = "GlobalAveragePool_62_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_62_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc52) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc52) + } -> tensor<1x120x1x920xbf16> loc(#loc52) + %200 = xten_nn.subgraph (%arg5 = %199: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_62_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_62_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_62_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "460", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_62_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc52) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc52) + } -> tensor<1x120x1x1xbf16> loc(#loc52) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc52) + } -> tensor<1x120x1x1xbf16> loc(#loc52) + %201 = xten_nn.subgraph (%arg5 = %200: tensor<1x120x1x1xbf16>, %arg6 = %128: tensor<32x120x1x1xbf16>, %arg7 = %127: tensor<32xbf16>) attributes { + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "460", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_63", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "461", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "460", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_64", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc335) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc335) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_63", + PartOfOutputName = "Conv_63", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc53) + %465 = tosa.clamp %464 { + LayerName = "Relu_64", + OutputName = "Relu_64", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc54) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc335) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc54) + } -> tensor<1x32x1x1xbf16> loc(#loc335) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc335) + } -> tensor<1x32x1x1xbf16> loc(#loc335) + %202 = xten_nn.subgraph (%arg5 = %201: tensor<1x32x1x1xbf16>, %arg6 = %126: tensor<120x32x1x1xbf16>, %arg7 = %125: tensor<120xbf16>) attributes { + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "462", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_65", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "463", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "462", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.5.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_65", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc55) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc55) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_65", + PartOfOutputName = "Conv_65", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc55) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc55) + } -> tensor<1x120x1x1xbf16> loc(#loc55) + %203 = xten_nn.subgraph (%arg5 = %202: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_67", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "464", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_67", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_67", OutputName = "Add_67"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc56) + } -> tensor<1x120x1x1xbf16> loc(#loc56) + %204 = xten_nn.subgraph (%arg5 = %203: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_70", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "466", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_70", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_70", + OutputName = "Clip_70", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc57) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc57) + } -> tensor<1x120x1x1xbf16> loc(#loc57) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc57) + } -> tensor<1x120x1x1xbf16> loc(#loc57) + %205 = xten_nn.subgraph (%arg5 = %204: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_72", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_72", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_72", + OutputName = "Div_72", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc58) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc58) + } -> tensor<1x120x1x1xbf16> loc(#loc58) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc58) + } -> tensor<1x120x1x1xbf16> loc(#loc58) + %206 = xten_nn.subgraph (%arg5 = %205: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Mul_73_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_73_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc59) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc59) + } -> tensor<1x120x23x40xbf16> loc(#loc59) + %207 = xten_nn.subgraph (%arg5 = %206: tensor<1x120x23x40xbf16>, %arg6 = %198: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_73_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "471", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "469", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_73_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_73", + OutputName = "Mul_73", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc59) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc59) + } -> tensor<1x120x23x40xbf16> loc(#loc59) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc59) + } -> tensor<1x120x23x40xbf16> loc(#loc59) + %208 = xten_nn.subgraph (%arg5 = %207: tensor<1x120x23x40xbf16>, %arg6 = %124: tensor<40x120x1x1xbf16>, %arg7 = %123: tensor<40xbf16>, %arg8 = %196: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_74", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "471", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "472", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_74", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "471", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "472", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_74", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "970", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc60) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc60) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc60) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_74", + PartOfOutputName = "Conv_74", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc60) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc60) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc60) + } -> tensor<1x40x23x40xbf16> loc(#loc60) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_75", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "970", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "472", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_75", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_75", OutputName = "Add_75"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc61) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc61) + } -> tensor<1x40x23x40xbf16> loc(#loc61) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc61) + } -> tensor<1x40x23x40xbf16> loc(#loc336) + %209 = xten_nn.subgraph (%arg5 = %208: tensor<1x40x23x40xbf16>, %arg6 = %122: tensor<120x40x1x1xbf16>, %arg7 = %121: tensor<120xbf16>) attributes { + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "970", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "475", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<120x40x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_76", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "475", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "970", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "475", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_77", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc337) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x40x1x1xbf16>) -> tensor<120x1x1x40xbf16> loc(#loc337) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc337) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_76", + PartOfOutputName = "Conv_76", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<120x1x1x40xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc62) + %467 = tosa.clamp %466 { + LayerName = "Relu_77", + OutputName = "Relu_77", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc63) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc337) + xten_nn.output %468 : tensor<1x120x23x40xbf16> loc(#loc63) + } -> tensor<1x120x23x40xbf16> loc(#loc337) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc337) + } -> tensor<1x120x23x40xbf16> loc(#loc337) + %210 = xten_nn.subgraph (%arg5 = %209: tensor<1x120x23x40xbf16>, %arg6 = %120: tensor<120x1x5x5xbf16>, %arg7 = %119: tensor<120xbf16>) attributes { + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "973", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "977", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x23x40xbf16>, %arg9 = %arg6: tensor<120x1x5x5xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[2, 2], [2, 2]], + LayerName = "Conv_78", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "478", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "973", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 1, 5, 5]> : vector<4xindex> + }, + { + Name = "977", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_79", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + Traits = { + NonNegativeOut = true + }, + With = { + config.act = 1 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 5 : ui8, + config.kernel_width = 5 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc338) + %465 = tosa.transpose %arg9, %464 : (tensor<120x1x5x5xbf16>, tensor<4xi32>) -> tensor<5x5x120x1xbf16> loc(#loc338) + %466 = tosa.transpose %arg8, %463 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc338) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_78", + PartOfOutputName = "Conv_78", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<5x5x120x1xbf16>, tensor<120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc64) + %468 = tosa.clamp %467 { + LayerName = "Relu_79", + OutputName = "Relu_79", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x120xbf16>) -> tensor<1x23x40x120xbf16> loc(#loc65) + %469 = tosa.transpose %468, %462 : (tensor<1x23x40x120xbf16>, tensor<4xi32>) -> tensor<1x120x23x40xbf16> loc(#loc338) + xten_nn.output %469 : tensor<1x120x23x40xbf16> loc(#loc65) + } -> tensor<1x120x23x40xbf16> loc(#loc338) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc338) + } -> tensor<1x120x23x40xbf16> loc(#loc338) + %211 = xten_nn.subgraph (%arg5 = %210: tensor<1x120x23x40xbf16>) attributes { + LayerName = "GlobalAveragePool_80_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_80_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 23 : ui32, + config.dim_1 = 15 : ui32, + config.dim_2 = 40 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x120x23x40xbf16>) -> tensor<1x120x1x920xbf16> loc(#loc66) + xten_nn.output %461 : tensor<1x120x1x920xbf16> loc(#loc66) + } -> tensor<1x120x1x920xbf16> loc(#loc66) + %212 = xten_nn.subgraph (%arg5 = %211: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_80_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_80_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x920xbf16>) attributes { + LayerName = "GlobalAveragePool_80_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "481", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 920]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_80_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 120 : ui32, + config.full_height = 1 : ui32, + config.full_width = 920 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x120x1x920xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc66) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc66) + } -> tensor<1x120x1x1xbf16> loc(#loc66) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc66) + } -> tensor<1x120x1x1xbf16> loc(#loc66) + %213 = xten_nn.subgraph (%arg5 = %212: tensor<1x120x1x1xbf16>, %arg6 = %118: tensor<32x120x1x1xbf16>, %arg7 = %117: tensor<32xbf16>) attributes { + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "481", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<32x120x1x1xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_81", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "482", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "481", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_82", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<32x120x1x1xbf16>) -> tensor<32x1x1x120xbf16> loc(#loc339) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc339) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_81", + PartOfOutputName = "Conv_81", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<32x1x1x120xbf16>, tensor<32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc67) + %465 = tosa.clamp %464 { + LayerName = "Relu_82", + OutputName = "Relu_82", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x32xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc68) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x32xbf16>) -> tensor<1x32x1x1xbf16> loc(#loc339) + xten_nn.output %466 : tensor<1x32x1x1xbf16> loc(#loc68) + } -> tensor<1x32x1x1xbf16> loc(#loc339) + xten_nn.output %461 : tensor<1x32x1x1xbf16> loc(#loc339) + } -> tensor<1x32x1x1xbf16> loc(#loc339) + %214 = xten_nn.subgraph (%arg5 = %213: tensor<1x32x1x1xbf16>, %arg6 = %116: tensor<120x32x1x1xbf16>, %arg7 = %115: tensor<120xbf16>) attributes { + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "483", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x1x1xbf16>, %arg9 = %arg6: tensor<120x32x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_83", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "484", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "483", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 32, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.6.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_83", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x32x1x1xbf16>) -> tensor<120x1x1x32xbf16> loc(#loc69) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x32x1x1xbf16>) -> tensor<1x1x1x32xbf16> loc(#loc69) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_83", + PartOfOutputName = "Conv_83", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x32xbf16>, tensor<120x1x1x32xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc69) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %465 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc69) + } -> tensor<1x120x1x1xbf16> loc(#loc69) + %215 = xten_nn.subgraph (%arg5 = %214: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Add_85", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "485", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_85", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_85", OutputName = "Add_85"} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc70) + } -> tensor<1x120x1x1xbf16> loc(#loc70) + %216 = xten_nn.subgraph (%arg5 = %215: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Clip_88", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "487", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_88", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_88", + OutputName = "Clip_88", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc71) + xten_nn.output %462 : tensor<1x120x1x1xbf16> loc(#loc71) + } -> tensor<1x120x1x1xbf16> loc(#loc71) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc71) + } -> tensor<1x120x1x1xbf16> loc(#loc71) + %217 = xten_nn.subgraph (%arg5 = %216: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Div_90", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_90", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_90", + OutputName = "Div_90", + shift = 0 : i8} : (tensor<1x120x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc72) + xten_nn.output %463 : tensor<1x120x1x1xbf16> loc(#loc72) + } -> tensor<1x120x1x1xbf16> loc(#loc72) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc72) + } -> tensor<1x120x1x1xbf16> loc(#loc72) + %218 = xten_nn.subgraph (%arg5 = %217: tensor<1x120x1x1xbf16>) attributes { + LayerName = "Mul_91_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_91_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 120 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 23 : ui32, + config.rep_dim_w = 40 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc73) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc73) + } -> tensor<1x120x23x40xbf16> loc(#loc73) + %219 = xten_nn.subgraph (%arg5 = %218: tensor<1x120x23x40xbf16>, %arg6 = %210: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x120x23x40xbf16>, %arg8 = %arg6: tensor<1x120x23x40xbf16>) attributes { + LayerName = "Mul_91_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "492", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "490", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_91_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_91", + OutputName = "Mul_91", + shift = 0 : i8} : (tensor<1x120x23x40xbf16>, tensor<1x120x23x40xbf16>) -> tensor<1x120x23x40xbf16> loc(#loc73) + xten_nn.output %462 : tensor<1x120x23x40xbf16> loc(#loc73) + } -> tensor<1x120x23x40xbf16> loc(#loc73) + xten_nn.output %461 : tensor<1x120x23x40xbf16> loc(#loc73) + } -> tensor<1x120x23x40xbf16> loc(#loc73) + %220 = xten_nn.subgraph (%arg5 = %219: tensor<1x120x23x40xbf16>, %arg6 = %114: tensor<40x120x1x1xbf16>, %arg7 = %113: tensor<40xbf16>, %arg8 = %208: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Conv_92", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "492", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "493", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x120x23x40xbf16>, %arg10 = %arg6: tensor<40x120x1x1xbf16>, %arg11 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_92", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 23, 40]> : vector<4xindex> + }, + { + Name = "492", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "493", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_92", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "979", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc74) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<40x120x1x1xbf16>) -> tensor<40x1x1x120xbf16> loc(#loc74) + %466 = tosa.transpose %arg9, %464 : (tensor<1x120x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x120xbf16> loc(#loc74) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_92", + PartOfOutputName = "Conv_92", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x120xbf16>, tensor<40x1x1x120xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc74) + %468 = tosa.transpose %467, %463 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc74) + xten_nn.output %468 : tensor<1x40x23x40xbf16> loc(#loc74) + } -> tensor<1x40x23x40xbf16> loc(#loc74) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x40x23x40xbf16>, %arg10 = %arg8: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_93", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "979", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "493", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_93", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_93", OutputName = "Add_93"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc75) + xten_nn.output %463 : tensor<1x40x23x40xbf16> loc(#loc75) + } -> tensor<1x40x23x40xbf16> loc(#loc75) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc75) + } -> tensor<1x40x23x40xbf16> loc(#loc340) + %221 = xten_nn.subgraph (%arg5 = %220: tensor<1x40x23x40xbf16>, %arg6 = %112: tensor<240x40x1x1xbf16>, %arg7 = %111: tensor<240xbf16>) attributes { + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "979", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "496", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x23x40xbf16>, %arg9 = %arg6: tensor<240x40x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_94", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + Name = "979", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 40, 1, 1]> : vector<4xindex> + }, + { + Name = "496", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_94", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc76) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x40x1x1xbf16>) -> tensor<240x1x1x40xbf16> loc(#loc76) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc76) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_94", + PartOfOutputName = "Conv_94", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x40xbf16>, tensor<240x1x1x40xbf16>, tensor<240xbf16>) -> tensor<1x23x40x240xbf16> loc(#loc76) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x240xbf16>, tensor<4xi32>) -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %467 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc76) + } -> tensor<1x240x23x40xbf16> loc(#loc76) + %222 = xten_nn.subgraph (%arg5 = %221: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Add_96", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_96", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_96", OutputName = "Add_96"} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc77) + } -> tensor<1x240x23x40xbf16> loc(#loc77) + %223 = xten_nn.subgraph (%arg5 = %222: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Clip_99", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "500", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Clip_99", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_99", + OutputName = "Clip_99", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc78) + } -> tensor<1x240x23x40xbf16> loc(#loc78) + %224 = xten_nn.subgraph (%arg5 = %223: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "505", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Div_101", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "503", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Div_101", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "505", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_101", + OutputName = "Div_101", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc79) + xten_nn.output %463 : tensor<1x240x23x40xbf16> loc(#loc79) + } -> tensor<1x240x23x40xbf16> loc(#loc79) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc79) + } -> tensor<1x240x23x40xbf16> loc(#loc79) + %225 = xten_nn.subgraph (%arg5 = %221: tensor<1x240x23x40xbf16>, %arg6 = %224: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x23x40xbf16>, %arg8 = %arg6: tensor<1x240x23x40xbf16>) attributes { + LayerName = "Mul_102", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_102", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_102", + OutputName = "Mul_102", + shift = 0 : i8} : (tensor<1x240x23x40xbf16>, tensor<1x240x23x40xbf16>) -> tensor<1x240x23x40xbf16> loc(#loc80) + xten_nn.output %462 : tensor<1x240x23x40xbf16> loc(#loc80) + } -> tensor<1x240x23x40xbf16> loc(#loc80) + xten_nn.output %461 : tensor<1x240x23x40xbf16> loc(#loc80) + } -> tensor<1x240x23x40xbf16> loc(#loc80) + %226 = xten_nn.subgraph (%arg5 = %225: tensor<1x240x23x40xbf16>, %arg6 = %110: tensor<240x1x3x3xbf16>, %arg7 = %109: tensor<240xbf16>) attributes { + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "506", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x23x40xbf16>, %arg9 = %arg6: tensor<240x1x3x3xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 0]], + LayerName = "Conv_103", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "982", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "506", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_103", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 2 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc81) + %465 = tosa.transpose %arg9, %464 : (tensor<240x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x240x1xbf16> loc(#loc81) + %466 = tosa.transpose %arg8, %463 : (tensor<1x240x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x240xbf16> loc(#loc81) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_103", + PartOfOutputName = "Conv_103", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x240xbf16>, tensor<3x3x240x1xbf16>, tensor<240xbf16>) -> tensor<1x12x20x240xbf16> loc(#loc81) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x240xbf16>, tensor<4xi32>) -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %468 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc81) + } -> tensor<1x240x12x20xbf16> loc(#loc81) + %227 = xten_nn.subgraph (%arg5 = %226: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Add_105", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_105", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_105", OutputName = "Add_105"} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc82) + } -> tensor<1x240x12x20xbf16> loc(#loc82) + %228 = xten_nn.subgraph (%arg5 = %227: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Clip_108", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "510", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_108", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_108", + OutputName = "Clip_108", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc83) + } -> tensor<1x240x12x20xbf16> loc(#loc83) + %229 = xten_nn.subgraph (%arg5 = %228: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "515", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Div_110", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "513", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_110", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "515", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_110", + OutputName = "Div_110", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc84) + xten_nn.output %463 : tensor<1x240x12x20xbf16> loc(#loc84) + } -> tensor<1x240x12x20xbf16> loc(#loc84) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc84) + } -> tensor<1x240x12x20xbf16> loc(#loc84) + %230 = xten_nn.subgraph (%arg5 = %226: tensor<1x240x12x20xbf16>, %arg6 = %229: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x240x12x20xbf16>, %arg8 = %arg6: tensor<1x240x12x20xbf16>) attributes { + LayerName = "Mul_111", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "985", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "506", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_111", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_111", + OutputName = "Mul_111", + shift = 0 : i8} : (tensor<1x240x12x20xbf16>, tensor<1x240x12x20xbf16>) -> tensor<1x240x12x20xbf16> loc(#loc85) + xten_nn.output %462 : tensor<1x240x12x20xbf16> loc(#loc85) + } -> tensor<1x240x12x20xbf16> loc(#loc85) + xten_nn.output %461 : tensor<1x240x12x20xbf16> loc(#loc85) + } -> tensor<1x240x12x20xbf16> loc(#loc85) + %231 = xten_nn.subgraph (%arg5 = %230: tensor<1x240x12x20xbf16>, %arg6 = %108: tensor<80x240x1x1xbf16>, %arg7 = %107: tensor<80xbf16>) attributes { + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Name = "985", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "516", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x12x20xbf16>, %arg9 = %arg6: tensor<80x240x1x1xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_112", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "516", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 12, 20]> : vector<4xindex> + }, + { + Name = "985", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "516", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_112", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc86) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<80x240x1x1xbf16>) -> tensor<80x1x1x240xbf16> loc(#loc86) + %465 = tosa.transpose %arg8, %463 : (tensor<1x240x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x240xbf16> loc(#loc86) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_112", + PartOfOutputName = "Conv_112", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x240xbf16>, tensor<80x1x1x240xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc86) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc86) + xten_nn.output %467 : tensor<1x80x12x20xbf16> loc(#loc86) + } -> tensor<1x80x12x20xbf16> loc(#loc86) + xten_nn.output %461 : tensor<1x80x12x20xbf16> loc(#loc86) + } -> tensor<1x80x12x20xbf16> loc(#loc86) + %232 = xten_nn.subgraph (%arg5 = %231: tensor<1x80x12x20xbf16>, %arg6 = %106: tensor<200x80x1x1xbf16>, %arg7 = %105: tensor<200xbf16>) attributes { + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "516", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "992", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<200x80x1x1xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_113", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "516", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "992", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_113", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc87) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<200x80x1x1xbf16>) -> tensor<200x1x1x80xbf16> loc(#loc87) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc87) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_113", + PartOfOutputName = "Conv_113", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<200x1x1x80xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc87) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %467 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc87) + } -> tensor<1x200x12x20xbf16> loc(#loc87) + %233 = xten_nn.subgraph (%arg5 = %232: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_115", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_115", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_115", OutputName = "Add_115"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc88) + } -> tensor<1x200x12x20xbf16> loc(#loc88) + %234 = xten_nn.subgraph (%arg5 = %233: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_118", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "522", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_118", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_118", + OutputName = "Clip_118", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc89) + } -> tensor<1x200x12x20xbf16> loc(#loc89) + %235 = xten_nn.subgraph (%arg5 = %234: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "527", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_120", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "525", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_120", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "527", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_120", + OutputName = "Div_120", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc90) + } -> tensor<1x200x12x20xbf16> loc(#loc90) + %236 = xten_nn.subgraph (%arg5 = %232: tensor<1x200x12x20xbf16>, %arg6 = %235: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_121", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "988", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_121", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_121", + OutputName = "Mul_121", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc91) + } -> tensor<1x200x12x20xbf16> loc(#loc91) + %237 = xten_nn.subgraph (%arg5 = %236: tensor<1x200x12x20xbf16>, %arg6 = %104: tensor<200x1x3x3xbf16>, %arg7 = %103: tensor<200xbf16>) attributes { + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "528", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x200x12x20xbf16>, %arg9 = %arg6: tensor<200x1x3x3xbf16>, %arg10 = %arg7: tensor<200xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_122", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "991", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[200, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "528", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_122", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc92) + %465 = tosa.transpose %arg9, %464 : (tensor<200x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x200x1xbf16> loc(#loc92) + %466 = tosa.transpose %arg8, %463 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc92) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_122", + PartOfOutputName = "Conv_122", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<3x3x200x1xbf16>, tensor<200xbf16>) -> tensor<1x12x20x200xbf16> loc(#loc92) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x200xbf16>, tensor<4xi32>) -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %468 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc92) + } -> tensor<1x200x12x20xbf16> loc(#loc92) + %238 = xten_nn.subgraph (%arg5 = %237: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Add_124", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_124", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_124", OutputName = "Add_124"} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc93) + } -> tensor<1x200x12x20xbf16> loc(#loc93) + %239 = xten_nn.subgraph (%arg5 = %238: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Clip_127", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "532", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_127", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_127", + OutputName = "Clip_127", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc94) + } -> tensor<1x200x12x20xbf16> loc(#loc94) + %240 = xten_nn.subgraph (%arg5 = %239: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "537", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Div_129", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "535", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_129", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "537", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_129", + OutputName = "Div_129", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc95) + xten_nn.output %463 : tensor<1x200x12x20xbf16> loc(#loc95) + } -> tensor<1x200x12x20xbf16> loc(#loc95) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc95) + } -> tensor<1x200x12x20xbf16> loc(#loc95) + %241 = xten_nn.subgraph (%arg5 = %237: tensor<1x200x12x20xbf16>, %arg6 = %240: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x200x12x20xbf16>, %arg8 = %arg6: tensor<1x200x12x20xbf16>) attributes { + LayerName = "Mul_130", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "994", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "528", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_130", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_130", + OutputName = "Mul_130", + shift = 0 : i8} : (tensor<1x200x12x20xbf16>, tensor<1x200x12x20xbf16>) -> tensor<1x200x12x20xbf16> loc(#loc96) + xten_nn.output %462 : tensor<1x200x12x20xbf16> loc(#loc96) + } -> tensor<1x200x12x20xbf16> loc(#loc96) + xten_nn.output %461 : tensor<1x200x12x20xbf16> loc(#loc96) + } -> tensor<1x200x12x20xbf16> loc(#loc96) + %242 = xten_nn.subgraph (%arg5 = %241: tensor<1x200x12x20xbf16>, %arg6 = %102: tensor<80x200x1x1xbf16>, %arg7 = %101: tensor<80xbf16>, %arg8 = %231: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_131", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Name = "994", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Name = "538", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x200x12x20xbf16>, %arg10 = %arg6: tensor<80x200x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_131", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 200, 12, 20]> : vector<4xindex> + }, + { + Name = "994", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 200, 1, 1]> : vector<4xindex> + }, + { + Name = "538", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_131", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "997", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc97) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x200x1x1xbf16>) -> tensor<80x1x1x200xbf16> loc(#loc97) + %466 = tosa.transpose %arg9, %464 : (tensor<1x200x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x200xbf16> loc(#loc97) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_131", + PartOfOutputName = "Conv_131", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x200xbf16>, tensor<80x1x1x200xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc97) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc97) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc97) + } -> tensor<1x80x12x20xbf16> loc(#loc97) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_132", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "997", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "538", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_132", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_132", OutputName = "Add_132"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc98) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc98) + } -> tensor<1x80x12x20xbf16> loc(#loc98) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc98) + } -> tensor<1x80x12x20xbf16> loc(#loc341) + %243 = xten_nn.subgraph (%arg5 = %242: tensor<1x80x12x20xbf16>, %arg6 = %100: tensor<184x80x1x1xbf16>, %arg7 = %99: tensor<184xbf16>) attributes { + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "997", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "541", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_133", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "997", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "541", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_133", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc99) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc99) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc99) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_133", + PartOfOutputName = "Conv_133", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc99) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc99) + } -> tensor<1x184x12x20xbf16> loc(#loc99) + %244 = xten_nn.subgraph (%arg5 = %243: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_135", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_135", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_135", OutputName = "Add_135"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc100) + } -> tensor<1x184x12x20xbf16> loc(#loc100) + %245 = xten_nn.subgraph (%arg5 = %244: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_138", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "545", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_138", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_138", + OutputName = "Clip_138", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc101) + } -> tensor<1x184x12x20xbf16> loc(#loc101) + %246 = xten_nn.subgraph (%arg5 = %245: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "550", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_140", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "548", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_140", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "550", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_140", + OutputName = "Div_140", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc102) + } -> tensor<1x184x12x20xbf16> loc(#loc102) + %247 = xten_nn.subgraph (%arg5 = %243: tensor<1x184x12x20xbf16>, %arg6 = %246: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_141", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "541", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_141", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_141", + OutputName = "Mul_141", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc103) + } -> tensor<1x184x12x20xbf16> loc(#loc103) + %248 = xten_nn.subgraph (%arg5 = %247: tensor<1x184x12x20xbf16>, %arg6 = %98: tensor<184x1x3x3xbf16>, %arg7 = %97: tensor<184xbf16>) attributes { + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "551", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_142", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1000", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "551", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_142", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc104) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc104) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc104) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_142", + PartOfOutputName = "Conv_142", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc104) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc104) + } -> tensor<1x184x12x20xbf16> loc(#loc104) + %249 = xten_nn.subgraph (%arg5 = %248: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_144", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_144", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_144", OutputName = "Add_144"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc105) + } -> tensor<1x184x12x20xbf16> loc(#loc105) + %250 = xten_nn.subgraph (%arg5 = %249: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_147", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "555", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_147", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_147", + OutputName = "Clip_147", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc106) + } -> tensor<1x184x12x20xbf16> loc(#loc106) + %251 = xten_nn.subgraph (%arg5 = %250: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "560", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_149", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "558", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_149", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "560", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_149", + OutputName = "Div_149", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc107) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc107) + } -> tensor<1x184x12x20xbf16> loc(#loc107) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc107) + } -> tensor<1x184x12x20xbf16> loc(#loc107) + %252 = xten_nn.subgraph (%arg5 = %248: tensor<1x184x12x20xbf16>, %arg6 = %251: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_150", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1003", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "551", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_150", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_150", + OutputName = "Mul_150", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc108) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc108) + } -> tensor<1x184x12x20xbf16> loc(#loc108) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc108) + } -> tensor<1x184x12x20xbf16> loc(#loc108) + %253 = xten_nn.subgraph (%arg5 = %252: tensor<1x184x12x20xbf16>, %arg6 = %96: tensor<80x184x1x1xbf16>, %arg7 = %95: tensor<80xbf16>, %arg8 = %242: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_151", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1003", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "561", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_151", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1003", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "561", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_151", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1006", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc109) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc109) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc109) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_151", + PartOfOutputName = "Conv_151", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc109) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc109) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc109) + } -> tensor<1x80x12x20xbf16> loc(#loc109) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_152", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1006", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "561", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_152", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_152", OutputName = "Add_152"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc110) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc110) + } -> tensor<1x80x12x20xbf16> loc(#loc110) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc110) + } -> tensor<1x80x12x20xbf16> loc(#loc342) + %254 = xten_nn.subgraph (%arg5 = %253: tensor<1x80x12x20xbf16>, %arg6 = %94: tensor<184x80x1x1xbf16>, %arg7 = %93: tensor<184xbf16>) attributes { + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1006", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "564", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<184x80x1x1xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_153", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1006", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "564", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_153", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc111) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<184x80x1x1xbf16>) -> tensor<184x1x1x80xbf16> loc(#loc111) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc111) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_153", + PartOfOutputName = "Conv_153", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<184x1x1x80xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc111) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %467 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc111) + } -> tensor<1x184x12x20xbf16> loc(#loc111) + %255 = xten_nn.subgraph (%arg5 = %254: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_155", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_155", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_155", OutputName = "Add_155"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc112) + } -> tensor<1x184x12x20xbf16> loc(#loc112) + %256 = xten_nn.subgraph (%arg5 = %255: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_158", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "568", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_158", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_158", + OutputName = "Clip_158", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc113) + } -> tensor<1x184x12x20xbf16> loc(#loc113) + %257 = xten_nn.subgraph (%arg5 = %256: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "573", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_160", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "571", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_160", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "573", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_160", + OutputName = "Div_160", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc114) + } -> tensor<1x184x12x20xbf16> loc(#loc114) + %258 = xten_nn.subgraph (%arg5 = %254: tensor<1x184x12x20xbf16>, %arg6 = %257: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_161", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "564", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_161", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_161", + OutputName = "Mul_161", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc115) + } -> tensor<1x184x12x20xbf16> loc(#loc115) + %259 = xten_nn.subgraph (%arg5 = %258: tensor<1x184x12x20xbf16>, %arg6 = %92: tensor<184x1x3x3xbf16>, %arg7 = %91: tensor<184xbf16>) attributes { + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "574", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x184x12x20xbf16>, %arg9 = %arg6: tensor<184x1x3x3xbf16>, %arg10 = %arg7: tensor<184xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_162", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1009", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[184, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "574", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_162", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc116) + %465 = tosa.transpose %arg9, %464 : (tensor<184x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x184x1xbf16> loc(#loc116) + %466 = tosa.transpose %arg8, %463 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc116) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_162", + PartOfOutputName = "Conv_162", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<3x3x184x1xbf16>, tensor<184xbf16>) -> tensor<1x12x20x184xbf16> loc(#loc116) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x184xbf16>, tensor<4xi32>) -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %468 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc116) + } -> tensor<1x184x12x20xbf16> loc(#loc116) + %260 = xten_nn.subgraph (%arg5 = %259: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Add_164", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_164", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_164", OutputName = "Add_164"} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc117) + } -> tensor<1x184x12x20xbf16> loc(#loc117) + %261 = xten_nn.subgraph (%arg5 = %260: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Clip_167", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "578", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_167", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_167", + OutputName = "Clip_167", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc118) + } -> tensor<1x184x12x20xbf16> loc(#loc118) + %262 = xten_nn.subgraph (%arg5 = %261: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "583", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Div_169", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "581", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_169", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "583", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_169", + OutputName = "Div_169", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc119) + xten_nn.output %463 : tensor<1x184x12x20xbf16> loc(#loc119) + } -> tensor<1x184x12x20xbf16> loc(#loc119) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc119) + } -> tensor<1x184x12x20xbf16> loc(#loc119) + %263 = xten_nn.subgraph (%arg5 = %259: tensor<1x184x12x20xbf16>, %arg6 = %262: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x184x12x20xbf16>, %arg8 = %arg6: tensor<1x184x12x20xbf16>) attributes { + LayerName = "Mul_170", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1012", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "574", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_170", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_170", + OutputName = "Mul_170", + shift = 0 : i8} : (tensor<1x184x12x20xbf16>, tensor<1x184x12x20xbf16>) -> tensor<1x184x12x20xbf16> loc(#loc120) + xten_nn.output %462 : tensor<1x184x12x20xbf16> loc(#loc120) + } -> tensor<1x184x12x20xbf16> loc(#loc120) + xten_nn.output %461 : tensor<1x184x12x20xbf16> loc(#loc120) + } -> tensor<1x184x12x20xbf16> loc(#loc120) + %264 = xten_nn.subgraph (%arg5 = %263: tensor<1x184x12x20xbf16>, %arg6 = %90: tensor<80x184x1x1xbf16>, %arg7 = %89: tensor<80xbf16>, %arg8 = %253: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Conv_171", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1012", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "584", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x184x12x20xbf16>, %arg10 = %arg6: tensor<80x184x1x1xbf16>, %arg11 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_171", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 184, 12, 20]> : vector<4xindex> + }, + { + Name = "1012", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 184, 1, 1]> : vector<4xindex> + }, + { + Name = "584", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_171", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1015", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc121) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<80x184x1x1xbf16>) -> tensor<80x1x1x184xbf16> loc(#loc121) + %466 = tosa.transpose %arg9, %464 : (tensor<1x184x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x184xbf16> loc(#loc121) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_171", + PartOfOutputName = "Conv_171", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x184xbf16>, tensor<80x1x1x184xbf16>, tensor<80xbf16>) -> tensor<1x12x20x80xbf16> loc(#loc121) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x80xbf16>, tensor<4xi32>) -> tensor<1x80x12x20xbf16> loc(#loc121) + xten_nn.output %468 : tensor<1x80x12x20xbf16> loc(#loc121) + } -> tensor<1x80x12x20xbf16> loc(#loc121) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x80x12x20xbf16>, %arg10 = %arg8: tensor<1x80x12x20xbf16>) attributes { + LayerName = "Add_172", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1015", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "584", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_172", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_172", OutputName = "Add_172"} : (tensor<1x80x12x20xbf16>, tensor<1x80x12x20xbf16>) -> tensor<1x80x12x20xbf16> loc(#loc122) + xten_nn.output %463 : tensor<1x80x12x20xbf16> loc(#loc122) + } -> tensor<1x80x12x20xbf16> loc(#loc122) + xten_nn.output %462 : tensor<1x80x12x20xbf16> loc(#loc122) + } -> tensor<1x80x12x20xbf16> loc(#loc343) + %265 = xten_nn.subgraph (%arg5 = %264: tensor<1x80x12x20xbf16>, %arg6 = %88: tensor<480x80x1x1xbf16>, %arg7 = %87: tensor<480xbf16>) attributes { + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1015", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "587", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x12x20xbf16>, %arg9 = %arg6: tensor<480x80x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_173", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 12, 20]> : vector<4xindex> + }, + { + Name = "1015", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 80, 1, 1]> : vector<4xindex> + }, + { + Name = "587", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_173", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc123) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x80x1x1xbf16>) -> tensor<480x1x1x80xbf16> loc(#loc123) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x80xbf16> loc(#loc123) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_173", + PartOfOutputName = "Conv_173", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x80xbf16>, tensor<480x1x1x80xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc123) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %467 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc123) + } -> tensor<1x480x12x20xbf16> loc(#loc123) + %266 = xten_nn.subgraph (%arg5 = %265: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_175", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_175", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_175", OutputName = "Add_175"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc124) + } -> tensor<1x480x12x20xbf16> loc(#loc124) + %267 = xten_nn.subgraph (%arg5 = %266: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_178", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "591", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_178", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_178", + OutputName = "Clip_178", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc125) + } -> tensor<1x480x12x20xbf16> loc(#loc125) + %268 = xten_nn.subgraph (%arg5 = %267: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "596", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_180", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "594", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_180", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "596", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_180", + OutputName = "Div_180", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc126) + } -> tensor<1x480x12x20xbf16> loc(#loc126) + %269 = xten_nn.subgraph (%arg5 = %265: tensor<1x480x12x20xbf16>, %arg6 = %268: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_181", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "587", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_181", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_181", + OutputName = "Mul_181", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc127) + } -> tensor<1x480x12x20xbf16> loc(#loc127) + %270 = xten_nn.subgraph (%arg5 = %269: tensor<1x480x12x20xbf16>, %arg6 = %86: tensor<480x1x3x3xbf16>, %arg7 = %85: tensor<480xbf16>) attributes { + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "597", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<480x1x3x3xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_182", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1018", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "597", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_182", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc128) + %465 = tosa.transpose %arg9, %464 : (tensor<480x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x480x1xbf16> loc(#loc128) + %466 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc128) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_182", + PartOfOutputName = "Conv_182", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<3x3x480x1xbf16>, tensor<480xbf16>) -> tensor<1x12x20x480xbf16> loc(#loc128) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x480xbf16>, tensor<4xi32>) -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %468 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc128) + } -> tensor<1x480x12x20xbf16> loc(#loc128) + %271 = xten_nn.subgraph (%arg5 = %270: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Add_184", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_184", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_184", OutputName = "Add_184"} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc129) + } -> tensor<1x480x12x20xbf16> loc(#loc129) + %272 = xten_nn.subgraph (%arg5 = %271: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Clip_187", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "601", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_187", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_187", + OutputName = "Clip_187", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc130) + } -> tensor<1x480x12x20xbf16> loc(#loc130) + %273 = xten_nn.subgraph (%arg5 = %272: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "606", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Div_189", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "604", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_189", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "606", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_189", + OutputName = "Div_189", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc131) + xten_nn.output %463 : tensor<1x480x12x20xbf16> loc(#loc131) + } -> tensor<1x480x12x20xbf16> loc(#loc131) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc131) + } -> tensor<1x480x12x20xbf16> loc(#loc131) + %274 = xten_nn.subgraph (%arg5 = %270: tensor<1x480x12x20xbf16>, %arg6 = %273: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "597", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_190", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_190", + OutputName = "Mul_190", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc132) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc132) + } -> tensor<1x480x12x20xbf16> loc(#loc132) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc132) + } -> tensor<1x480x12x20xbf16> loc(#loc132) + %275 = xten_nn.subgraph (%arg5 = %274: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_190_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1021", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_191_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 60 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x480x12x20xbf16>) -> tensor<1x480x1x240xbf16> loc(#loc344) + xten_nn.output %461 : tensor<1x480x1x240xbf16> loc(#loc344) + } -> tensor<1x480x1x240xbf16> loc(#loc344) + %276 = xten_nn.subgraph (%arg5 = %275: tensor<1x480x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_191", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_191_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_191", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "607", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_191_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 480 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x480x1x240xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc133) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc133) + } -> tensor<1x480x1x1xbf16> loc(#loc133) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc133) + } -> tensor<1x480x1x1xbf16> loc(#loc133) + %277 = xten_nn.subgraph (%arg5 = %276: tensor<1x480x1x1xbf16>, %arg6 = %84: tensor<120x480x1x1xbf16>, %arg7 = %83: tensor<120xbf16>) attributes { + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "607", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x1x1xbf16>, %arg9 = %arg6: tensor<120x480x1x1xbf16>, %arg10 = %arg7: tensor<120xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_192", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "608", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "607", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[120, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_193", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<120x480x1x1xbf16>) -> tensor<120x1x1x480xbf16> loc(#loc345) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc345) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_192", + PartOfOutputName = "Conv_192", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x480xbf16>, tensor<120x1x1x480xbf16>, tensor<120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc134) + %465 = tosa.clamp %464 { + LayerName = "Relu_193", + OutputName = "Relu_193", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x120xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc135) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x120xbf16>) -> tensor<1x120x1x1xbf16> loc(#loc345) + xten_nn.output %466 : tensor<1x120x1x1xbf16> loc(#loc135) + } -> tensor<1x120x1x1xbf16> loc(#loc345) + xten_nn.output %461 : tensor<1x120x1x1xbf16> loc(#loc345) + } -> tensor<1x120x1x1xbf16> loc(#loc345) + %278 = xten_nn.subgraph (%arg5 = %277: tensor<1x120x1x1xbf16>, %arg6 = %82: tensor<480x120x1x1xbf16>, %arg7 = %81: tensor<480xbf16>) attributes { + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "609", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x120x1x1xbf16>, %arg9 = %arg6: tensor<480x120x1x1xbf16>, %arg10 = %arg7: tensor<480xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_194", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "610", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "609", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[480, 120, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.11.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_194", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<480x120x1x1xbf16>) -> tensor<480x1x1x120xbf16> loc(#loc136) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x120x1x1xbf16>) -> tensor<1x1x1x120xbf16> loc(#loc136) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_194", + PartOfOutputName = "Conv_194", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x120xbf16>, tensor<480x1x1x120xbf16>, tensor<480xbf16>) -> tensor<1x1x1x480xbf16> loc(#loc136) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x480xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %465 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc136) + } -> tensor<1x480x1x1xbf16> loc(#loc136) + %279 = xten_nn.subgraph (%arg5 = %278: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Add_196", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "611", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_196", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_196", OutputName = "Add_196"} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc137) + } -> tensor<1x480x1x1xbf16> loc(#loc137) + %280 = xten_nn.subgraph (%arg5 = %279: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Clip_199", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "613", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_199", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_199", + OutputName = "Clip_199", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc138) + xten_nn.output %462 : tensor<1x480x1x1xbf16> loc(#loc138) + } -> tensor<1x480x1x1xbf16> loc(#loc138) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc138) + } -> tensor<1x480x1x1xbf16> loc(#loc138) + %281 = xten_nn.subgraph (%arg5 = %280: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Div_201", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_201", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_201", + OutputName = "Div_201", + shift = 0 : i8} : (tensor<1x480x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x480x1x1xbf16> loc(#loc139) + xten_nn.output %463 : tensor<1x480x1x1xbf16> loc(#loc139) + } -> tensor<1x480x1x1xbf16> loc(#loc139) + xten_nn.output %461 : tensor<1x480x1x1xbf16> loc(#loc139) + } -> tensor<1x480x1x1xbf16> loc(#loc139) + %282 = xten_nn.subgraph (%arg5 = %281: tensor<1x480x1x1xbf16>) attributes { + LayerName = "Mul_202_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_202_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 480 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x480x1x1xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc140) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc140) + } -> tensor<1x480x12x20xbf16> loc(#loc140) + %283 = xten_nn.subgraph (%arg5 = %282: tensor<1x480x12x20xbf16>, %arg6 = %274: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x480x12x20xbf16>, %arg8 = %arg6: tensor<1x480x12x20xbf16>) attributes { + LayerName = "Mul_202_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "618", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "616", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_202_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_202", + OutputName = "Mul_202", + shift = 0 : i8} : (tensor<1x480x12x20xbf16>, tensor<1x480x12x20xbf16>) -> tensor<1x480x12x20xbf16> loc(#loc140) + xten_nn.output %462 : tensor<1x480x12x20xbf16> loc(#loc140) + } -> tensor<1x480x12x20xbf16> loc(#loc140) + xten_nn.output %461 : tensor<1x480x12x20xbf16> loc(#loc140) + } -> tensor<1x480x12x20xbf16> loc(#loc140) + %284 = xten_nn.subgraph (%arg5 = %283: tensor<1x480x12x20xbf16>, %arg6 = %80: tensor<112x480x1x1xbf16>, %arg7 = %79: tensor<112xbf16>) attributes { + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Name = "618", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "619", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x480x12x20xbf16>, %arg9 = %arg6: tensor<112x480x1x1xbf16>, %arg10 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_203", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "619", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 480, 12, 20]> : vector<4xindex> + }, + { + Name = "618", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 480, 1, 1]> : vector<4xindex> + }, + { + Name = "619", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_203", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc141) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<112x480x1x1xbf16>) -> tensor<112x1x1x480xbf16> loc(#loc141) + %465 = tosa.transpose %arg8, %463 : (tensor<1x480x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x480xbf16> loc(#loc141) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_203", + PartOfOutputName = "Conv_203", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x480xbf16>, tensor<112x1x1x480xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc141) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc141) + xten_nn.output %467 : tensor<1x112x12x20xbf16> loc(#loc141) + } -> tensor<1x112x12x20xbf16> loc(#loc141) + xten_nn.output %461 : tensor<1x112x12x20xbf16> loc(#loc141) + } -> tensor<1x112x12x20xbf16> loc(#loc141) + %285 = xten_nn.subgraph (%arg5 = %284: tensor<1x112x12x20xbf16>, %arg6 = %78: tensor<672x112x1x1xbf16>, %arg7 = %77: tensor<672xbf16>) attributes { + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "619", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "1028", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_204", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "619", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "1028", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_204", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc142) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc142) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc142) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_204", + PartOfOutputName = "Conv_204", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc142) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc142) + } -> tensor<1x672x12x20xbf16> loc(#loc142) + %286 = xten_nn.subgraph (%arg5 = %285: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_206", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_206", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_206", OutputName = "Add_206"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc143) + } -> tensor<1x672x12x20xbf16> loc(#loc143) + %287 = xten_nn.subgraph (%arg5 = %286: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_209", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "625", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_209", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_209", + OutputName = "Clip_209", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc144) + } -> tensor<1x672x12x20xbf16> loc(#loc144) + %288 = xten_nn.subgraph (%arg5 = %287: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "630", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_211", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "628", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_211", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "630", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_211", + OutputName = "Div_211", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc145) + } -> tensor<1x672x12x20xbf16> loc(#loc145) + %289 = xten_nn.subgraph (%arg5 = %285: tensor<1x672x12x20xbf16>, %arg6 = %288: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_212", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1024", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_212", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_212", + OutputName = "Mul_212", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc146) + } -> tensor<1x672x12x20xbf16> loc(#loc146) + %290 = xten_nn.subgraph (%arg5 = %289: tensor<1x672x12x20xbf16>, %arg6 = %76: tensor<672x1x3x3xbf16>, %arg7 = %75: tensor<672xbf16>) attributes { + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "631", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x3x3xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_213", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1027", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 3, 3]> : vector<4xindex> + }, + { + Name = "631", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_213", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 3 : ui8, + config.kernel_width = 3 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc147) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x3x3xbf16>, tensor<4xi32>) -> tensor<3x3x672x1xbf16> loc(#loc147) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc147) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_213", + PartOfOutputName = "Conv_213", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<3x3x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc147) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc147) + } -> tensor<1x672x12x20xbf16> loc(#loc147) + %291 = xten_nn.subgraph (%arg5 = %290: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_215", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_215", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_215", OutputName = "Add_215"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc148) + } -> tensor<1x672x12x20xbf16> loc(#loc148) + %292 = xten_nn.subgraph (%arg5 = %291: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_218", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "635", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_218", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_218", + OutputName = "Clip_218", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc149) + } -> tensor<1x672x12x20xbf16> loc(#loc149) + %293 = xten_nn.subgraph (%arg5 = %292: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "640", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_220", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "638", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_220", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "640", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_220", + OutputName = "Div_220", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc150) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc150) + } -> tensor<1x672x12x20xbf16> loc(#loc150) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc150) + } -> tensor<1x672x12x20xbf16> loc(#loc150) + %294 = xten_nn.subgraph (%arg5 = %290: tensor<1x672x12x20xbf16>, %arg6 = %293: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "631", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_221", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_221", + OutputName = "Mul_221", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc151) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc151) + } -> tensor<1x672x12x20xbf16> loc(#loc151) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc151) + } -> tensor<1x672x12x20xbf16> loc(#loc151) + %295 = xten_nn.subgraph (%arg5 = %294: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_221_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1030", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_222_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc346) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc346) + } -> tensor<1x672x1x240xbf16> loc(#loc346) + %296 = xten_nn.subgraph (%arg5 = %295: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_222", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_222_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_222", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "641", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_222_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc152) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc152) + } -> tensor<1x672x1x1xbf16> loc(#loc152) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc152) + } -> tensor<1x672x1x1xbf16> loc(#loc152) + %297 = xten_nn.subgraph (%arg5 = %296: tensor<1x672x1x1xbf16>, %arg6 = %74: tensor<168x672x1x1xbf16>, %arg7 = %73: tensor<168xbf16>) attributes { + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "641", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_223", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "642", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "641", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_224", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc347) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc347) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_223", + PartOfOutputName = "Conv_223", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc153) + %465 = tosa.clamp %464 { + LayerName = "Relu_224", + OutputName = "Relu_224", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc154) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc347) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc154) + } -> tensor<1x168x1x1xbf16> loc(#loc347) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc347) + } -> tensor<1x168x1x1xbf16> loc(#loc347) + %298 = xten_nn.subgraph (%arg5 = %297: tensor<1x168x1x1xbf16>, %arg6 = %72: tensor<672x168x1x1xbf16>, %arg7 = %71: tensor<672xbf16>) attributes { + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "643", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_225", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "644", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "643", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.12.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_225", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc155) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc155) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_225", + PartOfOutputName = "Conv_225", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc155) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc155) + } -> tensor<1x672x1x1xbf16> loc(#loc155) + %299 = xten_nn.subgraph (%arg5 = %298: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_227", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "645", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_227", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_227", OutputName = "Add_227"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc156) + } -> tensor<1x672x1x1xbf16> loc(#loc156) + %300 = xten_nn.subgraph (%arg5 = %299: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_230", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "647", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_230", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_230", + OutputName = "Clip_230", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc157) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc157) + } -> tensor<1x672x1x1xbf16> loc(#loc157) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc157) + } -> tensor<1x672x1x1xbf16> loc(#loc157) + %301 = xten_nn.subgraph (%arg5 = %300: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_232", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_232", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_232", + OutputName = "Div_232", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc158) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc158) + } -> tensor<1x672x1x1xbf16> loc(#loc158) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc158) + } -> tensor<1x672x1x1xbf16> loc(#loc158) + %302 = xten_nn.subgraph (%arg5 = %301: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Mul_233_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_233_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc159) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc159) + } -> tensor<1x672x12x20xbf16> loc(#loc159) + %303 = xten_nn.subgraph (%arg5 = %302: tensor<1x672x12x20xbf16>, %arg6 = %294: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_233_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "652", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "650", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_233_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_233", + OutputName = "Mul_233", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc159) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc159) + } -> tensor<1x672x12x20xbf16> loc(#loc159) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc159) + } -> tensor<1x672x12x20xbf16> loc(#loc159) + %304 = xten_nn.subgraph (%arg5 = %303: tensor<1x672x12x20xbf16>, %arg6 = %70: tensor<112x672x1x1xbf16>, %arg7 = %69: tensor<112xbf16>, %arg8 = %284: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Conv_234", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "652", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "653", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x672x12x20xbf16>, %arg10 = %arg6: tensor<112x672x1x1xbf16>, %arg11 = %arg7: tensor<112xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_234", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "652", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[112, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "653", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_234", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1033", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc160) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<112x672x1x1xbf16>) -> tensor<112x1x1x672xbf16> loc(#loc160) + %466 = tosa.transpose %arg9, %464 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc160) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_234", + PartOfOutputName = "Conv_234", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<112x1x1x672xbf16>, tensor<112xbf16>) -> tensor<1x12x20x112xbf16> loc(#loc160) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x112xbf16>, tensor<4xi32>) -> tensor<1x112x12x20xbf16> loc(#loc160) + xten_nn.output %468 : tensor<1x112x12x20xbf16> loc(#loc160) + } -> tensor<1x112x12x20xbf16> loc(#loc160) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x112x12x20xbf16>, %arg10 = %arg8: tensor<1x112x12x20xbf16>) attributes { + LayerName = "Add_235", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1033", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "653", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_235", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_235", OutputName = "Add_235"} : (tensor<1x112x12x20xbf16>, tensor<1x112x12x20xbf16>) -> tensor<1x112x12x20xbf16> loc(#loc161) + xten_nn.output %463 : tensor<1x112x12x20xbf16> loc(#loc161) + } -> tensor<1x112x12x20xbf16> loc(#loc161) + xten_nn.output %462 : tensor<1x112x12x20xbf16> loc(#loc161) + } -> tensor<1x112x12x20xbf16> loc(#loc348) + %305 = xten_nn.subgraph (%arg5 = %304: tensor<1x112x12x20xbf16>, %arg6 = %68: tensor<672x112x1x1xbf16>, %arg7 = %67: tensor<672xbf16>) attributes { + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "1033", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "656", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x112x12x20xbf16>, %arg9 = %arg6: tensor<672x112x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_236", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 112, 12, 20]> : vector<4xindex> + }, + { + Name = "1033", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 112, 1, 1]> : vector<4xindex> + }, + { + Name = "656", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_236", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc162) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x112x1x1xbf16>) -> tensor<672x1x1x112xbf16> loc(#loc162) + %465 = tosa.transpose %arg8, %463 : (tensor<1x112x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x112xbf16> loc(#loc162) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_236", + PartOfOutputName = "Conv_236", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x112xbf16>, tensor<672x1x1x112xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc162) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %467 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc162) + } -> tensor<1x672x12x20xbf16> loc(#loc162) + %306 = xten_nn.subgraph (%arg5 = %305: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_238", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_238", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_238", OutputName = "Add_238"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc163) + } -> tensor<1x672x12x20xbf16> loc(#loc163) + %307 = xten_nn.subgraph (%arg5 = %306: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_241", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "660", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_241", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_241", + OutputName = "Clip_241", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc164) + } -> tensor<1x672x12x20xbf16> loc(#loc164) + %308 = xten_nn.subgraph (%arg5 = %307: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "665", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_243", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "663", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_243", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "665", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_243", + OutputName = "Div_243", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc165) + } -> tensor<1x672x12x20xbf16> loc(#loc165) + %309 = xten_nn.subgraph (%arg5 = %305: tensor<1x672x12x20xbf16>, %arg6 = %308: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_244", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "656", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_244", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_244", + OutputName = "Mul_244", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc166) + } -> tensor<1x672x12x20xbf16> loc(#loc166) + %310 = xten_nn.subgraph (%arg5 = %309: tensor<1x672x12x20xbf16>, %arg6 = %66: tensor<672x1x9x9xbf16>, %arg7 = %65: tensor<672xbf16>) attributes { + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "666", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<672x1x9x9xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_245", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1036", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "666", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_245", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc167) + %465 = tosa.transpose %arg9, %464 : (tensor<672x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x672x1xbf16> loc(#loc167) + %466 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc167) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_245", + PartOfOutputName = "Conv_245", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<9x9x672x1xbf16>, tensor<672xbf16>) -> tensor<1x12x20x672xbf16> loc(#loc167) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x672xbf16>, tensor<4xi32>) -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %468 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc167) + } -> tensor<1x672x12x20xbf16> loc(#loc167) + %311 = xten_nn.subgraph (%arg5 = %310: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Add_247", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_247", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_247", OutputName = "Add_247"} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc168) + } -> tensor<1x672x12x20xbf16> loc(#loc168) + %312 = xten_nn.subgraph (%arg5 = %311: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Clip_250", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "670", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_250", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_250", + OutputName = "Clip_250", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc169) + } -> tensor<1x672x12x20xbf16> loc(#loc169) + %313 = xten_nn.subgraph (%arg5 = %312: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "675", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Div_252", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "673", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_252", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "675", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_252", + OutputName = "Div_252", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc170) + xten_nn.output %463 : tensor<1x672x12x20xbf16> loc(#loc170) + } -> tensor<1x672x12x20xbf16> loc(#loc170) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc170) + } -> tensor<1x672x12x20xbf16> loc(#loc170) + %314 = xten_nn.subgraph (%arg5 = %310: tensor<1x672x12x20xbf16>, %arg6 = %313: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "666", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_253", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_253", + OutputName = "Mul_253", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc171) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc171) + } -> tensor<1x672x12x20xbf16> loc(#loc171) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc171) + } -> tensor<1x672x12x20xbf16> loc(#loc171) + %315 = xten_nn.subgraph (%arg5 = %314: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_253_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1039", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_254_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 84 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x672x12x20xbf16>) -> tensor<1x672x1x240xbf16> loc(#loc349) + xten_nn.output %461 : tensor<1x672x1x240xbf16> loc(#loc349) + } -> tensor<1x672x1x240xbf16> loc(#loc349) + %316 = xten_nn.subgraph (%arg5 = %315: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_254", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_254_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_254", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "676", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_254_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 672 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x672x1x240xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc172) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc172) + } -> tensor<1x672x1x1xbf16> loc(#loc172) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc172) + } -> tensor<1x672x1x1xbf16> loc(#loc172) + %317 = xten_nn.subgraph (%arg5 = %316: tensor<1x672x1x1xbf16>, %arg6 = %64: tensor<168x672x1x1xbf16>, %arg7 = %63: tensor<168xbf16>) attributes { + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "676", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x1x1xbf16>, %arg9 = %arg6: tensor<168x672x1x1xbf16>, %arg10 = %arg7: tensor<168xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_255", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "677", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "676", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[168, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_256", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<168x672x1x1xbf16>) -> tensor<168x1x1x672xbf16> loc(#loc350) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc350) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_255", + PartOfOutputName = "Conv_255", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x672xbf16>, tensor<168x1x1x672xbf16>, tensor<168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc173) + %465 = tosa.clamp %464 { + LayerName = "Relu_256", + OutputName = "Relu_256", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x168xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc174) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x168xbf16>) -> tensor<1x168x1x1xbf16> loc(#loc350) + xten_nn.output %466 : tensor<1x168x1x1xbf16> loc(#loc174) + } -> tensor<1x168x1x1xbf16> loc(#loc350) + xten_nn.output %461 : tensor<1x168x1x1xbf16> loc(#loc350) + } -> tensor<1x168x1x1xbf16> loc(#loc350) + %318 = xten_nn.subgraph (%arg5 = %317: tensor<1x168x1x1xbf16>, %arg6 = %62: tensor<672x168x1x1xbf16>, %arg7 = %61: tensor<672xbf16>) attributes { + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "678", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x168x1x1xbf16>, %arg9 = %arg6: tensor<672x168x1x1xbf16>, %arg10 = %arg7: tensor<672xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_257", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "679", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "678", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[672, 168, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.13.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_257", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<672x168x1x1xbf16>) -> tensor<672x1x1x168xbf16> loc(#loc175) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x168x1x1xbf16>) -> tensor<1x1x1x168xbf16> loc(#loc175) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_257", + PartOfOutputName = "Conv_257", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x168xbf16>, tensor<672x1x1x168xbf16>, tensor<672xbf16>) -> tensor<1x1x1x672xbf16> loc(#loc175) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x672xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %465 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc175) + } -> tensor<1x672x1x1xbf16> loc(#loc175) + %319 = xten_nn.subgraph (%arg5 = %318: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Add_259", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "680", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_259", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_259", OutputName = "Add_259"} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc176) + } -> tensor<1x672x1x1xbf16> loc(#loc176) + %320 = xten_nn.subgraph (%arg5 = %319: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Clip_262", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "682", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_262", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_262", + OutputName = "Clip_262", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc177) + xten_nn.output %462 : tensor<1x672x1x1xbf16> loc(#loc177) + } -> tensor<1x672x1x1xbf16> loc(#loc177) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc177) + } -> tensor<1x672x1x1xbf16> loc(#loc177) + %321 = xten_nn.subgraph (%arg5 = %320: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Div_264", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_264", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_264", + OutputName = "Div_264", + shift = 0 : i8} : (tensor<1x672x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x672x1x1xbf16> loc(#loc178) + xten_nn.output %463 : tensor<1x672x1x1xbf16> loc(#loc178) + } -> tensor<1x672x1x1xbf16> loc(#loc178) + xten_nn.output %461 : tensor<1x672x1x1xbf16> loc(#loc178) + } -> tensor<1x672x1x1xbf16> loc(#loc178) + %322 = xten_nn.subgraph (%arg5 = %321: tensor<1x672x1x1xbf16>) attributes { + LayerName = "Mul_265_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_265_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 672 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x672x1x1xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc179) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc179) + } -> tensor<1x672x12x20xbf16> loc(#loc179) + %323 = xten_nn.subgraph (%arg5 = %322: tensor<1x672x12x20xbf16>, %arg6 = %314: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x672x12x20xbf16>, %arg8 = %arg6: tensor<1x672x12x20xbf16>) attributes { + LayerName = "Mul_265_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "687", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "685", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_265_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_265", + OutputName = "Mul_265", + shift = 0 : i8} : (tensor<1x672x12x20xbf16>, tensor<1x672x12x20xbf16>) -> tensor<1x672x12x20xbf16> loc(#loc179) + xten_nn.output %462 : tensor<1x672x12x20xbf16> loc(#loc179) + } -> tensor<1x672x12x20xbf16> loc(#loc179) + xten_nn.output %461 : tensor<1x672x12x20xbf16> loc(#loc179) + } -> tensor<1x672x12x20xbf16> loc(#loc179) + %324 = xten_nn.subgraph (%arg5 = %323: tensor<1x672x12x20xbf16>, %arg6 = %60: tensor<160x672x1x1xbf16>, %arg7 = %59: tensor<160xbf16>) attributes { + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "687", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "688", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x672x12x20xbf16>, %arg9 = %arg6: tensor<160x672x1x1xbf16>, %arg10 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_266", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "688", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 672, 12, 20]> : vector<4xindex> + }, + { + Name = "687", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 672, 1, 1]> : vector<4xindex> + }, + { + Name = "688", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_266", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc180) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<160x672x1x1xbf16>) -> tensor<160x1x1x672xbf16> loc(#loc180) + %465 = tosa.transpose %arg8, %463 : (tensor<1x672x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x672xbf16> loc(#loc180) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_266", + PartOfOutputName = "Conv_266", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x672xbf16>, tensor<160x1x1x672xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc180) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc180) + xten_nn.output %467 : tensor<1x160x12x20xbf16> loc(#loc180) + } -> tensor<1x160x12x20xbf16> loc(#loc180) + xten_nn.output %461 : tensor<1x160x12x20xbf16> loc(#loc180) + } -> tensor<1x160x12x20xbf16> loc(#loc180) + %325 = xten_nn.subgraph (%arg5 = %324: tensor<1x160x12x20xbf16>, %arg6 = %58: tensor<960x160x1x1xbf16>, %arg7 = %57: tensor<960xbf16>) attributes { + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "688", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "1046", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_267", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "688", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "1046", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_267", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc181) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc181) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc181) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_267", + PartOfOutputName = "Conv_267", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc181) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc181) + } -> tensor<1x960x12x20xbf16> loc(#loc181) + %326 = xten_nn.subgraph (%arg5 = %325: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_269", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_269", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_269", OutputName = "Add_269"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc182) + } -> tensor<1x960x12x20xbf16> loc(#loc182) + %327 = xten_nn.subgraph (%arg5 = %326: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_272", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "694", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_272", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_272", + OutputName = "Clip_272", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc183) + } -> tensor<1x960x12x20xbf16> loc(#loc183) + %328 = xten_nn.subgraph (%arg5 = %327: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "699", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_274", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "697", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_274", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "699", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_274", + OutputName = "Div_274", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc184) + } -> tensor<1x960x12x20xbf16> loc(#loc184) + %329 = xten_nn.subgraph (%arg5 = %325: tensor<1x960x12x20xbf16>, %arg6 = %328: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_275", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1042", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_275", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_275", + OutputName = "Mul_275", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc185) + } -> tensor<1x960x12x20xbf16> loc(#loc185) + %330 = xten_nn.subgraph (%arg5 = %329: tensor<1x960x12x20xbf16>, %arg6 = %56: tensor<960x1x9x9xbf16>, %arg7 = %55: tensor<960xbf16>) attributes { + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "700", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_276", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1045", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "700", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_276", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc186) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc186) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc186) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_276", + PartOfOutputName = "Conv_276", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc186) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc186) + } -> tensor<1x960x12x20xbf16> loc(#loc186) + %331 = xten_nn.subgraph (%arg5 = %330: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_278", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_278", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_278", OutputName = "Add_278"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc187) + } -> tensor<1x960x12x20xbf16> loc(#loc187) + %332 = xten_nn.subgraph (%arg5 = %331: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_281", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "704", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_281", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_281", + OutputName = "Clip_281", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc188) + } -> tensor<1x960x12x20xbf16> loc(#loc188) + %333 = xten_nn.subgraph (%arg5 = %332: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "709", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_283", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "707", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_283", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "709", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_283", + OutputName = "Div_283", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc189) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc189) + } -> tensor<1x960x12x20xbf16> loc(#loc189) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc189) + } -> tensor<1x960x12x20xbf16> loc(#loc189) + %334 = xten_nn.subgraph (%arg5 = %330: tensor<1x960x12x20xbf16>, %arg6 = %333: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "700", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_284", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_284", + OutputName = "Mul_284", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc190) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc190) + } -> tensor<1x960x12x20xbf16> loc(#loc190) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc190) + } -> tensor<1x960x12x20xbf16> loc(#loc190) + %335 = xten_nn.subgraph (%arg5 = %334: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_284_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1048", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_285_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc351) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc351) + } -> tensor<1x960x1x240xbf16> loc(#loc351) + %336 = xten_nn.subgraph (%arg5 = %335: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_285", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_285_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_285", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "710", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_285_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc191) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc191) + } -> tensor<1x960x1x1xbf16> loc(#loc191) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc191) + } -> tensor<1x960x1x1xbf16> loc(#loc191) + %337 = xten_nn.subgraph (%arg5 = %336: tensor<1x960x1x1xbf16>, %arg6 = %54: tensor<240x960x1x1xbf16>, %arg7 = %53: tensor<240xbf16>) attributes { + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "710", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_286", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "711", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "710", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_287", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc352) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc352) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_286", + PartOfOutputName = "Conv_286", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc192) + %465 = tosa.clamp %464 { + LayerName = "Relu_287", + OutputName = "Relu_287", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc193) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc352) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc193) + } -> tensor<1x240x1x1xbf16> loc(#loc352) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc352) + } -> tensor<1x240x1x1xbf16> loc(#loc352) + %338 = xten_nn.subgraph (%arg5 = %337: tensor<1x240x1x1xbf16>, %arg6 = %52: tensor<960x240x1x1xbf16>, %arg7 = %51: tensor<960xbf16>) attributes { + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "712", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_288", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "713", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "712", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.14.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_288", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc194) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc194) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_288", + PartOfOutputName = "Conv_288", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc194) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc194) + } -> tensor<1x960x1x1xbf16> loc(#loc194) + %339 = xten_nn.subgraph (%arg5 = %338: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_290", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "714", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_290", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_290", OutputName = "Add_290"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc195) + } -> tensor<1x960x1x1xbf16> loc(#loc195) + %340 = xten_nn.subgraph (%arg5 = %339: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_293", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "716", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_293", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_293", + OutputName = "Clip_293", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc196) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc196) + } -> tensor<1x960x1x1xbf16> loc(#loc196) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc196) + } -> tensor<1x960x1x1xbf16> loc(#loc196) + %341 = xten_nn.subgraph (%arg5 = %340: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_295", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_295", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_295", + OutputName = "Div_295", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc197) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc197) + } -> tensor<1x960x1x1xbf16> loc(#loc197) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc197) + } -> tensor<1x960x1x1xbf16> loc(#loc197) + %342 = xten_nn.subgraph (%arg5 = %341: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Mul_296_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_296_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc198) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc198) + } -> tensor<1x960x12x20xbf16> loc(#loc198) + %343 = xten_nn.subgraph (%arg5 = %342: tensor<1x960x12x20xbf16>, %arg6 = %334: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_296_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "721", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "719", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_296_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_296", + OutputName = "Mul_296", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc198) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc198) + } -> tensor<1x960x12x20xbf16> loc(#loc198) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc198) + } -> tensor<1x960x12x20xbf16> loc(#loc198) + %344 = xten_nn.subgraph (%arg5 = %343: tensor<1x960x12x20xbf16>, %arg6 = %50: tensor<160x960x1x1xbf16>, %arg7 = %49: tensor<160xbf16>, %arg8 = %324: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_297", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "721", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "722", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_297", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "721", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "722", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_297", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1051", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc199) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc199) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc199) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_297", + PartOfOutputName = "Conv_297", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc199) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc199) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc199) + } -> tensor<1x160x12x20xbf16> loc(#loc199) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_298", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1051", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "722", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_298", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_298", OutputName = "Add_298"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc200) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc200) + } -> tensor<1x160x12x20xbf16> loc(#loc200) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc200) + } -> tensor<1x160x12x20xbf16> loc(#loc353) + %345 = xten_nn.subgraph (%arg5 = %344: tensor<1x160x12x20xbf16>, %arg6 = %48: tensor<960x160x1x1xbf16>, %arg7 = %47: tensor<960xbf16>) attributes { + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1051", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "725", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_299", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1051", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "725", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_299", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc201) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc201) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc201) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_299", + PartOfOutputName = "Conv_299", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc201) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc201) + } -> tensor<1x960x12x20xbf16> loc(#loc201) + %346 = xten_nn.subgraph (%arg5 = %345: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_301", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_301", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_301", OutputName = "Add_301"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc202) + } -> tensor<1x960x12x20xbf16> loc(#loc202) + %347 = xten_nn.subgraph (%arg5 = %346: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_304", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "729", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_304", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_304", + OutputName = "Clip_304", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc203) + } -> tensor<1x960x12x20xbf16> loc(#loc203) + %348 = xten_nn.subgraph (%arg5 = %347: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "734", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_306", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "732", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_306", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "734", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_306", + OutputName = "Div_306", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc204) + } -> tensor<1x960x12x20xbf16> loc(#loc204) + %349 = xten_nn.subgraph (%arg5 = %345: tensor<1x960x12x20xbf16>, %arg6 = %348: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_307", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "725", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_307", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_307", + OutputName = "Mul_307", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc205) + } -> tensor<1x960x12x20xbf16> loc(#loc205) + %350 = xten_nn.subgraph (%arg5 = %349: tensor<1x960x12x20xbf16>, %arg6 = %46: tensor<960x1x9x9xbf16>, %arg7 = %45: tensor<960xbf16>) attributes { + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "735", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x12x20xbf16>, %arg9 = %arg6: tensor<960x1x9x9xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[4, 4], [4, 4]], + LayerName = "Conv_308", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "CMHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1054", + Port = "data_io.wts", + SubPort = "wts_data", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 1, 9, 9]> : vector<4xindex> + }, + { + Name = "735", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_308", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "DepthwiseConv2dBf16", + With = { + config.act = 0 : ui8, + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.kernel_height = 9 : ui8, + config.kernel_width = 9 : ui8, + config.stride = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[2, 3, 0, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc206) + %465 = tosa.transpose %arg9, %464 : (tensor<960x1x9x9xbf16>, tensor<4xi32>) -> tensor<9x9x960x1xbf16> loc(#loc206) + %466 = tosa.transpose %arg8, %463 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc206) + %467 = tosa.depthwise_conv2d %466, %465, %arg10 { + PartOfLayerName = "Conv_308", + PartOfOutputName = "Conv_308", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<9x9x960x1xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc206) + %468 = tosa.transpose %467, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %468 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc206) + } -> tensor<1x960x12x20xbf16> loc(#loc206) + %351 = xten_nn.subgraph (%arg5 = %350: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_310", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_310", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_310", OutputName = "Add_310"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc207) + } -> tensor<1x960x12x20xbf16> loc(#loc207) + %352 = xten_nn.subgraph (%arg5 = %351: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_313", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "739", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_313", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_313", + OutputName = "Clip_313", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc208) + } -> tensor<1x960x12x20xbf16> loc(#loc208) + %353 = xten_nn.subgraph (%arg5 = %352: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "744", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_315", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "742", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_315", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "744", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_315", + OutputName = "Div_315", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc209) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc209) + } -> tensor<1x960x12x20xbf16> loc(#loc209) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc209) + } -> tensor<1x960x12x20xbf16> loc(#loc209) + %354 = xten_nn.subgraph (%arg5 = %350: tensor<1x960x12x20xbf16>, %arg6 = %353: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "735", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_316", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_316", + OutputName = "Mul_316", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc210) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc210) + } -> tensor<1x960x12x20xbf16> loc(#loc210) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc210) + } -> tensor<1x960x12x20xbf16> loc(#loc210) + %355 = xten_nn.subgraph (%arg5 = %354: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_316_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1057", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_317_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc354) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc354) + } -> tensor<1x960x1x240xbf16> loc(#loc354) + %356 = xten_nn.subgraph (%arg5 = %355: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_317", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_317_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_317", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "745", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_317_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc211) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc211) + } -> tensor<1x960x1x1xbf16> loc(#loc211) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc211) + } -> tensor<1x960x1x1xbf16> loc(#loc211) + %357 = xten_nn.subgraph (%arg5 = %356: tensor<1x960x1x1xbf16>, %arg6 = %44: tensor<240x960x1x1xbf16>, %arg7 = %43: tensor<240xbf16>) attributes { + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "745", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<240x960x1x1xbf16>, %arg10 = %arg7: tensor<240xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_318", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "746", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "745", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[240, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_319", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<240x960x1x1xbf16>) -> tensor<240x1x1x960xbf16> loc(#loc355) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc355) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_318", + PartOfOutputName = "Conv_318", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<240x1x1x960xbf16>, tensor<240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc212) + %465 = tosa.clamp %464 { + LayerName = "Relu_319", + OutputName = "Relu_319", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x1x240xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc213) + %466 = tosa.reshape %465 {new_shape = array} : (tensor<1x1x1x240xbf16>) -> tensor<1x240x1x1xbf16> loc(#loc355) + xten_nn.output %466 : tensor<1x240x1x1xbf16> loc(#loc213) + } -> tensor<1x240x1x1xbf16> loc(#loc355) + xten_nn.output %461 : tensor<1x240x1x1xbf16> loc(#loc355) + } -> tensor<1x240x1x1xbf16> loc(#loc355) + %358 = xten_nn.subgraph (%arg5 = %357: tensor<1x240x1x1xbf16>, %arg6 = %42: tensor<960x240x1x1xbf16>, %arg7 = %41: tensor<960xbf16>) attributes { + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "747", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc2.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x240x1x1xbf16>, %arg9 = %arg6: tensor<960x240x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_320", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "748", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "747", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 240, 1, 1]> : vector<4xindex> + }, + { + Name = "backbone.features.15.block.2.fc2.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_320", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x240x1x1xbf16>) -> tensor<960x1x1x240xbf16> loc(#loc214) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x240x1x1xbf16>) -> tensor<1x1x1x240xbf16> loc(#loc214) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_320", + PartOfOutputName = "Conv_320", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x240xbf16>, tensor<960x1x1x240xbf16>, tensor<960xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc214) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x960xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %465 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc214) + } -> tensor<1x960x1x1xbf16> loc(#loc214) + %359 = xten_nn.subgraph (%arg5 = %358: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Add_322", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "749", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Add_322", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_322", OutputName = "Add_322"} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc215) + } -> tensor<1x960x1x1xbf16> loc(#loc215) + %360 = xten_nn.subgraph (%arg5 = %359: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Clip_325", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "751", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Clip_325", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_325", + OutputName = "Clip_325", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc216) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc216) + } -> tensor<1x960x1x1xbf16> loc(#loc216) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc216) + } -> tensor<1x960x1x1xbf16> loc(#loc216) + %361 = xten_nn.subgraph (%arg5 = %360: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Div_327", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Div_327", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_327", + OutputName = "Div_327", + shift = 0 : i8} : (tensor<1x960x1x1xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc217) + xten_nn.output %463 : tensor<1x960x1x1xbf16> loc(#loc217) + } -> tensor<1x960x1x1xbf16> loc(#loc217) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc217) + } -> tensor<1x960x1x1xbf16> loc(#loc217) + %362 = xten_nn.subgraph (%arg5 = %361: tensor<1x960x1x1xbf16>) attributes { + LayerName = "Mul_328_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_328_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 960 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc218) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc218) + } -> tensor<1x960x12x20xbf16> loc(#loc218) + %363 = xten_nn.subgraph (%arg5 = %362: tensor<1x960x12x20xbf16>, %arg6 = %354: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_328_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "756", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "754", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_328_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_328", + OutputName = "Mul_328", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc218) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc218) + } -> tensor<1x960x12x20xbf16> loc(#loc218) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc218) + } -> tensor<1x960x12x20xbf16> loc(#loc218) + %364 = xten_nn.subgraph (%arg5 = %363: tensor<1x960x12x20xbf16>, %arg6 = %40: tensor<160x960x1x1xbf16>, %arg7 = %39: tensor<160xbf16>, %arg8 = %344: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Conv_329", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "756", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "757", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<160x960x1x1xbf16>, %arg11 = %arg7: tensor<160xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_329", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "756", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[160, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "757", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_329", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1060", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc219) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<160x960x1x1xbf16>) -> tensor<160x1x1x960xbf16> loc(#loc219) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc219) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_329", + PartOfOutputName = "Conv_329", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<160x1x1x960xbf16>, tensor<160xbf16>) -> tensor<1x12x20x160xbf16> loc(#loc219) + %468 = tosa.transpose %467, %463 : (tensor<1x12x20x160xbf16>, tensor<4xi32>) -> tensor<1x160x12x20xbf16> loc(#loc219) + xten_nn.output %468 : tensor<1x160x12x20xbf16> loc(#loc219) + } -> tensor<1x160x12x20xbf16> loc(#loc219) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x160x12x20xbf16>, %arg10 = %arg8: tensor<1x160x12x20xbf16>) attributes { + LayerName = "Add_330", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1060", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "757", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_330", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.add %arg9, %arg10 {LayerName = "Add_330", OutputName = "Add_330"} : (tensor<1x160x12x20xbf16>, tensor<1x160x12x20xbf16>) -> tensor<1x160x12x20xbf16> loc(#loc220) + xten_nn.output %463 : tensor<1x160x12x20xbf16> loc(#loc220) + } -> tensor<1x160x12x20xbf16> loc(#loc220) + xten_nn.output %462 : tensor<1x160x12x20xbf16> loc(#loc220) + } -> tensor<1x160x12x20xbf16> loc(#loc356) + %365 = xten_nn.subgraph (%arg5 = %364: tensor<1x160x12x20xbf16>, %arg6 = %38: tensor<960x160x1x1xbf16>, %arg7 = %37: tensor<960xbf16>) attributes { + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1060", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "760", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x160x12x20xbf16>, %arg9 = %arg6: tensor<960x160x1x1xbf16>, %arg10 = %arg7: tensor<960xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_331", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 160, 12, 20]> : vector<4xindex> + }, + { + Name = "1060", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[960, 160, 1, 1]> : vector<4xindex> + }, + { + Name = "760", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_331", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc221) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<960x160x1x1xbf16>) -> tensor<960x1x1x160xbf16> loc(#loc221) + %465 = tosa.transpose %arg8, %463 : (tensor<1x160x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x160xbf16> loc(#loc221) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_331", + PartOfOutputName = "Conv_331", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x160xbf16>, tensor<960x1x1x160xbf16>, tensor<960xbf16>) -> tensor<1x12x20x960xbf16> loc(#loc221) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x960xbf16>, tensor<4xi32>) -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %467 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc221) + } -> tensor<1x960x12x20xbf16> loc(#loc221) + %366 = xten_nn.subgraph (%arg5 = %365: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Add_333", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_333", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 3.000000e+00 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<3.000000e+00> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.add %arg6, %462 {LayerName = "Add_333", OutputName = "Add_333"} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc222) + } -> tensor<1x960x12x20xbf16> loc(#loc222) + %367 = xten_nn.subgraph (%arg5 = %366: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Clip_336", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "764", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Clip_336", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 6.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_336", + OutputName = "Clip_336", + max_fp = 6.000000e+00 : f32, + max_int = 6 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc223) + } -> tensor<1x960x12x20xbf16> loc(#loc223) + %368 = xten_nn.subgraph (%arg5 = %367: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "769", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Div_338", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "767", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Div_338", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "769", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulAttributeBroadcastingBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.num_kernel_iters = 0 : ui16, + config.scalar = 1.660160e-01 : bf16, + config.scalar_position = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<1.660160e-01> : tensor<1x1x1x1xbf16>}> : () -> tensor<1x1x1x1xbf16> loc(#loc) + %463 = tosa.mul %arg6, %462 { + LayerName = "Div_338", + OutputName = "Div_338", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x1x1x1xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc224) + xten_nn.output %463 : tensor<1x960x12x20xbf16> loc(#loc224) + } -> tensor<1x960x12x20xbf16> loc(#loc224) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc224) + } -> tensor<1x960x12x20xbf16> loc(#loc224) + %369 = xten_nn.subgraph (%arg5 = %365: tensor<1x960x12x20xbf16>, %arg6 = %368: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x960x12x20xbf16>, %arg8 = %arg6: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "760", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_339", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_339", + OutputName = "Mul_339", + shift = 0 : i8} : (tensor<1x960x12x20xbf16>, tensor<1x960x12x20xbf16>) -> tensor<1x960x12x20xbf16> loc(#loc225) + xten_nn.output %462 : tensor<1x960x12x20xbf16> loc(#loc225) + } -> tensor<1x960x12x20xbf16> loc(#loc225) + xten_nn.output %461 : tensor<1x960x12x20xbf16> loc(#loc225) + } -> tensor<1x960x12x20xbf16> loc(#loc225) + %370 = xten_nn.subgraph (%arg5 = %369: tensor<1x960x12x20xbf16>) attributes { + LayerName = "Mul_339_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1063", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_342_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + Specializes = "Transpose4dAdf", + With = { + config.aie_arch = "aie2p", + config.dim_0 = 12 : ui32, + config.dim_1 = 120 : ui32, + config.dim_2 = 20 : ui32, + config.dim_3 = 8 : ui32, + config.dtype = "bfloat16", + config.perm = 6 : ui32 + }} { + %461 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x960x12x20xbf16>) -> tensor<1x960x1x240xbf16> loc(#loc357) + xten_nn.output %461 : tensor<1x960x1x240xbf16> loc(#loc357) + } -> tensor<1x960x1x240xbf16> loc(#loc357) + %371 = xten_nn.subgraph (%arg5 = %370: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_342", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_342_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x960x1x240xbf16>) attributes { + LayerName = "GlobalAveragePool_342", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 240]> : vector<4xindex> + } + ], + OutputName = "GlobalAveragePool_342_Duplicated#1", + PadValue = 0.000000e+00 : bf16, + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + } + ], + Specializes = "ReduceMeanC8Bf16", + Traits = { + Reduce = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.full_channel = 960 : ui32, + config.full_height = 1 : ui32, + config.full_width = 240 : ui32, + config.reduce_dim = "W" + }} { + %462 = xten_nn.reduce_mean %arg6 {axes = array, keepdims = 1 : i64} : (tensor<1x960x1x240xbf16>) -> tensor<1x960x1x1xbf16> loc(#loc226) + xten_nn.output %462 : tensor<1x960x1x1xbf16> loc(#loc226) + } -> tensor<1x960x1x1xbf16> loc(#loc226) + xten_nn.output %461 : tensor<1x960x1x1xbf16> loc(#loc226) + } -> tensor<1x960x1x1xbf16> loc(#loc226) + %372 = xten_nn.subgraph (%arg5 = %371: tensor<1x960x1x1xbf16>, %arg6 = %36: tensor<128x960x1x1xbf16>, %arg7 = %35: tensor<128xbf16>) attributes { + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "aspp.aspp2.1.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x960x1x1xbf16>, %arg9 = %arg6: tensor<128x960x1x1xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_343", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "774", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "aspp.aspp2.1.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_343", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = tosa.reshape %arg9 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc227) + %463 = tosa.reshape %arg8 {new_shape = array} : (tensor<1x960x1x1xbf16>) -> tensor<1x1x1x960xbf16> loc(#loc227) + %464 = tosa.conv2d %463, %462, %arg10 { + PartOfLayerName = "Conv_343", + PartOfOutputName = "Conv_343", + dilation = array, + pad = array, + stride = array} : (tensor<1x1x1x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x1x1x128xbf16> loc(#loc227) + %465 = tosa.reshape %464 {new_shape = array} : (tensor<1x1x1x128xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc227) + xten_nn.output %465 : tensor<1x128x1x1xbf16> loc(#loc227) + } -> tensor<1x128x1x1xbf16> loc(#loc227) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc227) + } -> tensor<1x128x1x1xbf16> loc(#loc227) + %373 = xten_nn.subgraph (%arg5 = %372: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "776", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Sigmoid_344", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "775", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_344", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "776", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_344", OutputName = "Sigmoid_344"} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x1x1xbf16> loc(#loc228) + xten_nn.output %462 : tensor<1x128x1x1xbf16> loc(#loc228) + } -> tensor<1x128x1x1xbf16> loc(#loc228) + xten_nn.output %461 : tensor<1x128x1x1xbf16> loc(#loc228) + } -> tensor<1x128x1x1xbf16> loc(#loc228) + %374 = xten_nn.subgraph (%arg5 = %373: tensor<1x128x1x1xbf16>) attributes { + LayerName = "Mul_345_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "773", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 1, 1]> : vector<4xindex> + } + ], + OutputName = "Mul_345_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TileAdf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.i_dim_c = 128 : ui32, + config.i_dim_h = 1 : ui32, + config.i_dim_n = 1 : ui32, + config.i_dim_w = 1 : ui32, + config.rep_dim_c = 1 : ui32, + config.rep_dim_h = 12 : ui32, + config.rep_dim_w = 20 : ui32 + }} { + %461 = tosa.tile %arg5 {multiples = array} : (tensor<1x128x1x1xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc229) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc229) + %375 = xten_nn.subgraph (%arg5 = %369: tensor<1x960x12x20xbf16>, %arg6 = %34: tensor<128x960x1x1xbf16>, %arg7 = %33: tensor<128xbf16>, %arg8 = %374: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Conv_340", + OfmShare = 3 : index, + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "1063", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + UnknownDataFormat = true + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1066", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg9 = %arg5: tensor<1x960x12x20xbf16>, %arg10 = %arg6: tensor<128x960x1x1xbf16>, %arg11 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_340", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "770", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 960, 12, 20]> : vector<4xindex> + }, + { + Name = "1063", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 960, 1, 1]> : vector<4xindex> + }, + { + Name = "770", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_341", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "773", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %463 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc359) + %465 = tosa.reshape %arg10 {new_shape = array} : (tensor<128x960x1x1xbf16>) -> tensor<128x1x1x960xbf16> loc(#loc359) + %466 = tosa.transpose %arg9, %464 : (tensor<1x960x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x960xbf16> loc(#loc359) + %467 = tosa.conv2d %466, %465, %arg11 { + PartOfLayerName = "Conv_340", + PartOfOutputName = "Conv_340", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x960xbf16>, tensor<128x1x1x960xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc230) + %468 = tosa.clamp %467 { + LayerName = "Relu_341", + OutputName = "Relu_341", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x12x20x128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc231) + %469 = tosa.transpose %468, %463 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc359) + xten_nn.output %469 : tensor<1x128x12x20xbf16> loc(#loc231) + } -> tensor<1x128x12x20xbf16> loc(#loc359) + %462 = xten_nn.subgraph (%arg9 = %461: tensor<1x128x12x20xbf16>, %arg10 = %arg8: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Mul_345_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "773", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "1066", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_345_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %463 = tosa.mul %arg9, %arg10 { + LayerName = "Mul_345", + OutputName = "Mul_345", + shift = 0 : i8} : (tensor<1x128x12x20xbf16>, tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc229) + xten_nn.output %463 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc229) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc229) + } -> tensor<1x128x12x20xbf16> loc(#loc358) + %376 = xten_nn.subgraph (%arg5 = %375: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "781", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc232) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc232) + } -> tensor<1x64x12x20xbf16> loc(#loc232) + %377 = xten_nn.subgraph (%arg5 = %375: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_349_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_349_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "781", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_349", + PartOfOutputName = "Split_349", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc232) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc232) + } -> tensor<1x64x12x20xbf16> loc(#loc232) + %378 = xten_nn.subgraph (%arg5 = %377: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_350", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_350", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "783", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_350", + OutputName = "Concat_350", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc233) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc233) + } -> tensor<1x128x12x20xbf16> loc(#loc233) + %379 = xten_nn.subgraph (%arg5 = %378: tensor<1x128x12x20xbf16>, %arg6 = %32: tensor<128x128x3x3xbf16>, %arg7 = %31: tensor<128xbf16>) attributes { + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "783", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "397", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<128x128x3x3xbf16>, %arg10 = %arg7: tensor<128xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_351", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "783", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "397", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[128, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_351", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<128x128x3x3xbf16>, tensor<4xi32>) -> tensor<128x3x3x128xbf16> loc(#loc234) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc234) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_351", + PartOfOutputName = "Conv_351", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<128x3x3x128xbf16>, tensor<128xbf16>) -> tensor<1x12x20x128xbf16> loc(#loc234) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x128xbf16>, tensor<4xi32>) -> tensor<1x128x12x20xbf16> loc(#loc234) + xten_nn.output %467 : tensor<1x128x12x20xbf16> loc(#loc234) + } -> tensor<1x128x12x20xbf16> loc(#loc234) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc234) + } -> tensor<1x128x12x20xbf16> loc(#loc234) + %380 = xten_nn.subgraph (%arg5 = %379: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Sigmoid_352", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "784", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_352", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_352", OutputName = "Sigmoid_352"} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc235) + xten_nn.output %462 : tensor<1x128x12x20xbf16> loc(#loc235) + } -> tensor<1x128x12x20xbf16> loc(#loc235) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc235) + } -> tensor<1x128x12x20xbf16> loc(#loc235) + %381 = xten_nn.subgraph (%arg5 = %380: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 64 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc236) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc236) + } -> tensor<1x64x12x20xbf16> loc(#loc236) + %382 = xten_nn.subgraph (%arg5 = %380: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Split_353_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Split_353_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 128 : ui32, + config.dim_h = 12 : ui32, + config.dim_w = 20 : ui32, + config.dtype = "bfloat16", + config.end = 128 : ui32, + config.start = 64 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_353", + PartOfOutputName = "Split_353", + size = array, + start = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc236) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc236) + } -> tensor<1x64x12x20xbf16> loc(#loc236) + %383 = xten_nn.subgraph (%arg5 = %30: tensor<1x64x12x20xbf16>, %arg6 = %382: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Sub_359", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Sub_359", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_359", OutputName = "Sub_359"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc5) + } -> tensor<1x64x12x20xbf16> loc(#loc5) + %384 = xten_nn.subgraph (%arg5 = %383: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_360", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_360", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_360", + OutputName = "Mul_360", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc237) + } -> tensor<1x64x12x20xbf16> loc(#loc237) + %385 = xten_nn.subgraph (%arg5 = %381: tensor<1x64x12x20xbf16>, %arg6 = %arg4: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "788", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_354", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "786", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "785", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_354", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "788", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_354", + OutputName = "Mul_354", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc238) + } -> tensor<1x64x12x20xbf16> loc(#loc238) + %386 = xten_nn.subgraph (%arg5 = %377: tensor<1x64x12x20xbf16>, %arg6 = %385: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_355", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "782", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "788", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_355", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "789", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_355", + OutputName = "Concat_355", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc239) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc239) + } -> tensor<1x128x12x20xbf16> loc(#loc239) + %387 = xten_nn.subgraph (%arg5 = %386: tensor<1x128x12x20xbf16>, %arg6 = %29: tensor<64x128x3x3xbf16>, %arg7 = %28: tensor<64xbf16>) attributes { + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "789", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "788", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x128x12x20xbf16>, %arg9 = %arg6: tensor<64x128x3x3xbf16>, %arg10 = %arg7: tensor<64xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_356", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "789", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + }, + { + Name = "788", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[64, 128, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode4.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_356", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<64x128x3x3xbf16>, tensor<4xi32>) -> tensor<64x3x3x128xbf16> loc(#loc240) + %465 = tosa.transpose %arg8, %463 : (tensor<1x128x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x128xbf16> loc(#loc240) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_356", + PartOfOutputName = "Conv_356", + dilation = array, + pad = array, + stride = array} : (tensor<1x12x20x128xbf16>, tensor<64x3x3x128xbf16>, tensor<64xbf16>) -> tensor<1x12x20x64xbf16> loc(#loc240) + %467 = tosa.transpose %466, %462 : (tensor<1x12x20x64xbf16>, tensor<4xi32>) -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %467 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc240) + } -> tensor<1x64x12x20xbf16> loc(#loc240) + %388 = xten_nn.subgraph (%arg5 = %387: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Tanh_357", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "790", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Tanh_357", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_357", OutputName = "Tanh_357"} : (tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc241) + } -> tensor<1x64x12x20xbf16> loc(#loc241) + %389 = xten_nn.subgraph (%arg5 = %382: tensor<1x64x12x20xbf16>, %arg6 = %388: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "795", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Mul_361", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "787", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "791", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Mul_361", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "795", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_361", + OutputName = "Mul_361", + shift = 0 : i8} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc242) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc242) + } -> tensor<1x64x12x20xbf16> loc(#loc242) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc242) + } -> tensor<1x64x12x20xbf16> loc(#loc242) + %390 = xten_nn.subgraph (%arg5 = %384: tensor<1x64x12x20xbf16>, %arg6 = %389: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "794", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "796", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x64x12x20xbf16>, %arg8 = %arg6: tensor<1x64x12x20xbf16>) attributes { + LayerName = "Add_362", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "794", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "793", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Add_362", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "796", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_362", OutputName = "Add_362"} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x64x12x20xbf16> loc(#loc243) + xten_nn.output %462 : tensor<1x64x12x20xbf16> loc(#loc243) + } -> tensor<1x64x12x20xbf16> loc(#loc243) + xten_nn.output %461 : tensor<1x64x12x20xbf16> loc(#loc243) + } -> tensor<1x64x12x20xbf16> loc(#loc243) + %391 = xten_nn.subgraph (%arg5 = %376: tensor<1x64x12x20xbf16>, %arg6 = %390: tensor<1x64x12x20xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_363", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "781", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "777", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 64, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Concat_363", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "797", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_363", + OutputName = "Concat_363", + axis = 1 : i32} : (tensor<1x64x12x20xbf16>, tensor<1x64x12x20xbf16>) -> tensor<1x128x12x20xbf16> loc(#loc244) + xten_nn.output %461 : tensor<1x128x12x20xbf16> loc(#loc244) + } -> tensor<1x128x12x20xbf16> loc(#loc244) + %392 = xten_nn.subgraph (%arg5 = %391: tensor<1x128x12x20xbf16>) attributes { + LayerName = "Resize_365", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "797", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 12, 20]> : vector<4xindex> + } + ], + OutputName = "Resize_365", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "802", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 128 : ui32, + config.dim_2 = 12 : ui32, + config.dim_3 = 20 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 24 : ui32, + config.output_W = 40 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_365", + OutputName = "Resize_365", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x128x12x20xbf16>) -> tensor<1x128x24x40xbf16> loc(#loc245) + xten_nn.output %461 : tensor<1x128x24x40xbf16> loc(#loc245) + } -> tensor<1x128x24x40xbf16> loc(#loc245) + %393 = xten_nn.subgraph (%arg5 = %392: tensor<1x128x24x40xbf16>) attributes { + LayerName = "Slice_371", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "802", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 24, 40]> : vector<4xindex> + } + ], + OutputName = "Slice_371", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "812", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 128 : ui32, + config.dim_h = 24 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 23 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_371", + OutputName = "Slice_371", + size = array, + start = array} : (tensor<1x128x24x40xbf16>) -> tensor<1x128x23x40xbf16> loc(#loc246) + xten_nn.output %461 : tensor<1x128x23x40xbf16> loc(#loc246) + } -> tensor<1x128x23x40xbf16> loc(#loc246) + %394 = xten_nn.subgraph (%arg5 = %393: tensor<1x128x23x40xbf16>, %arg6 = %220: tensor<1x40x23x40xbf16>, %arg7 = %169: tensor<1x3x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_372", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "812", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 128, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "802", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "496", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_372", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "813", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_372", + OutputName = "Concat_372", + axis = 1 : i32} : (tensor<1x128x23x40xbf16>, tensor<1x40x23x40xbf16>, tensor<1x3x23x40xbf16>) -> tensor<1x171x23x40xbf16> loc(#loc247) + xten_nn.output %461 : tensor<1x171x23x40xbf16> loc(#loc247) + } -> tensor<1x171x23x40xbf16> loc(#loc247) + %395 = xten_nn.subgraph (%arg5 = %394: tensor<1x171x23x40xbf16>, %arg6 = %27: tensor<80x171x3x3xbf16>, %arg7 = %26: tensor<80xbf16>) attributes { + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "813", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Name = "812", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Name = "813", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x171x23x40xbf16>, %arg9 = %arg6: tensor<80x171x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_373", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "813", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 171, 23, 40]> : vector<4xindex> + }, + { + Name = "812", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[80, 171, 3, 3]> : vector<4xindex> + }, + { + Name = "813", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_374", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x171x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x171xbf16> loc(#loc360) + %465 = tosa.transpose %arg8, %463 : (tensor<1x171x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x171xbf16> loc(#loc360) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_373", + PartOfOutputName = "Conv_373", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x171xbf16>, tensor<80x3x3x171xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc248) + %467 = tosa.clamp %466 { + LayerName = "Relu_374", + OutputName = "Relu_374", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x23x40x80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc249) + %468 = tosa.transpose %467, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc360) + xten_nn.output %468 : tensor<1x80x23x40xbf16> loc(#loc249) + } -> tensor<1x80x23x40xbf16> loc(#loc360) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc360) + } -> tensor<1x80x23x40xbf16> loc(#loc360) + %396 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "817", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %397 = xten_nn.subgraph (%arg5 = %395: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_375_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_375_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "817", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_375", + PartOfOutputName = "Split_375", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc250) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc250) + } -> tensor<1x40x23x40xbf16> loc(#loc250) + %398 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_376", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_376", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "819", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_376", + OutputName = "Concat_376", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc251) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc251) + } -> tensor<1x80x23x40xbf16> loc(#loc251) + %399 = xten_nn.subgraph (%arg5 = %398: tensor<1x80x23x40xbf16>, %arg6 = %25: tensor<80x80x3x3xbf16>, %arg7 = %24: tensor<80xbf16>) attributes { + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "819", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "396", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<80x80x3x3xbf16>, %arg10 = %arg7: tensor<80xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_377", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "819", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "396", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[80, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_377", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<80x80x3x3xbf16>, tensor<4xi32>) -> tensor<80x3x3x80xbf16> loc(#loc252) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_377", + PartOfOutputName = "Conv_377", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<80x3x3x80xbf16>, tensor<80xbf16>) -> tensor<1x23x40x80xbf16> loc(#loc252) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x80xbf16>, tensor<4xi32>) -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %467 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc252) + } -> tensor<1x80x23x40xbf16> loc(#loc252) + %400 = xten_nn.subgraph (%arg5 = %399: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Sigmoid_378", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "820", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_378", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_378", OutputName = "Sigmoid_378"} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %462 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc253) + } -> tensor<1x80x23x40xbf16> loc(#loc253) + %401 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %402 = xten_nn.subgraph (%arg5 = %400: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Split_379_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Split_379_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 80 : ui32, + config.dim_h = 23 : ui32, + config.dim_w = 40 : ui32, + config.dtype = "bfloat16", + config.end = 80 : ui32, + config.start = 40 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_379", + PartOfOutputName = "Split_379", + size = array, + start = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc254) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc254) + } -> tensor<1x40x23x40xbf16> loc(#loc254) + %403 = xten_nn.subgraph (%arg5 = %23: tensor<1x40x23x40xbf16>, %arg6 = %402: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Sub_385", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Sub_385", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_385", OutputName = "Sub_385"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc4) + } -> tensor<1x40x23x40xbf16> loc(#loc4) + %404 = xten_nn.subgraph (%arg5 = %403: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_386", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_386", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_386", + OutputName = "Mul_386", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc255) + } -> tensor<1x40x23x40xbf16> loc(#loc255) + %405 = xten_nn.subgraph (%arg5 = %401: tensor<1x40x23x40xbf16>, %arg6 = %arg3: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "824", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_380", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "822", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "821", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_380", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "824", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_380", + OutputName = "Mul_380", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc256) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc256) + } -> tensor<1x40x23x40xbf16> loc(#loc256) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc256) + } -> tensor<1x40x23x40xbf16> loc(#loc256) + %406 = xten_nn.subgraph (%arg5 = %397: tensor<1x40x23x40xbf16>, %arg6 = %405: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_381", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "818", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "824", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_381", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "825", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_381", + OutputName = "Concat_381", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc257) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc257) + } -> tensor<1x80x23x40xbf16> loc(#loc257) + %407 = xten_nn.subgraph (%arg5 = %406: tensor<1x80x23x40xbf16>, %arg6 = %22: tensor<40x80x3x3xbf16>, %arg7 = %21: tensor<40xbf16>) attributes { + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "825", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "824", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x80x23x40xbf16>, %arg9 = %arg6: tensor<40x80x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_382", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "825", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + }, + { + Name = "824", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 80, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode3.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_382", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x80x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x80xbf16> loc(#loc258) + %465 = tosa.transpose %arg8, %463 : (tensor<1x80x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x80xbf16> loc(#loc258) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_382", + PartOfOutputName = "Conv_382", + dilation = array, + pad = array, + stride = array} : (tensor<1x23x40x80xbf16>, tensor<40x3x3x80xbf16>, tensor<40xbf16>) -> tensor<1x23x40x40xbf16> loc(#loc258) + %467 = tosa.transpose %466, %462 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %467 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc258) + } -> tensor<1x40x23x40xbf16> loc(#loc258) + %408 = xten_nn.subgraph (%arg5 = %407: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Tanh_383", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "826", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Tanh_383", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_383", OutputName = "Tanh_383"} : (tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc259) + } -> tensor<1x40x23x40xbf16> loc(#loc259) + %409 = xten_nn.subgraph (%arg5 = %402: tensor<1x40x23x40xbf16>, %arg6 = %408: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "831", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Mul_387", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "823", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "827", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Mul_387", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "831", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_387", + OutputName = "Mul_387", + shift = 0 : i8} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc260) + } -> tensor<1x40x23x40xbf16> loc(#loc260) + %410 = xten_nn.subgraph (%arg5 = %404: tensor<1x40x23x40xbf16>, %arg6 = %409: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "830", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "832", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x40x23x40xbf16>, %arg8 = %arg6: tensor<1x40x23x40xbf16>) attributes { + LayerName = "Add_388", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "830", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "829", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Add_388", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "832", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_388", OutputName = "Add_388"} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %462 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + xten_nn.output %461 : tensor<1x40x23x40xbf16> loc(#loc261) + } -> tensor<1x40x23x40xbf16> loc(#loc261) + %411 = xten_nn.subgraph (%arg5 = %396: tensor<1x40x23x40xbf16>, %arg6 = %410: tensor<1x40x23x40xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_389", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "817", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "816", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Concat_389", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "833", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_389", + OutputName = "Concat_389", + axis = 1 : i32} : (tensor<1x40x23x40xbf16>, tensor<1x40x23x40xbf16>) -> tensor<1x80x23x40xbf16> loc(#loc262) + xten_nn.output %461 : tensor<1x80x23x40xbf16> loc(#loc262) + } -> tensor<1x80x23x40xbf16> loc(#loc262) + %412 = xten_nn.subgraph (%arg5 = %411: tensor<1x80x23x40xbf16>) attributes { + LayerName = "Resize_391", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "833", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 23, 40]> : vector<4xindex> + } + ], + OutputName = "Resize_391", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "838", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 80 : ui32, + config.dim_2 = 23 : ui32, + config.dim_3 = 40 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 46 : ui32, + config.output_W = 80 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_391", + OutputName = "Resize_391", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x80x23x40xbf16>) -> tensor<1x80x46x80xbf16> loc(#loc263) + xten_nn.output %461 : tensor<1x80x46x80xbf16> loc(#loc263) + } -> tensor<1x80x46x80xbf16> loc(#loc263) + %413 = xten_nn.subgraph (%arg5 = %412: tensor<1x80x46x80xbf16>) attributes { + LayerName = "Slice_397", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "838", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 46, 80]> : vector<4xindex> + } + ], + OutputName = "Slice_397", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "848", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "H", + config.dim_c = 80 : ui32, + config.dim_h = 46 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 45 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + LayerName = "Slice_397", + OutputName = "Slice_397", + size = array, + start = array} : (tensor<1x80x46x80xbf16>) -> tensor<1x80x45x80xbf16> loc(#loc264) + xten_nn.output %461 : tensor<1x80x45x80xbf16> loc(#loc264) + } -> tensor<1x80x45x80xbf16> loc(#loc264) + %414 = xten_nn.subgraph (%arg5 = %413: tensor<1x80x45x80xbf16>, %arg6 = %184: tensor<1x24x45x80xbf16>, %arg7 = %168: tensor<1x3x45x80xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_398", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "848", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 80, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "838", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 24, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "434", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_398", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "849", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_398", + OutputName = "Concat_398", + axis = 1 : i32} : (tensor<1x80x45x80xbf16>, tensor<1x24x45x80xbf16>, tensor<1x3x45x80xbf16>) -> tensor<1x107x45x80xbf16> loc(#loc265) + xten_nn.output %461 : tensor<1x107x45x80xbf16> loc(#loc265) + } -> tensor<1x107x45x80xbf16> loc(#loc265) + %415 = xten_nn.subgraph (%arg5 = %414: tensor<1x107x45x80xbf16>, %arg6 = %20: tensor<40x107x3x3xbf16>, %arg7 = %19: tensor<40xbf16>) attributes { + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "849", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Name = "848", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Name = "849", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x107x45x80xbf16>, %arg9 = %arg6: tensor<40x107x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_399", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "849", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 107, 45, 80]> : vector<4xindex> + }, + { + Name = "848", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[40, 107, 3, 3]> : vector<4xindex> + }, + { + Name = "849", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_400", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x107x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x107xbf16> loc(#loc361) + %465 = tosa.transpose %arg8, %463 : (tensor<1x107x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x107xbf16> loc(#loc361) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_399", + PartOfOutputName = "Conv_399", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x107xbf16>, tensor<40x3x3x107xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc266) + %467 = tosa.clamp %466 { + LayerName = "Relu_400", + OutputName = "Relu_400", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x45x80x40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc267) + %468 = tosa.transpose %467, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc361) + xten_nn.output %468 : tensor<1x40x45x80xbf16> loc(#loc267) + } -> tensor<1x40x45x80xbf16> loc(#loc361) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc361) + } -> tensor<1x40x45x80xbf16> loc(#loc361) + %416 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "853", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %417 = xten_nn.subgraph (%arg5 = %415: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_401_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_401_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "853", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_401", + PartOfOutputName = "Split_401", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc268) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc268) + } -> tensor<1x20x45x80xbf16> loc(#loc268) + %418 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_402", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_402", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "855", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_402", + OutputName = "Concat_402", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc269) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc269) + } -> tensor<1x40x45x80xbf16> loc(#loc269) + %419 = xten_nn.subgraph (%arg5 = %418: tensor<1x40x45x80xbf16>, %arg6 = %18: tensor<40x40x3x3xbf16>, %arg7 = %17: tensor<40xbf16>) attributes { + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "855", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "395", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<40x40x3x3xbf16>, %arg10 = %arg7: tensor<40xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_403", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "855", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "395", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[40, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_403", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<40x40x3x3xbf16>, tensor<4xi32>) -> tensor<40x3x3x40xbf16> loc(#loc270) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_403", + PartOfOutputName = "Conv_403", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<40x3x3x40xbf16>, tensor<40xbf16>) -> tensor<1x45x80x40xbf16> loc(#loc270) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x40xbf16>, tensor<4xi32>) -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %467 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc270) + } -> tensor<1x40x45x80xbf16> loc(#loc270) + %420 = xten_nn.subgraph (%arg5 = %419: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Sigmoid_404", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "856", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_404", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_404", OutputName = "Sigmoid_404"} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %462 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc271) + } -> tensor<1x40x45x80xbf16> loc(#loc271) + %421 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 20 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %422 = xten_nn.subgraph (%arg5 = %420: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Split_405_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Split_405_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 40 : ui32, + config.dim_h = 45 : ui32, + config.dim_w = 80 : ui32, + config.dtype = "bfloat16", + config.end = 40 : ui32, + config.start = 20 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_405", + PartOfOutputName = "Split_405", + size = array, + start = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc272) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc272) + } -> tensor<1x20x45x80xbf16> loc(#loc272) + %423 = xten_nn.subgraph (%arg5 = %16: tensor<1x20x45x80xbf16>, %arg6 = %422: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Sub_411", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Sub_411", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_411", OutputName = "Sub_411"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc3) + } -> tensor<1x20x45x80xbf16> loc(#loc3) + %424 = xten_nn.subgraph (%arg5 = %423: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_412", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_412", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_412", + OutputName = "Mul_412", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc273) + } -> tensor<1x20x45x80xbf16> loc(#loc273) + %425 = xten_nn.subgraph (%arg5 = %421: tensor<1x20x45x80xbf16>, %arg6 = %arg2: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "860", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_406", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "858", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "857", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_406", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "860", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_406", + OutputName = "Mul_406", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc274) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc274) + } -> tensor<1x20x45x80xbf16> loc(#loc274) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc274) + } -> tensor<1x20x45x80xbf16> loc(#loc274) + %426 = xten_nn.subgraph (%arg5 = %417: tensor<1x20x45x80xbf16>, %arg6 = %425: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_407", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "854", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "860", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_407", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "861", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_407", + OutputName = "Concat_407", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc275) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc275) + } -> tensor<1x40x45x80xbf16> loc(#loc275) + %427 = xten_nn.subgraph (%arg5 = %426: tensor<1x40x45x80xbf16>, %arg6 = %15: tensor<20x40x3x3xbf16>, %arg7 = %14: tensor<20xbf16>) attributes { + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "861", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "860", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x40x45x80xbf16>, %arg9 = %arg6: tensor<20x40x3x3xbf16>, %arg10 = %arg7: tensor<20xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_408", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "861", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + }, + { + Name = "860", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[20, 40, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode2.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_408", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<20x40x3x3xbf16>, tensor<4xi32>) -> tensor<20x3x3x40xbf16> loc(#loc276) + %465 = tosa.transpose %arg8, %463 : (tensor<1x40x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x40xbf16> loc(#loc276) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_408", + PartOfOutputName = "Conv_408", + dilation = array, + pad = array, + stride = array} : (tensor<1x45x80x40xbf16>, tensor<20x3x3x40xbf16>, tensor<20xbf16>) -> tensor<1x45x80x20xbf16> loc(#loc276) + %467 = tosa.transpose %466, %462 : (tensor<1x45x80x20xbf16>, tensor<4xi32>) -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %467 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc276) + } -> tensor<1x20x45x80xbf16> loc(#loc276) + %428 = xten_nn.subgraph (%arg5 = %427: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Tanh_409", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "862", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Tanh_409", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_409", OutputName = "Tanh_409"} : (tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc277) + } -> tensor<1x20x45x80xbf16> loc(#loc277) + %429 = xten_nn.subgraph (%arg5 = %422: tensor<1x20x45x80xbf16>, %arg6 = %428: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "867", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Mul_413", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "859", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "863", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Mul_413", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "867", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_413", + OutputName = "Mul_413", + shift = 0 : i8} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc278) + } -> tensor<1x20x45x80xbf16> loc(#loc278) + %430 = xten_nn.subgraph (%arg5 = %424: tensor<1x20x45x80xbf16>, %arg6 = %429: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "866", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "868", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x20x45x80xbf16>, %arg8 = %arg6: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Add_414", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "866", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "865", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Add_414", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "868", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_414", OutputName = "Add_414"} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %462 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + xten_nn.output %461 : tensor<1x20x45x80xbf16> loc(#loc279) + } -> tensor<1x20x45x80xbf16> loc(#loc279) + %431 = xten_nn.subgraph (%arg5 = %416: tensor<1x20x45x80xbf16>, %arg6 = %430: tensor<1x20x45x80xbf16>) attributes { + LayerName = "Concat_415", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "853", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "852", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 20, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Concat_415", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "869", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + Specializes = "ConcatC8Adf", + With = { + config.aie_arch = "aie2p", + config.dtype = "bfloat16", + config.in1_dim_c = 24 : ui32, + config.in1_dim_h = 45 : ui32, + config.in1_dim_w = 80 : ui32, + config.in2_dim_c = 24 : ui32, + config.in2_dim_h = 45 : ui32, + config.in2_dim_w = 80 : ui32, + config.num_eff_concat_input0_size = 20 : ui32, + config.num_eff_concat_input0_start = 0 : ui32, + config.num_eff_concat_input1_size = 20 : ui32, + config.num_eff_concat_input1_start = 0 : ui32 + }} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_415", + OutputName = "Concat_415", + axis = 1 : i32} : (tensor<1x20x45x80xbf16>, tensor<1x20x45x80xbf16>) -> tensor<1x40x45x80xbf16> loc(#loc280) + xten_nn.output %461 : tensor<1x40x45x80xbf16> loc(#loc280) + } -> tensor<1x40x45x80xbf16> loc(#loc280) + %432 = xten_nn.subgraph (%arg5 = %431: tensor<1x40x45x80xbf16>) attributes { + LayerName = "Resize_417", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "869", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 45, 80]> : vector<4xindex> + } + ], + OutputName = "Resize_417", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "874", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 40 : ui32, + config.dim_2 = 45 : ui32, + config.dim_3 = 80 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 90 : ui32, + config.output_W = 160 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_417", + OutputName = "Resize_417", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x40x45x80xbf16>) -> tensor<1x40x90x160xbf16> loc(#loc281) + xten_nn.output %461 : tensor<1x40x90x160xbf16> loc(#loc281) + } -> tensor<1x40x90x160xbf16> loc(#loc281) + %433 = xten_nn.subgraph (%arg5 = %432: tensor<1x40x90x160xbf16>, %arg6 = %178: tensor<1x16x90x160xbf16>, %arg7 = %167: tensor<1x3x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_418", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "874", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 40, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "869", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "417", + Port = "data_io.ifm3", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_418", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "875", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6, %arg7 { + LayerName = "Concat_418", + OutputName = "Concat_418", + axis = 1 : i32} : (tensor<1x40x90x160xbf16>, tensor<1x16x90x160xbf16>, tensor<1x3x90x160xbf16>) -> tensor<1x59x90x160xbf16> loc(#loc282) + xten_nn.output %461 : tensor<1x59x90x160xbf16> loc(#loc282) + } -> tensor<1x59x90x160xbf16> loc(#loc282) + %434 = xten_nn.subgraph (%arg5 = %433: tensor<1x59x90x160xbf16>, %arg6 = %13: tensor<32x59x3x3xbf16>, %arg7 = %12: tensor<32xbf16>) attributes { + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "875", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Name = "874", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Name = "875", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x59x90x160xbf16>, %arg9 = %arg6: tensor<32x59x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_419", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "875", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 59, 90, 160]> : vector<4xindex> + }, + { + Name = "874", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[32, 59, 3, 3]> : vector<4xindex> + }, + { + Name = "875", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_420", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x59x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x59xbf16> loc(#loc362) + %465 = tosa.transpose %arg8, %463 : (tensor<1x59x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x59xbf16> loc(#loc362) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_419", + PartOfOutputName = "Conv_419", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x59xbf16>, tensor<32x3x3x59xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc283) + %467 = tosa.clamp %466 { + LayerName = "Relu_420", + OutputName = "Relu_420", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x90x160x32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc284) + %468 = tosa.transpose %467, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc362) + xten_nn.output %468 : tensor<1x32x90x160xbf16> loc(#loc284) + } -> tensor<1x32x90x160xbf16> loc(#loc362) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc362) + } -> tensor<1x32x90x160xbf16> loc(#loc362) + %435 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "879", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %436 = xten_nn.subgraph (%arg5 = %434: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_421_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_421_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "879", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_421", + PartOfOutputName = "Split_421", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc285) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc285) + } -> tensor<1x16x90x160xbf16> loc(#loc285) + %437 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_422", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_422", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "881", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_422", + OutputName = "Concat_422", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc286) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc286) + } -> tensor<1x32x90x160xbf16> loc(#loc286) + %438 = xten_nn.subgraph (%arg5 = %437: tensor<1x32x90x160xbf16>, %arg6 = %11: tensor<32x32x3x3xbf16>, %arg7 = %10: tensor<32xbf16>) attributes { + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "881", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "394", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.ih.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<32x32x3x3xbf16>, %arg10 = %arg7: tensor<32xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_423", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "881", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "394", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[32, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.ih.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_423", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<32x32x3x3xbf16>, tensor<4xi32>) -> tensor<32x3x3x32xbf16> loc(#loc287) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_423", + PartOfOutputName = "Conv_423", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<32x3x3x32xbf16>, tensor<32xbf16>) -> tensor<1x90x160x32xbf16> loc(#loc287) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x32xbf16>, tensor<4xi32>) -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %467 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc287) + } -> tensor<1x32x90x160xbf16> loc(#loc287) + %439 = xten_nn.subgraph (%arg5 = %438: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Sigmoid_424", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "882", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sigmoid_424", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SigmoidTemplatedBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.sigmoid %arg6 {LayerName = "Sigmoid_424", OutputName = "Sigmoid_424"} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %462 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc288) + } -> tensor<1x32x90x160xbf16> loc(#loc288) + %440 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 16 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %441 = xten_nn.subgraph (%arg5 = %439: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Split_425_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Split_425_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 32 : ui32, + config.dim_h = 90 : ui32, + config.dim_w = 160 : ui32, + config.dtype = "bfloat16", + config.end = 32 : ui32, + config.start = 16 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_425", + PartOfOutputName = "Split_425", + size = array, + start = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc289) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc289) + } -> tensor<1x16x90x160xbf16> loc(#loc289) + %442 = xten_nn.subgraph (%arg5 = %9: tensor<1x16x90x160xbf16>, %arg6 = %441: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Sub_431", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "890", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Sub_431", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "SubBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.sub %arg7, %arg8 {LayerName = "Sub_431", OutputName = "Sub_431"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc2) + } -> tensor<1x16x90x160xbf16> loc(#loc2) + %443 = xten_nn.subgraph (%arg5 = %442: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_432", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_432", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_432", + OutputName = "Mul_432", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc290) + } -> tensor<1x16x90x160xbf16> loc(#loc290) + %444 = xten_nn.subgraph (%arg5 = %440: tensor<1x16x90x160xbf16>, %arg6 = %arg1: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "886", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_426", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "884", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "883", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_426", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "886", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_426", + OutputName = "Mul_426", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc291) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc291) + } -> tensor<1x16x90x160xbf16> loc(#loc291) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc291) + } -> tensor<1x16x90x160xbf16> loc(#loc291) + %445 = xten_nn.subgraph (%arg5 = %436: tensor<1x16x90x160xbf16>, %arg6 = %444: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_427", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "880", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "886", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_427", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "887", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_427", + OutputName = "Concat_427", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc292) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc292) + } -> tensor<1x32x90x160xbf16> loc(#loc292) + %446 = xten_nn.subgraph (%arg5 = %445: tensor<1x32x90x160xbf16>, %arg6 = %8: tensor<16x32x3x3xbf16>, %arg7 = %7: tensor<16xbf16>) attributes { + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "887", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "886", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.hh.0.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x32x90x160xbf16>, %arg9 = %arg6: tensor<16x32x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_428", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "887", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + }, + { + Name = "886", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 32, 3, 3]> : vector<4xindex> + }, + { + Name = "decoder.decode1.gru.hh.0.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_428", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x32x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x32xbf16> loc(#loc293) + %465 = tosa.transpose %arg8, %463 : (tensor<1x32x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x32xbf16> loc(#loc293) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_428", + PartOfOutputName = "Conv_428", + dilation = array, + pad = array, + stride = array} : (tensor<1x90x160x32xbf16>, tensor<16x3x3x32xbf16>, tensor<16xbf16>) -> tensor<1x90x160x16xbf16> loc(#loc293) + %467 = tosa.transpose %466, %462 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %467 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc293) + } -> tensor<1x16x90x160xbf16> loc(#loc293) + %447 = xten_nn.subgraph (%arg5 = %446: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "single", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Tanh_429", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "888", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Tanh_429", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "TanhTemplatedBf16", + Traits = { + Elementwise = true, + Unary = true + }, + With = { + config.ENABLE_FP16_AS_BF16 = 0 : ui8, + config.aie_arch = "aie2p", + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.tanh %arg6 {LayerName = "Tanh_429", OutputName = "Tanh_429"} : (tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc294) + } -> tensor<1x16x90x160xbf16> loc(#loc294) + %448 = xten_nn.subgraph (%arg5 = %441: tensor<1x16x90x160xbf16>, %arg6 = %447: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "893", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Mul_433", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "885", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "889", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Mul_433", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "893", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "MulBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.mul %arg7, %arg8 { + LayerName = "Mul_433", + OutputName = "Mul_433", + shift = 0 : i8} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc295) + } -> tensor<1x16x90x160xbf16> loc(#loc295) + %449 = xten_nn.subgraph (%arg5 = %443: tensor<1x16x90x160xbf16>, %arg6 = %448: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "892", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "894", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x16x90x160xbf16>, %arg8 = %arg6: tensor<1x16x90x160xbf16>) attributes { + LayerName = "Add_434", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "892", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "891", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Add_434", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "894", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_434", OutputName = "Add_434"} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %462 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + xten_nn.output %461 : tensor<1x16x90x160xbf16> loc(#loc296) + } -> tensor<1x16x90x160xbf16> loc(#loc296) + %450 = xten_nn.subgraph (%arg5 = %435: tensor<1x16x90x160xbf16>, %arg6 = %449: tensor<1x16x90x160xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_435", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "879", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "878", + Port = "data_io.ifm2", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Concat_435", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "895", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_435", + OutputName = "Concat_435", + axis = 1 : i32} : (tensor<1x16x90x160xbf16>, tensor<1x16x90x160xbf16>) -> tensor<1x32x90x160xbf16> loc(#loc297) + xten_nn.output %461 : tensor<1x32x90x160xbf16> loc(#loc297) + } -> tensor<1x32x90x160xbf16> loc(#loc297) + %451 = xten_nn.subgraph (%arg5 = %450: tensor<1x32x90x160xbf16>) attributes { + LayerName = "Resize_437", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "895", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 90, 160]> : vector<4xindex> + } + ], + OutputName = "Resize_437", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "900", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ResizeAdf", + With = { + config.co_trans_mode = 1 : ui32, + config.dim_0 = 1 : ui32, + config.dim_1 = 32 : ui32, + config.dim_2 = 90 : ui32, + config.dim_3 = 160 : ui32, + config.dtype = "bfloat16", + config.mode = 1 : ui32, + config.nearest_mode = 0 : ui32, + config.output_H = 180 : ui32, + config.output_W = 320 : ui32 + }} { + %461 = xten_nn.resize %arg5 { + LayerName = "Resize_437", + OutputName = "Resize_437", + coordinate_transformation_mode = 1 : i64, + mode = 1 : i64, + nearest_mode = 0 : i64, + scales = array} : (tensor<1x32x90x160xbf16>) -> tensor<1x32x180x320xbf16> loc(#loc298) + xten_nn.output %461 : tensor<1x32x180x320xbf16> loc(#loc298) + } -> tensor<1x32x180x320xbf16> loc(#loc298) + %452 = xten_nn.subgraph (%arg5 = %451: tensor<1x32x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + Axis = 1 : i32, + LayerName = "Concat_438", + Op = "Concat", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "900", + Port = "data_io.ifm1", + l3_extend_end = dense<0> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 32, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "895", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Concat_438", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "PseudoOp", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "901", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + } + ], + current_data_format = "NCHW", + data_format = "HCWN"} { + %461 = tosa.concat %arg5, %arg6 { + LayerName = "Concat_438", + OutputName = "Concat_438", + axis = 1 : i32} : (tensor<1x32x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x35x180x320xbf16> loc(#loc299) + xten_nn.output %461 : tensor<1x35x180x320xbf16> loc(#loc299) + } -> tensor<1x35x180x320xbf16> loc(#loc299) + %453 = xten_nn.subgraph (%arg5 = %452: tensor<1x35x180x320xbf16>, %arg6 = %6: tensor<16x35x3x3xbf16>, %arg7 = %5: tensor<16xbf16>) attributes { + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "901", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Name = "900", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Name = "901", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x35x180x320xbf16>, %arg9 = %arg6: tensor<16x35x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_439", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "901", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 35, 180, 320]> : vector<4xindex> + }, + { + Name = "900", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[16, 35, 3, 3]> : vector<4xindex> + }, + { + Name = "901", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_440", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x35x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x35xbf16> loc(#loc363) + %465 = tosa.transpose %arg8, %463 : (tensor<1x35x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x35xbf16> loc(#loc363) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_439", + PartOfOutputName = "Conv_439", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x35xbf16>, tensor<16x3x3x35xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc300) + %467 = tosa.clamp %466 { + LayerName = "Relu_440", + OutputName = "Relu_440", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc301) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc363) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc301) + } -> tensor<1x16x180x320xbf16> loc(#loc363) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc363) + } -> tensor<1x16x180x320xbf16> loc(#loc363) + %454 = xten_nn.subgraph (%arg5 = %453: tensor<1x16x180x320xbf16>, %arg6 = %4: tensor<16x16x3x3xbf16>, %arg7 = %3: tensor<16xbf16>) attributes { + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1078", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Name = "1082", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<16x16x3x3xbf16>, %arg10 = %arg7: tensor<16xbf16>) attributes { + Dilations = array, + HWPadding = [[1, 1], [1, 1]], + LayerName = "Conv_441", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "904", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1078", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[16, 16, 3, 3]> : vector<4xindex> + }, + { + Name = "1082", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Relu_442", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + Port = "data_io.ofm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true, + NonNegativeOut = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 1 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 3 : ui8, + config.ksize.width = 3 : ui8, + config.lrelu_alpha = 0.000000e+00 : bf16, + config.lrelu_alpha_kernel = 0.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %464 = tosa.transpose %arg9, %463 : (tensor<16x16x3x3xbf16>, tensor<4xi32>) -> tensor<16x3x3x16xbf16> loc(#loc364) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc364) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_441", + PartOfOutputName = "Conv_441", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<16x3x3x16xbf16>, tensor<16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc302) + %467 = tosa.clamp %466 { + LayerName = "Relu_442", + OutputName = "Relu_442", + max_fp = 3.40282347E+38 : f32, + max_int = 2147483647 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x180x320x16xbf16>) -> tensor<1x180x320x16xbf16> loc(#loc303) + %468 = tosa.transpose %467, %462 : (tensor<1x180x320x16xbf16>, tensor<4xi32>) -> tensor<1x16x180x320xbf16> loc(#loc364) + xten_nn.output %468 : tensor<1x16x180x320xbf16> loc(#loc303) + } -> tensor<1x16x180x320xbf16> loc(#loc364) + xten_nn.output %461 : tensor<1x16x180x320xbf16> loc(#loc364) + } -> tensor<1x16x180x320xbf16> loc(#loc364) + %455 = xten_nn.subgraph (%arg5 = %454: tensor<1x16x180x320xbf16>, %arg6 = %2: tensor<4x16x1x1xbf16>, %arg7 = %1: tensor<4xbf16>) attributes { + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1081", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "project_mat.conv.weight", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg8 = %arg5: tensor<1x16x180x320xbf16>, %arg9 = %arg6: tensor<4x16x1x1xbf16>, %arg10 = %arg7: tensor<4xbf16>) attributes { + Dilations = array, + HWPadding = [[0, 0], [0, 0]], + LayerName = "Conv_443", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "907", + Port = "data_io.ifm", + l3_extend_end = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 16, 180, 320]> : vector<4xindex> + }, + { + Name = "1081", + Port = "data_io.wts", + SubPort = "wts_data", + UnknownDataFormat = true, + l3_extend_end = dense<[4, 0, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[4, 16, 1, 1]> : vector<4xindex> + }, + { + Name = "project_mat.conv.weight", + Port = "data_io.wts", + SubPort = "bias", + UnknownDataFormat = true + } + ], + OutputName = "Conv_443", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + Specializes = "Conv2DBf16", + Traits = { + AllowDMAOptimization = true + }, + With = { + config.AIE_API_EMULATE_BFLOAT16_MMUL_WITH_BFP16 = 1 : ui8, + config.act = 0 : ui8, + config.act_type = "RELU", + config.aie_arch = "aie2p", + config.batch_size = 1 : ui8, + config.compiler = "chess", + config.conv_type = [0 : ui8, 12 : ui8, 64 : ui8], + config.dtype_ifm = "bfloat16", + config.dtype_ofm = "bfloat16", + config.dtype_wts = "bfloat16", + config.ksize.height = 1 : ui8, + config.ksize.width = 1 : ui8, + config.lrelu_alpha = 1.000000e+00 : bf16, + config.lrelu_alpha_kernel = 1.000000e+00 : bf16, + config.stride_h = 1 : ui8, + config.stride_w = 1 : ui8 + }} { + %462 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc) + %463 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc304) + %464 = tosa.reshape %arg9 {new_shape = array} : (tensor<4x16x1x1xbf16>) -> tensor<4x1x1x16xbf16> loc(#loc304) + %465 = tosa.transpose %arg8, %463 : (tensor<1x16x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x16xbf16> loc(#loc304) + %466 = tosa.conv2d %465, %464, %arg10 { + PartOfLayerName = "Conv_443", + PartOfOutputName = "Conv_443", + dilation = array, + pad = array, + stride = array} : (tensor<1x180x320x16xbf16>, tensor<4x1x1x16xbf16>, tensor<4xbf16>) -> tensor<1x180x320x4xbf16> loc(#loc304) + %467 = tosa.transpose %466, %462 : (tensor<1x180x320x4xbf16>, tensor<4xi32>) -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %467 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + xten_nn.output %461 : tensor<1x4x180x320xbf16> loc(#loc304) + } -> tensor<1x4x180x320xbf16> loc(#loc304) + %456 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#0", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#0", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 3 : ui32, + config.start = 0 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc305) + } -> tensor<1x3x180x320xbf16> loc(#loc305) + %457 = xten_nn.subgraph (%arg5 = %456: tensor<1x3x180x320xbf16>, %arg6 = %166: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445_Duplicated#1", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg7 = %arg5: tensor<1x3x180x320xbf16>, %arg8 = %arg6: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Add_445", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + Port = "data_io.ifm1", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + }, + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ifm2", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Add_445_Duplicated#1", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "AddBf16", + Traits = { + Binary = true, + Elementwise = true + }, + With = { + config.act = 0 : ui8, + config.act_type = "LINEAR", + config.aie_arch = "aie2p", + config.compiler = "chess", + config.dtype = "bfloat16", + config.num_kernel_iters = 0 : ui16 + }} { + %462 = tosa.add %arg7, %arg8 {LayerName = "Add_445", OutputName = "Add_445"} : (tensor<1x3x180x320xbf16>, tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc11) + } -> tensor<1x3x180x320xbf16> loc(#loc11) + %458 = xten_nn.subgraph (%arg5 = %457: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "916", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x3x180x320xbf16>) attributes { + LayerName = "Clip_446", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "911", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_446", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "916", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 5, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 3, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_446", + OutputName = "Clip_446", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x3x180x320xbf16>) -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %462 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + xten_nn.output %461 : tensor<1x3x180x320xbf16> loc(#loc306) + } -> tensor<1x3x180x320xbf16> loc(#loc306) + %459 = xten_nn.subgraph (%arg5 = %455: tensor<1x4x180x320xbf16>) attributes { + LayerName = "Split_444_Duplicated#1", + Operands = [ + { + CurrentDataFormat = "NCHW", + External = false, + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "908", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 4, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 4, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Split_444_Duplicated#1", + Overlay = "1x1_1x1_unspecifiedConnectivity", + Reason = "TemplatedGraph", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "909", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_extend_start = dense<0> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "SliceHCWC8Adf", + With = { + config.aie_arch = "aie2p", + config.axis_letter = "C", + config.dim_c = 8 : ui32, + config.dim_h = 180 : ui32, + config.dim_w = 320 : ui32, + config.dtype = "bfloat16", + config.end = 4 : ui32, + config.start = 3 : ui32, + config.step = 1 : ui32 + }} { + %461 = tosa.slice %arg5 { + PartOfLayerName = "Split_444", + PartOfOutputName = "Split_444", + size = array, + start = array} : (tensor<1x4x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc305) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc305) + } -> tensor<1x1x180x320xbf16> loc(#loc305) + %460 = xten_nn.subgraph (%arg5 = %459: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "910", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Overlay = "4x4_1x4_vertBroadcastLeft_horizBroadcastRight", + Reason = "InCoreChain", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "921", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + memory_configuration = { + L1 = {layout = "strict"}, + L2 = {feature_maps_buffering = "double", layout = "flexible"} + }} { + %461 = xten_nn.subgraph (%arg6 = %arg5: tensor<1x1x180x320xbf16>) attributes { + LayerName = "Clip_447", + Operands = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "910", + Port = "data_io.ifm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + OutputName = "Clip_447", + Reason = "MllibKernel", + Results = [ + { + CurrentDataFormat = "NCHW", + L3DataFormat = "HCWN", + L3Vectorization = "C:8", + Name = "921", + Port = "data_io.ofm", + l3_extend_end = dense<[0, 7, 0, 0]> : vector<4xindex>, + l3_tile_count = dense<[1, 1, 180, 320]> : vector<4xindex> + } + ], + Specializes = "ClipBf16", + Traits = { + Elementwise = true, + NonNegativeOut = true, + Unary = true + }, + With = { + config.aie_arch = "aie2p", + config.clamp_max = 1.000000e+00 : bf16, + config.clamp_min = 0.000000e+00 : bf16, + config.compiler = "chess", + config.ifm_shift = 0 : si8, + config.num_kernel_iters = 0 : ui16, + config.ofm_shift = 0 : si8 + }} { + %462 = tosa.clamp %arg6 { + LayerName = "Clip_447", + OutputName = "Clip_447", + max_fp = 1.000000e+00 : f32, + max_int = 1 : i64, + min_fp = 0.000000e+00 : f32, + min_int = 0 : i64} : (tensor<1x1x180x320xbf16>) -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %462 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + xten_nn.output %461 : tensor<1x1x180x320xbf16> loc(#loc307) + } -> tensor<1x1x180x320xbf16> loc(#loc307) + return %449, %430, %410, %390, %458, %460 : tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16> loc(#loc319) + } loc(#loc319) + func.func @forward(%arg0: tensor<1x180x320x4xui8> {onnx.name = "src"} loc(unknown), %arg1: tensor<1x90x160x16xbf16> {onnx.name = "r1i"} loc(unknown), %arg2: tensor<1x45x80x20xbf16> {onnx.name = "r2i"} loc(unknown), %arg3: tensor<1x23x40x40xbf16> {onnx.name = "r3i"} loc(unknown), %arg4: tensor<1x12x20x64xbf16> {onnx.name = "r4i"} loc(unknown)) -> (tensor<1x180x320x3xbf16> {onnx.name = "fgr"}, tensor<1x180x320x1xbf16> {onnx.name = "pha"}, tensor<1x90x160x16xbf16> {onnx.name = "r1o"}, tensor<1x45x80x20xbf16> {onnx.name = "r2o"}, tensor<1x23x40x40xbf16> {onnx.name = "r3o"}, tensor<1x12x20x64xbf16> {onnx.name = "r4o"}) { + %0 = xten_nn.subgraph (%arg5 = %arg1: tensor<1x90x160x16xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %12 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc309) + %13 = tosa.transpose %arg5, %12 : (tensor<1x90x160x16xbf16>, tensor<4xi32>) -> tensor<1x16x90x160xbf16> loc(#loc309) + xten_nn.output %13 : tensor<1x16x90x160xbf16> loc(#loc309) + } -> tensor<1x16x90x160xbf16> loc(#loc309) + %1 = xten_nn.subgraph (%arg5 = %arg2: tensor<1x45x80x20xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %12 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc310) + %13 = tosa.transpose %arg5, %12 : (tensor<1x45x80x20xbf16>, tensor<4xi32>) -> tensor<1x20x45x80xbf16> loc(#loc310) + xten_nn.output %13 : tensor<1x20x45x80xbf16> loc(#loc310) + } -> tensor<1x20x45x80xbf16> loc(#loc310) + %2 = xten_nn.subgraph (%arg5 = %arg3: tensor<1x23x40x40xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %12 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc311) + %13 = tosa.transpose %arg5, %12 : (tensor<1x23x40x40xbf16>, tensor<4xi32>) -> tensor<1x40x23x40xbf16> loc(#loc311) + xten_nn.output %13 : tensor<1x40x23x40xbf16> loc(#loc311) + } -> tensor<1x40x23x40xbf16> loc(#loc311) + %3 = tosa.cast %arg0 {LayerName = "Cast_0", OutputName = "Cast_0"} : (tensor<1x180x320x4xui8>) -> tensor<1x180x320x4xbf16> loc(#loc308) + %4 = xten_nn.subgraph (%arg5 = %arg4: tensor<1x12x20x64xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %12 = "tosa.const"() <{value = dense<[0, 3, 1, 2]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc312) + %13 = tosa.transpose %arg5, %12 : (tensor<1x12x20x64xbf16>, tensor<4xi32>) -> tensor<1x64x12x20xbf16> loc(#loc312) + xten_nn.output %13 : tensor<1x64x12x20xbf16> loc(#loc312) + } -> tensor<1x64x12x20xbf16> loc(#loc312) + %5:6 = call @forward_outlined_part_0(%3, %0, %1, %2, %4) : (tensor<1x180x320x4xbf16>, tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>) -> (tensor<1x16x90x160xbf16>, tensor<1x20x45x80xbf16>, tensor<1x40x23x40xbf16>, tensor<1x64x12x20xbf16>, tensor<1x3x180x320xbf16>, tensor<1x1x180x320xbf16>) loc(#loc319) + %6 = xten_nn.subgraph (%arg5 = %5#4: tensor<1x3x180x320xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %12 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc313) + %13 = tosa.transpose %arg5, %12 : (tensor<1x3x180x320xbf16>, tensor<4xi32>) -> tensor<1x180x320x3xbf16> loc(#loc313) + xten_nn.output %13 : tensor<1x180x320x3xbf16> loc(#loc313) + } -> tensor<1x180x320x3xbf16> loc(#loc313) + %7 = xten_nn.subgraph (%arg5 = %5#3: tensor<1x64x12x20xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %12 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc314) + %13 = tosa.transpose %arg5, %12 : (tensor<1x64x12x20xbf16>, tensor<4xi32>) -> tensor<1x12x20x64xbf16> loc(#loc314) + xten_nn.output %13 : tensor<1x12x20x64xbf16> loc(#loc314) + } -> tensor<1x12x20x64xbf16> loc(#loc314) + %8 = xten_nn.subgraph (%arg5 = %5#2: tensor<1x40x23x40xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %12 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc315) + %13 = tosa.transpose %arg5, %12 : (tensor<1x40x23x40xbf16>, tensor<4xi32>) -> tensor<1x23x40x40xbf16> loc(#loc315) + xten_nn.output %13 : tensor<1x23x40x40xbf16> loc(#loc315) + } -> tensor<1x23x40x40xbf16> loc(#loc315) + %9 = xten_nn.subgraph (%arg5 = %5#1: tensor<1x20x45x80xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %12 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc316) + %13 = tosa.transpose %arg5, %12 : (tensor<1x20x45x80xbf16>, tensor<4xi32>) -> tensor<1x45x80x20xbf16> loc(#loc316) + xten_nn.output %13 : tensor<1x45x80x20xbf16> loc(#loc316) + } -> tensor<1x45x80x20xbf16> loc(#loc316) + %10 = xten_nn.subgraph (%arg5 = %5#0: tensor<1x16x90x160xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %12 = "tosa.const"() <{value = dense<[0, 2, 3, 1]> : tensor<4xi32>}> : () -> tensor<4xi32> loc(#loc317) + %13 = tosa.transpose %arg5, %12 : (tensor<1x16x90x160xbf16>, tensor<4xi32>) -> tensor<1x90x160x16xbf16> loc(#loc317) + xten_nn.output %13 : tensor<1x90x160x16xbf16> loc(#loc317) + } -> tensor<1x90x160x16xbf16> loc(#loc317) + %11 = xten_nn.subgraph (%arg5 = %5#5: tensor<1x1x180x320xbf16>) attributes {Message = "onnx.Transpose in function edge.", Reason = "CpuBecause"} { + %12 = tosa.reshape %arg5 {new_shape = array} : (tensor<1x1x180x320xbf16>) -> tensor<1x180x320x1xbf16> loc(#loc318) + xten_nn.output %12 : tensor<1x180x320x1xbf16> loc(#loc318) + } -> tensor<1x180x320x1xbf16> loc(#loc318) + return %6, %11, %10, %9, %8, %7 : tensor<1x180x320x3xbf16>, tensor<1x180x320x1xbf16>, tensor<1x90x160x16xbf16>, tensor<1x45x80x20xbf16>, tensor<1x23x40x40xbf16>, tensor<1x12x20x64xbf16> loc(#loc) + } loc(#loc) +} loc(#loc) +#loc1 = loc("Div_2") +#loc2 = loc("Sub_431") +#loc3 = loc("Sub_411") +#loc4 = loc("Sub_385") +#loc5 = loc("Sub_359") +#loc6 = loc("Div_16") +#loc7 = loc("Sub_14") +#loc8 = loc("Initializer_398") +#loc9 = loc("Slice_7") +#loc10 = loc("CompilerGeneratedLoc") +#loc11 = loc("Add_445") +#loc12 = loc("AveragePool_346") +#loc13 = loc("AveragePool_347") +#loc14 = loc("AveragePool_348") +#loc15 = loc("Conv_17") +#loc16 = loc("Add_19") +#loc17 = loc("Clip_22") +#loc18 = loc("Div_24") +#loc19 = loc("Mul_25") +#loc20 = loc("Conv_26") +#loc21 = loc("Relu_27") +#loc22 = loc("Conv_28") +#loc23 = loc("Add_29") +#loc24 = loc("Conv_30") +#loc25 = loc("Relu_31") +#loc26 = loc("Conv_32") +#loc27 = loc("Relu_33") +#loc28 = loc("Conv_34") +#loc29 = loc("Conv_35") +#loc30 = loc("Relu_36") +#loc31 = loc("Conv_37") +#loc32 = loc("Relu_38") +#loc33 = loc("Conv_39") +#loc34 = loc("Add_40") +#loc35 = loc("Conv_41") +#loc36 = loc("Relu_42") +#loc37 = loc("Conv_43") +#loc38 = loc("Relu_44") +#loc39 = loc("GlobalAveragePool_45") +#loc40 = loc("Conv_46") +#loc41 = loc("Relu_47") +#loc42 = loc("Conv_48") +#loc43 = loc("Add_50") +#loc44 = loc("Clip_53") +#loc45 = loc("Div_55") +#loc46 = loc("Mul_56") +#loc47 = loc("Conv_57") +#loc48 = loc("Conv_58") +#loc49 = loc("Relu_59") +#loc50 = loc("Conv_60") +#loc51 = loc("Relu_61") +#loc52 = loc("GlobalAveragePool_62") +#loc53 = loc("Conv_63") +#loc54 = loc("Relu_64") +#loc55 = loc("Conv_65") +#loc56 = loc("Add_67") +#loc57 = loc("Clip_70") +#loc58 = loc("Div_72") +#loc59 = loc("Mul_73") +#loc60 = loc("Conv_74") +#loc61 = loc("Add_75") +#loc62 = loc("Conv_76") +#loc63 = loc("Relu_77") +#loc64 = loc("Conv_78") +#loc65 = loc("Relu_79") +#loc66 = loc("GlobalAveragePool_80") +#loc67 = loc("Conv_81") +#loc68 = loc("Relu_82") +#loc69 = loc("Conv_83") +#loc70 = loc("Add_85") +#loc71 = loc("Clip_88") +#loc72 = loc("Div_90") +#loc73 = loc("Mul_91") +#loc74 = loc("Conv_92") +#loc75 = loc("Add_93") +#loc76 = loc("Conv_94") +#loc77 = loc("Add_96") +#loc78 = loc("Clip_99") +#loc79 = loc("Div_101") +#loc80 = loc("Mul_102") +#loc81 = loc("Conv_103") +#loc82 = loc("Add_105") +#loc83 = loc("Clip_108") +#loc84 = loc("Div_110") +#loc85 = loc("Mul_111") +#loc86 = loc("Conv_112") +#loc87 = loc("Conv_113") +#loc88 = loc("Add_115") +#loc89 = loc("Clip_118") +#loc90 = loc("Div_120") +#loc91 = loc("Mul_121") +#loc92 = loc("Conv_122") +#loc93 = loc("Add_124") +#loc94 = loc("Clip_127") +#loc95 = loc("Div_129") +#loc96 = loc("Mul_130") +#loc97 = loc("Conv_131") +#loc98 = loc("Add_132") +#loc99 = loc("Conv_133") +#loc100 = loc("Add_135") +#loc101 = loc("Clip_138") +#loc102 = loc("Div_140") +#loc103 = loc("Mul_141") +#loc104 = loc("Conv_142") +#loc105 = loc("Add_144") +#loc106 = loc("Clip_147") +#loc107 = loc("Div_149") +#loc108 = loc("Mul_150") +#loc109 = loc("Conv_151") +#loc110 = loc("Add_152") +#loc111 = loc("Conv_153") +#loc112 = loc("Add_155") +#loc113 = loc("Clip_158") +#loc114 = loc("Div_160") +#loc115 = loc("Mul_161") +#loc116 = loc("Conv_162") +#loc117 = loc("Add_164") +#loc118 = loc("Clip_167") +#loc119 = loc("Div_169") +#loc120 = loc("Mul_170") +#loc121 = loc("Conv_171") +#loc122 = loc("Add_172") +#loc123 = loc("Conv_173") +#loc124 = loc("Add_175") +#loc125 = loc("Clip_178") +#loc126 = loc("Div_180") +#loc127 = loc("Mul_181") +#loc128 = loc("Conv_182") +#loc129 = loc("Add_184") +#loc130 = loc("Clip_187") +#loc131 = loc("Div_189") +#loc132 = loc("Mul_190") +#loc133 = loc("GlobalAveragePool_191") +#loc134 = loc("Conv_192") +#loc135 = loc("Relu_193") +#loc136 = loc("Conv_194") +#loc137 = loc("Add_196") +#loc138 = loc("Clip_199") +#loc139 = loc("Div_201") +#loc140 = loc("Mul_202") +#loc141 = loc("Conv_203") +#loc142 = loc("Conv_204") +#loc143 = loc("Add_206") +#loc144 = loc("Clip_209") +#loc145 = loc("Div_211") +#loc146 = loc("Mul_212") +#loc147 = loc("Conv_213") +#loc148 = loc("Add_215") +#loc149 = loc("Clip_218") +#loc150 = loc("Div_220") +#loc151 = loc("Mul_221") +#loc152 = loc("GlobalAveragePool_222") +#loc153 = loc("Conv_223") +#loc154 = loc("Relu_224") +#loc155 = loc("Conv_225") +#loc156 = loc("Add_227") +#loc157 = loc("Clip_230") +#loc158 = loc("Div_232") +#loc159 = loc("Mul_233") +#loc160 = loc("Conv_234") +#loc161 = loc("Add_235") +#loc162 = loc("Conv_236") +#loc163 = loc("Add_238") +#loc164 = loc("Clip_241") +#loc165 = loc("Div_243") +#loc166 = loc("Mul_244") +#loc167 = loc("Conv_245") +#loc168 = loc("Add_247") +#loc169 = loc("Clip_250") +#loc170 = loc("Div_252") +#loc171 = loc("Mul_253") +#loc172 = loc("GlobalAveragePool_254") +#loc173 = loc("Conv_255") +#loc174 = loc("Relu_256") +#loc175 = loc("Conv_257") +#loc176 = loc("Add_259") +#loc177 = loc("Clip_262") +#loc178 = loc("Div_264") +#loc179 = loc("Mul_265") +#loc180 = loc("Conv_266") +#loc181 = loc("Conv_267") +#loc182 = loc("Add_269") +#loc183 = loc("Clip_272") +#loc184 = loc("Div_274") +#loc185 = loc("Mul_275") +#loc186 = loc("Conv_276") +#loc187 = loc("Add_278") +#loc188 = loc("Clip_281") +#loc189 = loc("Div_283") +#loc190 = loc("Mul_284") +#loc191 = loc("GlobalAveragePool_285") +#loc192 = loc("Conv_286") +#loc193 = loc("Relu_287") +#loc194 = loc("Conv_288") +#loc195 = loc("Add_290") +#loc196 = loc("Clip_293") +#loc197 = loc("Div_295") +#loc198 = loc("Mul_296") +#loc199 = loc("Conv_297") +#loc200 = loc("Add_298") +#loc201 = loc("Conv_299") +#loc202 = loc("Add_301") +#loc203 = loc("Clip_304") +#loc204 = loc("Div_306") +#loc205 = loc("Mul_307") +#loc206 = loc("Conv_308") +#loc207 = loc("Add_310") +#loc208 = loc("Clip_313") +#loc209 = loc("Div_315") +#loc210 = loc("Mul_316") +#loc211 = loc("GlobalAveragePool_317") +#loc212 = loc("Conv_318") +#loc213 = loc("Relu_319") +#loc214 = loc("Conv_320") +#loc215 = loc("Add_322") +#loc216 = loc("Clip_325") +#loc217 = loc("Div_327") +#loc218 = loc("Mul_328") +#loc219 = loc("Conv_329") +#loc220 = loc("Add_330") +#loc221 = loc("Conv_331") +#loc222 = loc("Add_333") +#loc223 = loc("Clip_336") +#loc224 = loc("Div_338") +#loc225 = loc("Mul_339") +#loc226 = loc("GlobalAveragePool_342") +#loc227 = loc("Conv_343") +#loc228 = loc("Sigmoid_344") +#loc229 = loc("Mul_345") +#loc230 = loc("Conv_340") +#loc231 = loc("Relu_341") +#loc232 = loc("Split_349") +#loc233 = loc("Concat_350") +#loc234 = loc("Conv_351") +#loc235 = loc("Sigmoid_352") +#loc236 = loc("Split_353") +#loc237 = loc("Mul_360") +#loc238 = loc("Mul_354") +#loc239 = loc("Concat_355") +#loc240 = loc("Conv_356") +#loc241 = loc("Tanh_357") +#loc242 = loc("Mul_361") +#loc243 = loc("Add_362") +#loc244 = loc("Concat_363") +#loc245 = loc("Resize_365") +#loc246 = loc("Slice_371") +#loc247 = loc("Concat_372") +#loc248 = loc("Conv_373") +#loc249 = loc("Relu_374") +#loc250 = loc("Split_375") +#loc251 = loc("Concat_376") +#loc252 = loc("Conv_377") +#loc253 = loc("Sigmoid_378") +#loc254 = loc("Split_379") +#loc255 = loc("Mul_386") +#loc256 = loc("Mul_380") +#loc257 = loc("Concat_381") +#loc258 = loc("Conv_382") +#loc259 = loc("Tanh_383") +#loc260 = loc("Mul_387") +#loc261 = loc("Add_388") +#loc262 = loc("Concat_389") +#loc263 = loc("Resize_391") +#loc264 = loc("Slice_397") +#loc265 = loc("Concat_398") +#loc266 = loc("Conv_399") +#loc267 = loc("Relu_400") +#loc268 = loc("Split_401") +#loc269 = loc("Concat_402") +#loc270 = loc("Conv_403") +#loc271 = loc("Sigmoid_404") +#loc272 = loc("Split_405") +#loc273 = loc("Mul_412") +#loc274 = loc("Mul_406") +#loc275 = loc("Concat_407") +#loc276 = loc("Conv_408") +#loc277 = loc("Tanh_409") +#loc278 = loc("Mul_413") +#loc279 = loc("Add_414") +#loc280 = loc("Concat_415") +#loc281 = loc("Resize_417") +#loc282 = loc("Concat_418") +#loc283 = loc("Conv_419") +#loc284 = loc("Relu_420") +#loc285 = loc("Split_421") +#loc286 = loc("Concat_422") +#loc287 = loc("Conv_423") +#loc288 = loc("Sigmoid_424") +#loc289 = loc("Split_425") +#loc290 = loc("Mul_432") +#loc291 = loc("Mul_426") +#loc292 = loc("Concat_427") +#loc293 = loc("Conv_428") +#loc294 = loc("Tanh_429") +#loc295 = loc("Mul_433") +#loc296 = loc("Add_434") +#loc297 = loc("Concat_435") +#loc298 = loc("Resize_437") +#loc299 = loc("Concat_438") +#loc300 = loc("Conv_439") +#loc301 = loc("Relu_440") +#loc302 = loc("Conv_441") +#loc303 = loc("Relu_442") +#loc304 = loc("Conv_443") +#loc305 = loc("Split_444") +#loc306 = loc("Clip_446") +#loc307 = loc("Clip_447") +#loc313 = loc("Transpose_452") +#loc314 = loc("Transpose_451") +#loc315 = loc("Transpose_450") +#loc316 = loc("Transpose_449") +#loc317 = loc("Transpose_448") +#loc318 = loc("Transpose_453") +#loc319 = loc(fused[#loc1, #loc2, #loc3, #loc4, #loc5, #loc6, #loc7, #loc8, #loc9, #loc10, #loc11, #loc12, #loc13, #loc14, #loc15, #loc16, #loc17, #loc18, #loc19, #loc20, #loc21, #loc22, #loc23, #loc24, #loc25, #loc26, #loc27, #loc28, #loc29, #loc30, #loc31, #loc32, #loc33, #loc34, #loc35, #loc36, #loc37, #loc38, #loc39, #loc40, #loc41, #loc42, #loc43, #loc44, #loc45, #loc46, #loc47, #loc48, #loc49, #loc50, #loc51, #loc52, #loc53, #loc54, #loc55, #loc56, #loc57, #loc58, #loc59, #loc60, #loc61, #loc62, #loc63, #loc64, #loc65, #loc66, #loc67, #loc68, #loc69, #loc70, #loc71, #loc72, #loc73, #loc74, #loc75, #loc76, #loc77, #loc78, #loc79, #loc80, #loc81, #loc82, #loc83, #loc84, #loc85, #loc86, #loc87, #loc88, #loc89, #loc90, #loc91, #loc92, #loc93, #loc94, #loc95, #loc96, #loc97, #loc98, #loc99, #loc100, #loc101, #loc102, #loc103, #loc104, #loc105, #loc106, #loc107, #loc108, #loc109, #loc110, #loc111, #loc112, #loc113, #loc114, #loc115, #loc116, #loc117, #loc118, #loc119, #loc120, #loc121, #loc122, #loc123, #loc124, #loc125, #loc126, #loc127, #loc128, #loc129, #loc130, #loc131, #loc132, #loc133, #loc134, #loc135, #loc136, #loc137, #loc138, #loc139, #loc140, #loc141, #loc142, #loc143, #loc144, #loc145, #loc146, #loc147, #loc148, #loc149, #loc150, #loc151, #loc152, #loc153, #loc154, #loc155, #loc156, #loc157, #loc158, #loc159, #loc160, #loc161, #loc162, #loc163, #loc164, #loc165, #loc166, #loc167, #loc168, #loc169, #loc170, #loc171, #loc172, #loc173, #loc174, #loc175, #loc176, #loc177, #loc178, #loc179, #loc180, #loc181, #loc182, #loc183, #loc184, #loc185, #loc186, #loc187, #loc188, #loc189, #loc190, #loc191, #loc192, #loc193, #loc194, #loc195, #loc196, #loc197, #loc198, #loc199, #loc200, #loc201, #loc202, #loc203, #loc204, #loc205, #loc206, #loc207, #loc208, #loc209, #loc210, #loc211, #loc212, #loc213, #loc214, #loc215, #loc216, #loc217, #loc218, #loc219, #loc220, #loc221, #loc222, #loc223, #loc224, #loc225, #loc226, #loc227, #loc228, #loc229, #loc230, #loc231, #loc232, #loc233, #loc234, #loc235, #loc236, #loc237, #loc238, #loc239, #loc240, #loc241, #loc242, #loc243, #loc244, #loc245, #loc246, #loc247, #loc248, #loc249, #loc250, #loc251, #loc252, #loc253, #loc254, #loc255, #loc256, #loc257, #loc258, #loc259, #loc260, #loc261, #loc262, #loc263, #loc264, #loc265, #loc266, #loc267, #loc268, #loc269, #loc270, #loc271, #loc272, #loc273, #loc274, #loc275, #loc276, #loc277, #loc278, #loc279, #loc280, #loc281, #loc282, #loc283, #loc284, #loc285, #loc286, #loc287, #loc288, #loc289, #loc290, #loc291, #loc292, #loc293, #loc294, #loc295, #loc296, #loc297, #loc298, #loc299, #loc300, #loc301, #loc302, #loc303, #loc304, #loc305, #loc306, #loc307]) +#loc320 = loc(fused[#loc7, #loc8]) +#loc321 = loc(fused[#loc11, #loc9, #loc12]) +#loc322 = loc(fused[#loc9, #loc12, #loc11]) +#loc323 = loc(fused[#loc20, #loc21]) +#loc324 = loc(fused[#loc22, #loc23]) +#loc325 = loc(fused[#loc24, #loc25]) +#loc326 = loc(fused[#loc26, #loc27]) +#loc327 = loc(fused[#loc29, #loc30]) +#loc328 = loc(fused[#loc31, #loc32]) +#loc329 = loc(fused[#loc33, #loc34]) +#loc330 = loc(fused[#loc35, #loc36]) +#loc331 = loc(fused[#loc37, #loc38]) +#loc332 = loc(fused[#loc40, #loc41]) +#loc333 = loc(fused[#loc48, #loc49]) +#loc334 = loc(fused[#loc50, #loc51]) +#loc335 = loc(fused[#loc53, #loc54]) +#loc336 = loc(fused[#loc60, #loc61]) +#loc337 = loc(fused[#loc62, #loc63]) +#loc338 = loc(fused[#loc64, #loc65]) +#loc339 = loc(fused[#loc67, #loc68]) +#loc340 = loc(fused[#loc74, #loc75]) +#loc341 = loc(fused[#loc97, #loc98]) +#loc342 = loc(fused[#loc109, #loc110]) +#loc343 = loc(fused[#loc121, #loc122]) +#loc344 = loc(fused[#loc132, #loc133]) +#loc345 = loc(fused[#loc134, #loc135]) +#loc346 = loc(fused[#loc151, #loc152]) +#loc347 = loc(fused[#loc153, #loc154]) +#loc348 = loc(fused[#loc160, #loc161]) +#loc349 = loc(fused[#loc171, #loc172]) +#loc350 = loc(fused[#loc173, #loc174]) +#loc351 = loc(fused[#loc190, #loc191]) +#loc352 = loc(fused[#loc192, #loc193]) +#loc353 = loc(fused[#loc199, #loc200]) +#loc354 = loc(fused[#loc210, #loc211]) +#loc355 = loc(fused[#loc212, #loc213]) +#loc356 = loc(fused[#loc219, #loc220]) +#loc357 = loc(fused[#loc225, #loc226]) +#loc358 = loc(fused[#loc230, #loc231, #loc229]) +#loc359 = loc(fused[#loc230, #loc231]) +#loc360 = loc(fused[#loc248, #loc249]) +#loc361 = loc(fused[#loc266, #loc267]) +#loc362 = loc(fused[#loc283, #loc284]) +#loc363 = loc(fused[#loc300, #loc301]) +#loc364 = loc(fused[#loc302, #loc303]) diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/aie_unsupported_original_ops.json b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/aie_unsupported_original_ops.json new file mode 100644 index 0000000000000000000000000000000000000000..66e0d29a6848fc5b9c7ec338ac2464b80c65870a --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/aie_unsupported_original_ops.json @@ -0,0 +1,13 @@ +[ + "Cast_0", + "Transpose_10", + "Transpose_11", + "Transpose_12", + "Transpose_448", + "Transpose_449", + "Transpose_450", + "Transpose_451", + "Transpose_452", + "Transpose_453", + "Transpose_9" +] diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/aie_unsupported_original_ops_with_reasons.json b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/aie_unsupported_original_ops_with_reasons.json new file mode 100644 index 0000000000000000000000000000000000000000..a299686af03f38e12db9c508d4adcf8e7bd6dfb1 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/aie_unsupported_original_ops_with_reasons.json @@ -0,0 +1,46 @@ +[ + { + "Operation": "Cast_0", + "Reason": "Operation is not implemented on AIE" + }, + { + "Operation": "Transpose_10", + "Reason": "onnx.Transpose in function edge." + }, + { + "Operation": "Transpose_11", + "Reason": "onnx.Transpose in function edge." + }, + { + "Operation": "Transpose_12", + "Reason": "onnx.Transpose in function edge." + }, + { + "Operation": "Transpose_448", + "Reason": "onnx.Transpose in function edge." + }, + { + "Operation": "Transpose_449", + "Reason": "onnx.Transpose in function edge." + }, + { + "Operation": "Transpose_450", + "Reason": "onnx.Transpose in function edge." + }, + { + "Operation": "Transpose_451", + "Reason": "onnx.Transpose in function edge." + }, + { + "Operation": "Transpose_452", + "Reason": "onnx.Transpose in function edge." + }, + { + "Operation": "Transpose_453", + "Reason": "onnx.Transpose in function edge." + }, + { + "Operation": "Transpose_9", + "Reason": "onnx.Transpose in function edge." + } +] diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/fail_safe_summary.json b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/fail_safe_summary.json new file mode 100644 index 0000000000000000000000000000000000000000..16f3b4c3a78456d9d15c9e557420a0931309c720 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/fail_safe_summary.json @@ -0,0 +1,6 @@ +{ + "offload_map": { + "AIE": 96.875, + "CPU": 3.125 + } +} \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/partition-info.json b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/partition-info.json new file mode 100644 index 0000000000000000000000000000000000000000..674184c6ebbb9c50fcb109e39c1efa124c12e9d2 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/partition-info.json @@ -0,0 +1,19 @@ +{ + "compiler": "flexml.compile", + "dry_run": false, + "sha1sum": "4b0d733eefdc636c83713aa34f2a653a540a924c", + "device": "stx", + "status": "completed", + "num_aie_partitions": 1, + "xrt": "na", + "flexml": "1.5.0.dev20250320125000+g1d4c9c71", + "os": "Ubuntu 22.04", + "target_os": "linux", + "modelname": "OnnxModel", + "ai_analyzer_profiling_enabled": false, + "runner_type": "hw", + "output_type": "incore-chaining", + "aie_partition_call_order": [ + "forward_outlined_part_0" + ] +} \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/tilingEngine/select-layout-options.json b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/tilingEngine/select-layout-options.json new file mode 100644 index 0000000000000000000000000000000000000000..f699946a290c9f9f53633f1e1a87905106635c55 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/tilingEngine/select-layout-options.json @@ -0,0 +1 @@ +{"blockOptions":{"useLegacyBackendConstraints":false,"constraintFilter":"","disableEstimationCache":false,"disablePerformanceConstraints":false,"disableSolutionCache":false,"enableCoreTileBDProgramming":false,"exhaustive":false,"forceSpilling":false,"geometryAndBufferLayout":"","l2FmDoubleBuffering":false,"optimizingSolver":true,"solutionSeeds":false,"stepsBeforeFailsafeExploration":40000,"tileIcwForWtsDoubleBuffering":true,"tilingRecipe":"","backend":{"optimizeLevel":2,"shimDmaForWeights":1,"packetOrderedMerge":true}},"controlOptions":{"debugConstraints":false,"dumpInputs":false,"dumpSolution":false,"dumpAllDesigns":false,"failOnInvalidOverride":false,"outputDirectory":"./segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml","teSolutionCacheDir":"./segmentation_1_4_0_fp32_combined/cache/1d4c9c71f3a11a5a3ebd0c4f9b9709fb907ad630","allowFailures":false}} \ No newline at end of file diff --git a/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/tilingEngine/unwrapping-options.json b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/tilingEngine/unwrapping-options.json new file mode 100644 index 0000000000000000000000000000000000000000..f699946a290c9f9f53633f1e1a87905106635c55 --- /dev/null +++ b/segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml/tilingEngine/unwrapping-options.json @@ -0,0 +1 @@ +{"blockOptions":{"useLegacyBackendConstraints":false,"constraintFilter":"","disableEstimationCache":false,"disablePerformanceConstraints":false,"disableSolutionCache":false,"enableCoreTileBDProgramming":false,"exhaustive":false,"forceSpilling":false,"geometryAndBufferLayout":"","l2FmDoubleBuffering":false,"optimizingSolver":true,"solutionSeeds":false,"stepsBeforeFailsafeExploration":40000,"tileIcwForWtsDoubleBuffering":true,"tilingRecipe":"","backend":{"optimizeLevel":2,"shimDmaForWeights":1,"packetOrderedMerge":true}},"controlOptions":{"debugConstraints":false,"dumpInputs":false,"dumpSolution":false,"dumpAllDesigns":false,"failOnInvalidOverride":false,"outputDirectory":"./segmentation_1_4_0_fp32_combined/vaiml_partition_fe.flexml","teSolutionCacheDir":"./segmentation_1_4_0_fp32_combined/cache/1d4c9c71f3a11a5a3ebd0c4f9b9709fb907ad630","allowFailures":false}} \ No newline at end of file diff --git a/vitisai_config.json b/vitisai_config.json new file mode 100644 index 0000000000000000000000000000000000000000..56beb25346878254610e4f4fca3009c1cdf1d78e --- /dev/null +++ b/vitisai_config.json @@ -0,0 +1,19 @@ +{ + "passes": [ + { + "name": "init", + "plugin": "vaip-pass_init" + }, + { + "name": "vaiml_partition", + "plugin": "vaip-pass_vaiml_partition", + "vaiml_config": { + "keep_outputs": true, + "device": "stx", + "optimize_level": 2, + "enable_f32_to_bf16_conversion": true + } + } + ] +} +